import scraperwiki
import urllib2, lxml.etree
url = 'http://eci.nic.in/eci_main/statisticalreports/SE_1998/StatisticalReport-DEL98.pdf'
pdfdata = urllib2.urlopen(url).read()
xmldata = scraperwiki.pdftoxml(pdfdata)
root = lxml.etree.fromstring(xmldata)
# how many pages in PDF
pages = list(root)
print "There are",len(pages),"pages"
#from page 86 to 107
for page in pages[86:107]:
for el in page:
data = {}
if el.tag == "text":
if int(el.attrib['left']) < 215: data = { 'Rank': el.text }
elif int(el.attrib['left']) < 230: data['Name'] = el.text
elif int(el.attrib['left']) < 592: data['Sex'] = el.text
elif int(el.attrib['left']) < 624: data['Party'] = el.text
elif int(el.attrib['left']) < 750: data['Votes'] = el.text
elif int(el.attrib['left']) < 801: data['Percentage'] = el.text
print data
现在我想知道如何将这些数据保存在scraperwiki
. 我尝试了一些命令,例如
scraperwiki.sqlite.save(unique_keys=[], table_name='ecidata1998', data=data)
但是当我检查数据集时,他们没有给我所需的结果,代码或最后一条语句是否有问题。请帮忙。Python 编程和 Scraperwiki 的新手。