我正在尝试使用bulkloader功能将一个中等大小的csv文件上传到google应用程序引擎,但它似乎在某种程度上失败了,结果如下:
[INFO ] Logging to bulkloader-log-20110328.181531
[INFO ] Throttling transfers:
[INFO ] Bandwidth: 250000 bytes/second
[INFO ] HTTP connections: 8/second
[INFO ] Entities inserted/fetched/modified: 20/second
[INFO ] Batch Size: 10
[INFO ] Opening database: bulkloader-progress-20110328.181531.sql3
[INFO ] Connecting to notmyrealappname.appspot.com/_ah/remote_api
[INFO ] Starting import; maximum 10 entities per post
...............................................................[INFO ] Unexpected thread death: WorkerThread-7
[INFO ] An error occurred. Shutting down...
.........[ERROR ] Error in WorkerThread-7: <urlopen error [Errno -2] Name or service not known>
[INFO ] 1740 entites total, 0 previously transferred
[INFO ] 720 entities (472133 bytes) transferred in 32.3 seconds
[INFO ] Some entities not successfully transferred它上传了我试图上传的19k个条目中的大约700个,我想知道为什么它失败了。我检查了csv文件中的错误,比如额外的逗号可能会使python csv阅读器脱离,并且非ascii字符已经被剥离。
发布于 2011-11-02 10:05:39
取消批处理限制(batch_size)和rps限制(rps_limit)有效,我使用1000作为批处理大小,rps限制为500:
appcfg.py upload_data --url= --application= --filename= --email= --batch_size=1000 --rps_limit=500https://stackoverflow.com/questions/5466900
复制相似问题