python - Downloading a Large Number of Files from S3 -


What is the fastest way to obtain a large number of files from Python S3 (relatively small 10-50kB) from Python? (In the order of 200,000-million files)

I am currently using beto to generate signed URLs, and I am using PyCURL to get files one by one.

Will some types of concurrency help? PyCurl.CurlMulti Object?

I'm open to all the tips thanks!

In the case of dragon, because it is IO bound, multiple threads will use the CPU, but it probably Will use only one core if you have multiple cores, you may want to consider the new module. However, you want to use multiple threads for each process.

If you use many threads, then this is a good candidate for class.


Comments

Popular posts from this blog

c++ - Linux and clipboard -

Visual Studio 2005: How to speed up builds when a VSMDI is open? -

booting ubuntu from usb using virtualbox -