Saturday, 14 September 2013

Python Script To Upload 300,000 files- Sqlite3 or massive list?

Python Script To Upload 300,000 files- Sqlite3 or massive list?

I am writing a script that will upload files to a cloud account. There
will be various directories containing the files, but only one
depth...there will be no nested/directories inside of directories. Each
directory will be a container in which the files will go in. Sometimes the
files may be as large as 300,000 files. I will be using multiprocessing.
I would like to keep track of the filenames, outoutput information,
returns codes using sqlite, so I had a few questions:
1) If I only had sqlite3 run in memory rather than as flat files(since I
only need the info untill I'm done with the script) would it bloat the
memory? 2) Would there be a major performance impact using sqlite3 as
opposed to keeping track with a massive list of lists or a dictionary of
lists?

No comments:

Post a Comment