I was working on a script which reading a folder of files(each of size ranging from 20 MB to 100 MB), modifies some data in each line, and writes back to a copy of the file.
with open(inputPath, 'r+') as myRead:
my_list = myRead.readlines()
new_my_list = clean_data(my_list)
with open(outPath, 'w+') as myWrite:
tempT = time.time()
myWrite.writelines('\n'.join(new_my_list) + '\n')
print(time.time() - tempT)
print(inputPath, 'Cleaning Complete.')
On running this code with a 90 MB file (~900,000 lines), it printed 140 seconds as the time taken to write to the file. Here I used writelines()
. So I searched for different ways to improve file writing speed, and in most of the articles that I read, it said write()
and writelines()
should not show any difference since I am writing a single concatenated string. I also checked the time taken for only the following statement:
new_string = '\n'.join(new_my_list) + '\n'
And it took only 0.4 seconds, so the large time taken was not because of creating the list.
Just to try out write()
I tried this code:
with open(inputPath, 'r+') as myRead:
my_list = myRead.readlines()
new_my_list = clean_data(my_list)
with open(outPath, 'w+') as myWrite:
tempT = time.time()
myWrite.write('\n'.join(new_my_list) + '\n')
print(time.time() - tempT)
print(inputPath, 'Cleaning Complete.')
And it printed 2.5 seconds. Why is there such a large difference in the file writing time for write()
and writelines()
even though it is the same data? Is this normal behaviour or is there something wrong in my code? The output file seems to be the same for both cases, so I know that there is no loss in data.
file.writelines()
expects an iterable of strings. It then proceeds to loop and callfile.write()
for each string in the iterable. In Python, the method does this:You are passing in a single large string, and a string is an iterable of strings too. When iterating you get individual characters, strings of length 1. So in effect you are making
len(data)
separate calls tofile.write()
. And that is slow, because you are building up a write buffer a single character at a time.Don't pass in a single string to
file.writelines()
. Pass in a list or tuple or other iterable instead.You could send in individual lines with added newline in a generator expression, for example:
Now, if you could make
clean_data()
a generator, yielding cleaned lines, you could stream data from the input file, through your data cleaning generator, and out to the output file without using any more memory than is required for the read and write buffers and however much state is needed to clean your lines:In addition, I'd consider updating
clean_data()
to emit lines with newlines included.'write(arg)' method expects string as its argument. So once it calls, it will directly writes. this is the reason it is much faster. where as if you are using
writelines()
method, it expects list of string as iterator. so even if you are sending data towritelines
, it assumes that it got iterator and it tries to iterate over it. so since it is an iterator it will take some time to iterate over and write it.Is that clear ?
as a complement to Martijn answer, the best way would be to avoid to build the list using
join
in the first placeJust pass a generator comprehension to
writelines
, adding the newline in the end: no unnecessary memory allocation and no loop (besides the comprehension)