I have a very long file and I only need parts, a slice, of it.
There is new data coming in so the file will potentially get longer.
To load the data from the CSV I use numpy.genfromtxt
np.genfromtxt(filename, usecols={col}, delimiter=",", skip_header=skip_head)
This cuts off a certain parts of the file in the beginning which already substantially speeds up the process of loading the data.
But I can't use skip_footer
in the end to cut off the part after my slice that I want to use.
What I want is to only load a certain number of rows. e.g. lets say I skip the first 100 rows, then load the next 50 rows and skip the rest afterwards.
edit: I am using Python 3.4
edit: sample file: http://www.file-upload.net/download-10819938/sample.txt.html
You could get the slice using itertools, taking the column using itemgetter:
import numpy as np
from operator import itemgetter
import csv
with open(filename) as f:
from itertools import islice,imap
r = csv.reader(f)
np.genfromtxt(imap(itemgetter(1),islice(r, start, end+1)))
For python3, you can use fromiter
with the code above you need to specify the dtype:
import numpy as np
from operator import itemgetter
import csv
with open("sample.txt") as f:
from itertools import islice
r = csv.reader(f)
print(np.fromiter(map(itemgetter(0), islice(r, start, end+1)), dtype=float))
As in the other answer you can also pass the islice object directly to genfromtxt but for python3 you will need to open the file in binary mode:
with open("sample.txt", "rb") as f:
from itertools import islice
print(np.genfromtxt(islice(f, start, end+1), delimiter=",", usecols=cols))
Interestingly, for multiple columns using itertools.chain and reshaping is over twice as efficient if all your dtypes are the same:
from itertools import islice,chain
with open("sample.txt") as f:
r = csv.reader(f)
arr =np.fromiter(chain.from_iterable(map(itemgetter(0, 4, 10),
islice(r, 4, 10))), dtype=float).reshape(6, -1)
On you sample file:
In [27]: %%timeit
with open("sample.txt", "rb") as f:
(np.genfromtxt(islice(f, 4, 10), delimiter=",", usecols=(0, 4, 10),dtype=float))
....:
10000 loops, best of 3: 179 µs per loop
In [28]: %%timeit
with open("sample.txt") as f:
r = csv.reader(f) (np.fromiter(chain.from_iterable(map(itemgetter(0, 4, 10), islice(r, 4, 10))), dtype=float).reshape(6, -1))
10000 loops, best of 3: 86 µs per loop
Following this example, you should be able to use itertools.islice
, without needing imap
, map
or csv.reader
:
import numpy as np
import itertools
with open('sample.txt') as f:
# this will skip 100 lines, then read the next 50
d=np.genfromtxt(itertools.islice(f,100,150),delimiter=',',usecols={cols})