|
DATA MINING
Desktop Survival Guide by Graham Williams |
|
|||
Reading a Large File |
Suppose we have a very large dataset to score. We may not be able to load all of the data into R. One approach will be to partition the data and read blocks at a time to score and save results. The following snippet gives a hint as to how this might be done.
f <- file("hugedata.csv", "r")
skip <- 1562739
while (skip > 10000) {
junk <- readLines(f, 10000)
skip <- skip - 10000
}
junk <- readLines(f, skip)
readLines(f, 1)
|