I am having a .csv file with 1009725 rows and 85 columns in it. I am trying to read it with Microsoft R open. I used the following command,
data= read.csv("C:/Users/username/Desktop/data.csv")
But the number of rows that gets read are only 617200
(~65%). I am not sure what is the reason behind the file not getting completely read. The data columns are mostly integers like IDs, values and some of them have text. Can anybody help me in diagnosing the problem?
Also even for the 617200
rows, I am having so much of performance issues. Even the basic command such as,
nrow(data)
or length(unique(data$column1))
puts the Rstudio to Not responding status. The configuration of my system is 16GB RAM and i7 quad core processor. I feel this should be sufficient to crunch this data. But why am I not able to even run even the basic commands of the partial data that gets read ? Can anybody help me in diagnosing both the problems?
Thanks