We're trying to index the contents of a 3gb csv (not located on the box running the application). We're able to read the file with a BufferedReader, but we run into issues when we try to read efficiently. Someone suggested that we hash an id field with the contents of the line.
This seems like a good idea, but I cannot figure out how we can "buffer write" our hash map to a file. Seems like object writer only takes one massive "dump" object...
Anyone know of a way that we can continuously put entries to the same external hash map, and then read from these entires?
Thanks!