I am very new to hadoop and hbase and have some conceptual questions that are tripping me up during every tutorial I've found.
I have hadoop and hbase running on a single node within a ubuntu VM on my win 7 system. I have a csv file that I would like to load into a single hbase table.
The columns are: loan_number, borrower_name, current_distribution_date, loan_amount
I know that I need to write a MapReduce job to load this said csv file into hbase. The following tutorial describes the Java needed to write this MapReduce job. http://salsahpc.indiana.edu/ScienceCloud/hbase_hands_on_1.htm
What I'm missing is:
Where do I save these files and where do I compile them? Should I compile this on my win 7 machine running visual studio 12 and then move it to the ubuntu vm?
I read this SO question and answers but I guess I'm still missing the basics: Loading CSV File into Hbase table using MapReduce
I can't find anything covering these basic hadoop/hbase logistics. Any help would be greatly appreciated.