I am trying to run the Tom Whites' Chapter 2 example
When I run the command:
hadoop MaxTemperature input/ncdc/sample.txt output
11/12/31 18:08:28 INFO mapred.JobClient: Cleaning up the staging area hdfs://localhost:9000/tmp/hadoop-mymac/mapred/staging/mymac/.staging/job_201112311807_0001
11/12/31 18:08:28 ERROR security.UserGroupInformation: PriviledgedActionException as:mymac (auth:SIMPLE) cause:org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does not exist: hdfs://localhost:9000/user/mymac/input/ncdc/sample.txt
Exception in thread "main" org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does not exist: hdfs://localhost:9000/user/mymac/input/ncdc/sample.txt
hdfs-site.xml files are configured to use
localhost:9000. If this isn't what you expect (which is what I get from you post's title), what did you expect?
What version of Hadoop are we talking about? How did you install your Hadoop distribution? From your other question and the config files, I'm guessing you used CHD4. If you look over the instructions from Cloudera, can you see if you missed anything?
Before starting Hadoop, did you format HDFS?
$ hadoop namenode -format
Then, after starting Hadoop, do you get anything other than INFO messages?
Did you copy the input data into HDFS?
$ hadoop dfs -put /tmp/my/input/data input
Finally, what do you get from simple HDFS commands such as:
$ hadoop dfs -ls /
UPDATE: Run Word Count
hadoop dfs -ls /should work.
hadoop dfs -put text_files input_folder
hadoop dfs -ls .to see if your files got copied correctly.
hadoop-examples-X.Y.Z.jarfile on your system.
Navigate to whatever directory it's in, and run:
$ hadoop jar hadoop-examples-*.jar WordCount input_folder output_folder.
You should see the progress of the MapReduce application.
hadoop dfs -cat output_folder/*.