First you have to export the url
export url
"http://localhost.localdomain:1100/oozie"
Then submit the job
oozie job -oozie
http://localhost.localdomain:11000/oozie -config coord.properties -submit
For running a simple Oozie job you have to run a job.properties file but for time triggered you have run coord.properties file
Directory Structure for using Oozie:
ProjectFolder
Lib
Jars that you want
to run. If you have made a mapreduce job than make a jar from it and place It
here
DataFolder
Input output
files
CoordinatorFolder
Coord.xml
Job.properties
file ( should be on the local file system all other should be on hdfs,namenode
jobtracker are mentioned here)
Workflow.xml ( actual workflow file that specifies which job to run
which class to run and parameters are also specified here)
Coordinator Job
1-
The workflow job is started after the predicate
is satisfied. A predicate can reference to data, time and/or external events
2-
The outputs of last 4 runs of a workflow that
runs every 15 minutes become the input of another workflow that runs every 60
minutes. Chaining together these workflows result it is referred as a data
application pipeline
Useful links:
http://hadooped.blogspot.com/2013/06/apache-oozie-part-1-workflow-with-hdfs.html
ReplyDeleteFree Study Material R Analytics & Hadoop - http://www.hub4tech.com/big-data
Machine Learning Tutorial- http://www.hub4tech.com/machine-learning-tutorial
Hive Tutorial- http://www.hub4tech.com/hive-tutorial
Oozie Tutorial- http://www.hub4tech.com/oozie-tutorial
Pig Tutorial- http://www.hub4tech.com/pig-tutorial
HDFS Tutorial- http://www.hub4tech.com/hdfs-tutorial