300x250 AD TOP

Powered by Blogger.

Thursday, 12 June 2014

Tagged under: , ,

How To Run Oozie Coordinator Jobs

First you have to export the url
export url "http://localhost.localdomain:1100/oozie"

Then submit the job
oozie job -oozie http://localhost.localdomain:11000/oozie -config coord.properties -submit

For running a simple Oozie job you have to run a job.properties file but for time triggered you have run coord.properties file


Directory Structure for using Oozie:
ProjectFolder
                Lib
Jars that you want to run. If you have made a mapreduce job than make a jar from it and place It here
                DataFolder
                                Input output files
                CoordinatorFolder
                                Coord.xml
Job.properties file ( should be on the local file system all other should be on hdfs,namenode jobtracker are mentioned here)
                Workflow.xml ( actual workflow file that specifies which job to run which class to run and parameters are also specified here)
Coordinator Job
1-      The workflow job is started after the predicate is satisfied. A predicate can reference to data, time and/or external events
2-      The outputs of last 4 runs of a workflow that runs every 15 minutes become the input of another workflow that runs every 60 minutes. Chaining together these workflows result it is referred as a data application pipeline

Useful links:
http://hadooped.blogspot.com/2013/06/apache-oozie-part-1-workflow-with-hdfs.html

1 comments: