forked from D-Net/dnet-hadoop
Add ActionSet step
This commit is contained in:
parent
b5c252865c
commit
9dc8f0f05f
|
@ -63,6 +63,9 @@ oozieWorkflowPath=user/ilias.kanellos/workflow_example/
|
|||
# The directory where the workflow data is/should be stored
|
||||
workflowDataDir=user/ilias.kanellos/ranking_workflow
|
||||
|
||||
# Directory where json data containing scores will be output
|
||||
bipScorePath=${workflowDataDir}/openaire_universe_scores/
|
||||
|
||||
# Directory where dataframes are checkpointed
|
||||
checkpointDir=${nameNode}/${workflowDataDir}/check/
|
||||
|
||||
|
@ -84,3 +87,6 @@ wfAppPath=${nameNode}/${oozieWorkflowPath}
|
|||
# The following is needed as a property of a workflow
|
||||
oozie.wf.application.path=${wfAppPath}
|
||||
|
||||
# Path where the final output should be?
|
||||
actionSetOutputPath=${workflowDataDir}/bip_actionsets/
|
||||
|
||||
|
|
|
@ -552,11 +552,50 @@
|
|||
</spark>
|
||||
|
||||
<!-- Do this after finishing okay -->
|
||||
<ok to="end" />
|
||||
<ok to="deleteOutputPathForActionSet" />
|
||||
<!-- Go there if we have an error -->
|
||||
<error to="map-scores-fail" />
|
||||
|
||||
</action>
|
||||
</action>
|
||||
|
||||
<action name="deleteOutputPathForActionSet">
|
||||
<fs>
|
||||
<delete path="${actionSetOutputPath}"/>
|
||||
<mkdir path="${actionSetOutputPath}"/>
|
||||
<!--
|
||||
<delete path="${workingDir}"/>
|
||||
<mkdir path="${workingDir}"/>
|
||||
-->
|
||||
</fs>
|
||||
<ok to="createActionSet"/>
|
||||
<error to="actionset-delete-fail"/>
|
||||
</action>
|
||||
|
||||
|
||||
<action name="createActionSet">
|
||||
<spark xmlns="uri:oozie:spark-action:0.2">
|
||||
<master>yarn</master>
|
||||
<mode>cluster</mode>
|
||||
<name>Produces the atomic action with the bip finder scores for publications</name>
|
||||
<class>eu.dnetlib.dhp.actionmanager.bipfinder.SparkAtomicActionScoreJob</class>
|
||||
<jar>dhp-aggregation-${projectVersion}.jar</jar>
|
||||
<spark-opts>
|
||||
--executor-memory=${sparkExecutorMemory}
|
||||
--executor-cores=${sparkExecutorCores}
|
||||
--driver-memory=${sparkDriverMemory}
|
||||
--conf spark.extraListeners=${spark2ExtraListeners}
|
||||
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
|
||||
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
|
||||
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
|
||||
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
||||
</spark-opts>
|
||||
<arg>--inputPath</arg><arg>${bipScorePath}</arg>
|
||||
<arg>--outputPath</arg><arg>${actionSetOutputPath}</arg>
|
||||
</spark>
|
||||
<ok to="end"/>
|
||||
<error to="actionset-creation-fail"/>
|
||||
</action>
|
||||
|
||||
|
||||
|
||||
<!-- TODO: end the workflow-->
|
||||
|
@ -597,4 +636,12 @@
|
|||
<message>Mapping scores to DOIs failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
|
||||
</kill>
|
||||
|
||||
<kill name="actionset-delete-fail">
|
||||
<message>Deleting output path for actionsets failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
|
||||
</kill>
|
||||
|
||||
<kill name="actionset-creation-fail">
|
||||
<message>ActionSet creation failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
|
||||
</kill>
|
||||
|
||||
</workflow-app>
|
||||
|
|
Loading…
Reference in New Issue