forked from D-Net/dnet-hadoop
Fix scores & Workflow
This commit is contained in:
commit
0c433eccdd
|
@ -102,4 +102,4 @@ actionSetOutputPath=${workingDir}/bip_actionsets/
|
|||
# The directory to store project impact indicators
|
||||
projectImpactIndicatorsOutput=${workingDir}/project_indicators
|
||||
|
||||
resume=create-openaire-ranking-graph
|
||||
resume=entry-point-decision
|
||||
|
|
|
@ -340,12 +340,18 @@
|
|||
|
||||
</action>
|
||||
|
||||
<<<<<<< HEAD
|
||||
<!-- Removed for ser -->
|
||||
<!--
|
||||
JOIN ITERATIVE METHODS AND THEN END =
|
||||
<join name="join-iterative-rankings" to="end" />
|
||||
to="get-file-names"/>
|
||||
-->
|
||||
=======
|
||||
<!-- JOIN ITERATIVE METHODS AND THEN END -->
|
||||
<join name="join-iterative-rankings" to="get-file-names" />
|
||||
<!-- to="get-file-names"/> -->
|
||||
>>>>>>> 60f25b780de1c456762003cbb8b0011c9c82f93d
|
||||
|
||||
|
||||
<!-- This will be a shell action that will output key-value pairs for output files -->
|
||||
|
@ -358,7 +364,7 @@
|
|||
<!-- name of script to run -->
|
||||
<argument>get_ranking_files.sh</argument>
|
||||
<!-- We only pass the directory where we expect to find the rankings -->
|
||||
<argument>/${workingDir}</argument>
|
||||
<argument>${workingDir}</argument>
|
||||
|
||||
<!-- the name of the file run -->
|
||||
<file>${wfAppPath}/get_ranking_files.sh#get_ranking_files.sh</file>
|
||||
|
@ -512,7 +518,7 @@
|
|||
</spark-opts>
|
||||
|
||||
<!-- Script arguments here -->
|
||||
<arg>${openaireDataInput}</arg>
|
||||
<arg>${openaireDataInput}/</arg>
|
||||
<!-- number of partitions to be used on joins -->
|
||||
<arg>${synonymFolder}</arg>
|
||||
<!-- This needs to point to the file on the hdfs i think -->
|
||||
|
@ -570,6 +576,7 @@
|
|||
|
||||
<!-- Do this after finishing okay -->
|
||||
<ok to="deleteOutputPathForActionSet" />
|
||||
<!-- This is the initial code <ok to="deleteOutputPathForActionSet" /> -->
|
||||
<!-- Go there if we have an error -->
|
||||
<error to="map-scores-fail" />
|
||||
|
||||
|
|
Loading…
Reference in New Issue