forked from D-Net/dnet-hadoop
[OpenOrgsWf] cleanup
This commit is contained in:
parent
7f4e9479ec
commit
9237d55d7f
|
@ -16,10 +16,6 @@
|
|||
<name>workingPath</name>
|
||||
<description>path for the working directory</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>dedupGraphPath</name>
|
||||
<description>path for the output graph</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>cutConnectedComponent</name>
|
||||
<description>max number of elements in a connected component</description>
|
||||
|
|
|
@ -214,57 +214,6 @@
|
|||
<error to="Kill"/>
|
||||
</action>
|
||||
|
||||
<!--<action name="merge_claims_relation">-->
|
||||
<!--<spark xmlns="uri:oozie:spark-action:0.2">-->
|
||||
<!--<master>yarn</master>-->
|
||||
<!--<mode>cluster</mode>-->
|
||||
<!--<name>MergeClaims_relation</name>-->
|
||||
<!--<class>eu.dnetlib.dhp.oa.graph.raw.MergeClaimsApplication</class>-->
|
||||
<!--<jar>dhp-graph-mapper-${projectVersion}.jar</jar>-->
|
||||
<!--<spark-opts>-->
|
||||
<!----executor-memory ${sparkExecutorMemory}-->
|
||||
<!----executor-cores ${sparkExecutorCores}-->
|
||||
<!----driver-memory=${sparkDriverMemory}-->
|
||||
<!----conf spark.extraListeners=${spark2ExtraListeners}-->
|
||||
<!----conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}-->
|
||||
<!----conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}-->
|
||||
<!----conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}-->
|
||||
<!----conf spark.sql.shuffle.partitions=3840-->
|
||||
<!--</spark-opts>-->
|
||||
<!--<arg>--rawGraphPath</arg><arg>${workingDir}/graph_raw</arg>-->
|
||||
<!--<arg>--claimsGraphPath</arg><arg>${workingDir}/graph_claims</arg>-->
|
||||
<!--<arg>--outputRawGaphPath</arg><arg>${graphOutputPath}</arg>-->
|
||||
<!--<arg>--graphTableClassName</arg><arg>eu.dnetlib.dhp.schema.oaf.Relation</arg>-->
|
||||
<!--</spark>-->
|
||||
<!--<ok to="wait_merge"/>-->
|
||||
<!--<error to="Kill"/>-->
|
||||
<!--</action>-->
|
||||
|
||||
<!--<action name="merge_claims_organization">-->
|
||||
<!--<spark xmlns="uri:oozie:spark-action:0.2">-->
|
||||
<!--<master>yarn</master>-->
|
||||
<!--<mode>cluster</mode>-->
|
||||
<!--<name>MergeClaims_organization</name>-->
|
||||
<!--<class>eu.dnetlib.dhp.oa.graph.raw.MergeClaimsApplication</class>-->
|
||||
<!--<jar>dhp-graph-mapper-${projectVersion}.jar</jar>-->
|
||||
<!--<spark-opts>-->
|
||||
<!----executor-memory ${sparkExecutorMemory}-->
|
||||
<!----executor-cores ${sparkExecutorCores}-->
|
||||
<!----driver-memory=${sparkDriverMemory}-->
|
||||
<!----conf spark.extraListeners=${spark2ExtraListeners}-->
|
||||
<!----conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}-->
|
||||
<!----conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}-->
|
||||
<!----conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}-->
|
||||
<!----conf spark.sql.shuffle.partitions=200-->
|
||||
<!--</spark-opts>-->
|
||||
<!--<arg>--rawGraphPath</arg><arg>${workingDir}/graph_raw</arg>-->
|
||||
<!--<arg>--claimsGraphPath</arg><arg>${workingDir}/graph_claims</arg>-->
|
||||
<!--<arg>--outputRawGaphPath</arg><arg>${graphOutputPath}</arg>-->
|
||||
<!--<arg>--graphTableClassName</arg><arg>eu.dnetlib.dhp.schema.oaf.Organization</arg>-->
|
||||
<!--</spark>-->
|
||||
<!--<ok to="wait_merge"/>-->
|
||||
<!--<error to="Kill"/>-->
|
||||
<!--</action>-->
|
||||
|
||||
<end name="End"/>
|
||||
|
||||
</workflow-app>
|
Loading…
Reference in New Issue