sourcePath
the source path
allowedsemrels
the semantic relationships allowed for propagation
isLookUpUrl
the isLookup service endpoint
outputPath
the output path
Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]
${jobTracker}
${nameNode}
${nameNode}/${sourcePath}/relation
${nameNode}/${outputPath}/relation
${jobTracker}
${nameNode}
${nameNode}/${sourcePath}/organization
${nameNode}/${outputPath}/organization
${jobTracker}
${nameNode}
${nameNode}/${sourcePath}/project
${nameNode}/${outputPath}/project
${jobTracker}
${nameNode}
${nameNode}/${sourcePath}/datasource
${nameNode}/${outputPath}/datasource
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Publications
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--outputPath${workingDir}/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--isLookUpUrl${isLookUpUrl}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Dataset
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--outputPath${workingDir}/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--isLookUpUrl${isLookUpUrl}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-ORP
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--outputPath${workingDir}/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--isLookUpUrl${isLookUpUrl}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Software
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--outputPath${workingDir}/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--isLookUpUrl${isLookUpUrl}
yarn
cluster
ResultToCommunityEmRelPropagation-PreparePhase2
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep2
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${workingDir}/preparedInfo/targetCommunityAssoc
--outputPath${workingDir}/preparedInfo/mergedCommunityAssoc
yarn
cluster
Result2CommunitySemRelPropagation-Publication
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--conf spark.sql.shuffle.partitions=10000
--preparedInfoPath${workingDir}/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/publication
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--outputPath${outputPath}/publication
--saveGraph${saveGraph}
yarn
cluster
Result2CommunitySemRelPropagation-Dataset
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--conf spark.sql.shuffle.partitions=5000
--preparedInfoPath${workingDir}/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/dataset
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--outputPath${outputPath}/dataset
--saveGraph${saveGraph}
yarn
cluster
Result2CommunitySemRelPropagation-ORP
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--conf spark.sql.shuffle.partitions=2000
--preparedInfoPath${workingDir}/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/otherresearchproduct
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--outputPath${outputPath}/otherresearchproduct
--saveGraph${saveGraph}
yarn
cluster
Result2CommunitySemRelPropagation-Software
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--conf spark.sql.shuffle.partitions=1000
--preparedInfoPath${workingDir}/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/software
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--outputPath${outputPath}/software
--saveGraph${saveGraph}