sourcePath
the source path
allowedsemrels
the semantic relationships allowed for propagation
baseURL
the isLookup service endpoint
outputPath
the output path
startFrom
undelete
${wf:conf('startFrom') eq 'undelete'}
Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Publications
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--outputPath${workingDir}/communitysemrel/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--baseURL${baseURL}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Dataset
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--outputPath${workingDir}/communitysemrel/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--baseURL${baseURL}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-ORP
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--outputPath${workingDir}/communitysemrel/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--baseURL${baseURL}
yarn
cluster
ResultToCommunitySemRel-PreparePhase1-Software
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep1
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${sourcePath}
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--outputPath${workingDir}/communitysemrel/preparedInfo/targetCommunityAssoc
--allowedsemrels${allowedsemrels}
--baseURL${baseURL}
yarn
cluster
ResultToCommunityEmRelPropagation-PreparePhase2
eu.dnetlib.dhp.resulttocommunityfromsemrel.PrepareResultCommunitySetStep2
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${workingDir}/communitysemrel/preparedInfo/targetCommunityAssoc
--outputPath${workingDir}/communitysemrel/preparedInfo/mergedCommunityAssoc
yarn
cluster
Result2CommunitySemRelPropagation-Publication
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=4G
--conf spark.executor.memoryOverhead=5G
--conf spark.sql.shuffle.partitions=15000
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--preparedInfoPath${workingDir}/communitysemrel/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/publication
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--outputPath${workingDir}/communitysemrel/publication
yarn
cluster
Result2CommunitySemRelPropagation-Dataset
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--preparedInfoPath${workingDir}/communitysemrel/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/dataset
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--outputPath${workingDir}/communitysemrel/dataset
yarn
cluster
Result2CommunitySemRelPropagation-ORP
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--preparedInfoPath${workingDir}/communitysemrel/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/otherresearchproduct
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--outputPath${workingDir}/communitysemrel/otherresearchproduct
yarn
cluster
Result2CommunitySemRelPropagation-Software
eu.dnetlib.dhp.resulttocommunityfromsemrel.SparkResultToCommunityThroughSemRelJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--preparedInfoPath${workingDir}/communitysemrel/preparedInfo/mergedCommunityAssoc
--sourcePath${sourcePath}/software
--hive_metastore_uris${hive_metastore_uris}
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--outputPath${workingDir}/communitysemrel/software
yarn
cluster
move results
eu.dnetlib.dhp.MoveResult
dhp-enrichment-${projectVersion}.jar
--executor-cores=6
--executor-memory=5G
--conf spark.executor.memoryOverhead=3g
--conf spark.sql.shuffle.partitions=3284
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors}
--sourcePath${workingDir}/communitysemrel/
--outputPath${sourcePath}/