sourcePath the source path baseURL the base URL to use to select the right community APIs startFrom undelete ${jobTracker} ${nameNode} oozie.action.sharelib.for.spark ${oozieActionShareLibForSpark2} ${wf:conf('startFrom') eq 'undelete'} Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}] yarn cluster Prepare-Community-Result-Organization eu.dnetlib.dhp.resulttocommunityfromproject.PrepareResultCommunitySet dhp-enrichment-${projectVersion}.jar --executor-cores=6 --executor-memory=5G --conf spark.executor.memoryOverhead=3g --conf spark.sql.shuffle.partitions=3284 --driver-memory=${sparkDriverMemory} --conf spark.extraListeners=${spark2ExtraListeners} --conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners} --conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress} --conf spark.eventLog.dir=${nameNode}${spark2EventLogDir} --conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors} --sourcePath${sourcePath}/relation --outputPath${workingDir}/communitythroughproject/preparedInfo/resultCommunityList --baseURL${baseURL} yarn cluster community2resultfromproject eu.dnetlib.dhp.resulttocommunityfromproject.SparkResultToCommunityFromProject dhp-enrichment-${projectVersion}.jar --executor-cores=6 --executor-memory=5G --conf spark.executor.memoryOverhead=3g --conf spark.sql.shuffle.partitions=3284 --driver-memory=${sparkDriverMemory} --conf spark.extraListeners=${spark2ExtraListeners} --conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners} --conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress} --conf spark.eventLog.dir=${nameNode}${spark2EventLogDir} --conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors} --preparedInfoPath${workingDir}/communitythroughproject/preparedInfo/resultCommunityList --sourcePath${sourcePath}/ --outputPath${workingDir}/communitythroughproject/ yarn cluster move results eu.dnetlib.dhp.MoveResult dhp-enrichment-${projectVersion}.jar --executor-cores=6 --executor-memory=5G --conf spark.executor.memoryOverhead=3g --conf spark.sql.shuffle.partitions=3284 --driver-memory=${sparkDriverMemory} --conf spark.extraListeners=${spark2ExtraListeners} --conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners} --conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress} --conf spark.eventLog.dir=${nameNode}${spark2EventLogDir} --conf spark.dynamicAllocation.maxExecutors=${spark2MaxExecutors} --sourcePath${workingDir}/communitythroughproject/ --outputPath${sourcePath}/