sourcePath
the source path
whitelist
the white list
allowedtypes
the allowed types
outputPath
the output path
${jobTracker}
${nameNode}
oozie.action.sharelib.for.spark
${oozieActionShareLibForSpark2}
Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]
${nameNode}/${sourcePath}/relation
${nameNode}/${outputPath}/relation
${nameNode}/${sourcePath}/organization
${nameNode}/${outputPath}/organization
${nameNode}/${sourcePath}/project
${nameNode}/${outputPath}/project
${nameNode}/${sourcePath}/datasource
${nameNode}/${outputPath}/datasource
yarn
cluster
PrepareDatasourceCountryAssociation
eu.dnetlib.dhp.countrypropagation.PrepareDatasourceCountryAssociation
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}
--whitelist${whitelist}
--allowedtypes${allowedtypes}
--outputPath${workingDir}/preparedInfo
yarn
cluster
prepareResultCountry-Publication
eu.dnetlib.dhp.countrypropagation.PrepareResultCountrySet
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/publication
--outputPath${workingDir}/publication
--workingPath${workingDir}/workingP
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--preparedInfoPath${workingDir}/preparedInfo
yarn
cluster
prepareResultCountry-Dataset
eu.dnetlib.dhp.countrypropagation.PrepareResultCountrySet
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/dataset
--outputPath${workingDir}/dataset
--workingPath${workingDir}/workingD
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--preparedInfoPath${workingDir}/preparedInfo
yarn
cluster
prepareResultCountry-ORP
eu.dnetlib.dhp.countrypropagation.PrepareResultCountrySet
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/otherresearchproduct
--outputPath${workingDir}/otherresearchproduct
--workingPath${workingDir}/workingO
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--preparedInfoPath${workingDir}/preparedInfo
yarn
cluster
prepareResultCountry-Software
eu.dnetlib.dhp.countrypropagation.PrepareResultCountrySet
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/software
--outputPath${workingDir}/software
--workingPath${workingDir}/workingS
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--preparedInfoPath${workingDir}/preparedInfo
yarn
cluster
countryPropagationForPublications
eu.dnetlib.dhp.countrypropagation.SparkCountryPropagationJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/publication
--preparedInfoPath${workingDir}/publication
--resultTableNameeu.dnetlib.dhp.schema.oaf.Publication
--outputPath${outputPath}/publication
yarn
cluster
countryPropagationForDataset
eu.dnetlib.dhp.countrypropagation.SparkCountryPropagationJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/dataset
--preparedInfoPath${workingDir}/dataset
--resultTableNameeu.dnetlib.dhp.schema.oaf.Dataset
--outputPath${outputPath}/dataset
yarn
cluster
countryPropagationForORP
eu.dnetlib.dhp.countrypropagation.SparkCountryPropagationJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/otherresearchproduct
--preparedInfoPath${workingDir}/otherresearchproduct
--resultTableNameeu.dnetlib.dhp.schema.oaf.OtherResearchProduct
--outputPath${outputPath}/otherresearchproduct
yarn
cluster
countryPropagationForSoftware
eu.dnetlib.dhp.countrypropagation.SparkCountryPropagationJob
dhp-enrichment-${projectVersion}.jar
--executor-cores=${sparkExecutorCores}
--executor-memory=${sparkExecutorMemory}
--driver-memory=${sparkDriverMemory}
--conf spark.extraListeners=${spark2ExtraListeners}
--conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
--conf spark.speculation=false
--conf spark.hadoop.mapreduce.map.speculative=false
--conf spark.hadoop.mapreduce.reduce.speculative=false
--conf spark.sql.shuffle.partitions=3840
--sourcePath${sourcePath}/software
--preparedInfoPath${workingDir}/software
--resultTableNameeu.dnetlib.dhp.schema.oaf.Software
--outputPath${outputPath}/software