diff --git a/dhp-workflows/dhp-graph-mapper/src/main/resources/eu/dnetlib/dhp/oa/graph/raw_all/oozie_app/workflow.xml b/dhp-workflows/dhp-graph-mapper/src/main/resources/eu/dnetlib/dhp/oa/graph/raw_all/oozie_app/workflow.xml
index 80f33bd53..0821f04ea 100644
--- a/dhp-workflows/dhp-graph-mapper/src/main/resources/eu/dnetlib/dhp/oa/graph/raw_all/oozie_app/workflow.xml
+++ b/dhp-workflows/dhp-graph-mapper/src/main/resources/eu/dnetlib/dhp/oa/graph/raw_all/oozie_app/workflow.xml
@@ -40,6 +40,16 @@
false
should import content from the aggregator or reuse a previous version
+
+ reuseODF_hdfs
+ false
+ should import content from the aggregator or reuse a previous version
+
+
+ reuseOAF_hdfs
+ false
+ should import content from the aggregator or reuse a previous version
+
contentPath
path location to store (or reuse) content from the aggregator
@@ -289,7 +299,7 @@
${wf:conf('reuseOAF') eq false}
- ${wf:conf('reuseOAF') eq true}
+ ${wf:conf('reuseOAF') eq true}
@@ -324,10 +334,78 @@
--mdLayoutstore
--mdInterpretationintersection
-
+
+
+
+
+ ${wf:conf('reuseODF_hdfs') eq false}
+ ${wf:conf('reuseODF_hdfs') eq true}
+
+
+
+
+
+ yarn
+ cluster
+ ImportODF_hdfs
+ eu.dnetlib.dhp.oa.graph.raw.MigrateHdfsMdstoresApplication
+ dhp-graph-mapper-${projectVersion}.jar
+
+ --executor-memory ${sparkExecutorMemory}
+ --executor-cores ${sparkExecutorCores}
+ --driver-memory=${sparkDriverMemory}
+ --conf spark.extraListeners=${spark2ExtraListeners}
+ --conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
+ --conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
+ --conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
+
+ --hdfsPath${contentPath}/odf_records_hdfs
+ --mdstoreManagerUrl${mdstoreManagerUrl}
+ --mdFormatODF
+ --mdLayoutstore
+ --mdInterpretationcleaned
+
+
+
+
+
+
+
+ ${wf:conf('reuseOAF_hdfs') eq false}
+ ${wf:conf('reuseOAF_hdfs') eq true}
+
+
+
+
+
+
+ yarn
+ cluster
+ ImportOAF_hdfs
+ eu.dnetlib.dhp.oa.graph.raw.MigrateHdfsMdstoresApplication
+ dhp-graph-mapper-${projectVersion}.jar
+
+ --executor-memory ${sparkExecutorMemory}
+ --executor-cores ${sparkExecutorCores}
+ --driver-memory=${sparkDriverMemory}
+ --conf spark.extraListeners=${spark2ExtraListeners}
+ --conf spark.sql.queryExecutionListeners=${spark2SqlQueryExecutionListeners}
+ --conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
+ --conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
+
+ --hdfsPath${contentPath}/oaf_records_hdfs
+ --mdstoreManagerUrl${mdstoreManagerUrl}
+ --mdFormatOAF
+ --mdLayoutstore
+ --mdInterpretationcleaned
+
+
+
+
+
${wf:conf('reuseDBOpenorgs') eq false}
@@ -426,7 +504,7 @@
--conf spark.yarn.historyServer.address=${spark2YarnHistoryServerAddress}
--conf spark.eventLog.dir=${nameNode}${spark2EventLogDir}
- --sourcePaths${contentPath}/db_openaire,${contentPath}/db_openorgs,${contentPath}/oaf_records,${contentPath}/odf_records
+ --sourcePaths${contentPath}/db_openaire,${contentPath}/db_openorgs,${contentPath}/oaf_records,${contentPath}/odf_records,${contentPath}/oaf_records_hdfs,${contentPath}/odf_records_hdfs
--targetPath${workingDir}/entities
--isLookupUrl${isLookupUrl}
--shouldHashId${shouldHashId}