forked from D-Net/dnet-hadoop
changed because of D-Net/dnet-hadoop#61 (comment)
This commit is contained in:
parent
1df94b85b4
commit
f5e5e92a10
|
@ -121,7 +121,7 @@
|
||||||
<delete path="${outputPath}"/>
|
<delete path="${outputPath}"/>
|
||||||
<mkdir path="${outputPath}"/>
|
<mkdir path="${outputPath}"/>
|
||||||
</fs>
|
</fs>
|
||||||
<ok to="make_archive"/>
|
<ok to="save_community_map"/>
|
||||||
<error to="Kill"/>
|
<error to="Kill"/>
|
||||||
</action>
|
</action>
|
||||||
|
|
||||||
|
@ -274,7 +274,6 @@
|
||||||
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
||||||
</spark-opts>
|
</spark-opts>
|
||||||
<arg>--sourcePath</arg><arg>${workingDir}/result/publication</arg>
|
<arg>--sourcePath</arg><arg>${workingDir}/result/publication</arg>
|
||||||
<!-- <arg>--sourcePath</arg><arg>${sourcePath}/publication</arg>-->
|
|
||||||
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Publication</arg>
|
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Publication</arg>
|
||||||
<arg>--outputPath</arg><arg>${workingDir}/dump/publication</arg>
|
<arg>--outputPath</arg><arg>${workingDir}/dump/publication</arg>
|
||||||
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
||||||
|
@ -302,7 +301,6 @@
|
||||||
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
||||||
</spark-opts>
|
</spark-opts>
|
||||||
<arg>--sourcePath</arg><arg>${workingDir}/result/dataset</arg>
|
<arg>--sourcePath</arg><arg>${workingDir}/result/dataset</arg>
|
||||||
<!-- <arg>--sourcePath</arg><arg>${sourcePath}/dataset</arg>-->
|
|
||||||
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Dataset</arg>
|
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Dataset</arg>
|
||||||
<arg>--outputPath</arg><arg>${workingDir}/dump/dataset</arg>
|
<arg>--outputPath</arg><arg>${workingDir}/dump/dataset</arg>
|
||||||
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
||||||
|
@ -330,7 +328,6 @@
|
||||||
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
||||||
</spark-opts>
|
</spark-opts>
|
||||||
<arg>--sourcePath</arg><arg>${workingDir}/result/otherresearchproduct</arg>
|
<arg>--sourcePath</arg><arg>${workingDir}/result/otherresearchproduct</arg>
|
||||||
<!-- <arg>--sourcePath</arg><arg>${sourcePath}/otherresearchproduct</arg>-->
|
|
||||||
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.OtherResearchProduct</arg>
|
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.OtherResearchProduct</arg>
|
||||||
<arg>--outputPath</arg><arg>${workingDir}/dump/otherresearchproduct</arg>
|
<arg>--outputPath</arg><arg>${workingDir}/dump/otherresearchproduct</arg>
|
||||||
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
||||||
|
@ -358,7 +355,6 @@
|
||||||
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
--conf spark.sql.warehouse.dir=${sparkSqlWarehouseDir}
|
||||||
</spark-opts>
|
</spark-opts>
|
||||||
<arg>--sourcePath</arg><arg>${workingDir}/result/software</arg>
|
<arg>--sourcePath</arg><arg>${workingDir}/result/software</arg>
|
||||||
<!-- <arg>--sourcePath</arg><arg>${sourcePath}/software</arg>-->
|
|
||||||
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Software</arg>
|
<arg>--resultTableName</arg><arg>eu.dnetlib.dhp.schema.oaf.Software</arg>
|
||||||
<arg>--outputPath</arg><arg>${workingDir}/dump/software</arg>
|
<arg>--outputPath</arg><arg>${workingDir}/dump/software</arg>
|
||||||
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
<arg>--communityMapPath</arg><arg>${workingDir}/communityMap</arg>
|
||||||
|
@ -532,8 +528,7 @@
|
||||||
<main-class>eu.dnetlib.dhp.oa.graph.dump.MakeTar</main-class>
|
<main-class>eu.dnetlib.dhp.oa.graph.dump.MakeTar</main-class>
|
||||||
<arg>--hdfsPath</arg><arg>${outputPath}</arg>
|
<arg>--hdfsPath</arg><arg>${outputPath}</arg>
|
||||||
<arg>--nameNode</arg><arg>${nameNode}</arg>
|
<arg>--nameNode</arg><arg>${nameNode}</arg>
|
||||||
<!-- <arg>--sourcePath</arg><arg>${workingDir}/resultperfunder</arg>-->
|
<arg>--sourcePath</arg><arg>${workingDir}/resultperfunder</arg>
|
||||||
<arg>--sourcePath</arg><arg>/user/miriam.baglioni/graph_dump_whole_production_funder_results</arg>
|
|
||||||
</java>
|
</java>
|
||||||
<ok to="should_upload"/>
|
<ok to="should_upload"/>
|
||||||
<error to="Kill"/>
|
<error to="Kill"/>
|
||||||
|
|
Loading…
Reference in New Issue