forked from D-Net/dnet-hadoop
incremented number of repartition
This commit is contained in:
parent
071f5c3e52
commit
f09e065865
|
@ -37,7 +37,7 @@ public class SparkGenerateScholix {
|
|||
final JavaSparkContext sc = new JavaSparkContext(spark.sparkContext());
|
||||
|
||||
|
||||
final JavaRDD<String> relationToExport = sc.textFile(graphPath + "/relation").filter(ProvisionUtil::isNotDeleted);
|
||||
final JavaRDD<String> relationToExport = sc.textFile(graphPath + "/relation").filter(ProvisionUtil::isNotDeleted).repartition(4000);
|
||||
final JavaPairRDD<String,String> scholixSummary = sc.textFile(workingDirPath + "/summary").mapToPair((PairFunction<String, String, String>) i -> new Tuple2<>(DHPUtils.getJPathString(jsonIDPath, i), i));
|
||||
scholixSummary.join(
|
||||
relationToExport
|
||||
|
|
Loading…
Reference in New Issue