package eu.dnetlib.dhp.graph; import com.mongodb.*; import com.mongodb.client.FindIterable; import com.mongodb.client.MongoCollection; import com.mongodb.client.MongoDatabase; import eu.dnetlib.dhp.application.ArgumentApplicationParser; import eu.dnetlib.message.Message; import eu.dnetlib.message.MessageType; import org.apache.commons.io.IOUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.SequenceFile; import org.apache.hadoop.io.Text; import org.bson.Document; import org.bson.conversions.Bson; import java.io.IOException; import java.net.URI; import java.util.*; import java.util.concurrent.atomic.AtomicInteger; import java.util.function.Consumer; import java.util.stream.Collectors; public class ImportDataFromMongo { public static void main(String[] args) throws Exception { final ArgumentApplicationParser parser = new ArgumentApplicationParser(IOUtils.toString(SparkGraphImporterJob.class.getResourceAsStream("/eu/dnetlib/dhp/graph/import_from_mongo_parameters.json"))); parser.parseArgument(args); final int port = Integer.parseInt(parser.get("dbport")); final String host = parser.get("dbhost"); final String format = parser.get("format"); final String layout = parser.get("layout"); final String interpretation = parser.get("interpretation"); final String dbName = parser.get("dbName"); final MongoClient client = new MongoClient(host, port); MongoDatabase database = client.getDatabase(dbName); MongoCollection metadata = database.getCollection("metadata"); MongoCollection metadataManager = database.getCollection("metadataManager"); final DBObject query = QueryBuilder.start("format").is(format).and("layout").is(layout).and("interpretation").is(interpretation).get(); final List ids = new ArrayList<>(); metadata.find((Bson) query).forEach((Consumer) document -> ids.add(document.getString("mdId"))); List databaseId = ids.stream().map(it -> getCurrentId(it, metadataManager)).filter(Objects::nonNull).collect(Collectors.toList()); final String hdfsuri = parser.get("namenode"); // ====== Init HDFS File System Object Configuration conf = new Configuration(); // Set FileSystem URI conf.set("fs.defaultFS", hdfsuri); // Because of Maven conf.set("fs.hdfs.impl", org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()); conf.set("fs.file.impl", org.apache.hadoop.fs.LocalFileSystem.class.getName()); System.setProperty("HADOOP_USER_NAME", parser.get("user")); System.setProperty("hadoop.home.dir", "/"); FileSystem.get(URI.create(hdfsuri), conf); Path hdfswritepath = new Path(parser.get("targetPath")); final AtomicInteger counter = new AtomicInteger(0); try (SequenceFile.Writer writer = SequenceFile.createWriter(conf, SequenceFile.Writer.file(hdfswritepath), SequenceFile.Writer.keyClass(IntWritable.class), SequenceFile.Writer.valueClass(Text.class))) { final IntWritable key = new IntWritable(counter.get()); final Text value = new Text(); databaseId.forEach(id -> { System.out.println("Reading :"+id); MongoCollection collection = database.getCollection(id); collection.find().forEach((Consumer) document -> { key.set(counter.getAndIncrement()); value.set(document.getString("body")); if (counter.get() % 10000 == 0) { System.out.println("Added "+counter.get()); } try { writer.append(key, value); } catch (IOException e) { throw new RuntimeException(e); } } ); }); } } private static String getCurrentId(final String mdId, final MongoCollection metadataManager) { FindIterable result = metadataManager.find((Bson) QueryBuilder.start("mdId").is(mdId).get()); final Document item = result.first(); return item == null ? null : item.getString("currentId"); } }