2021-01-28 09:51:17 +01:00
|
|
|
|
|
|
|
package eu.dnetlib.dhp.collection.worker;
|
|
|
|
|
2021-02-02 12:28:21 +01:00
|
|
|
import static eu.dnetlib.dhp.aggregation.common.AggregationConstants.*;
|
2021-02-03 12:33:41 +01:00
|
|
|
import static eu.dnetlib.dhp.aggregation.common.AggregationUtility.*;
|
|
|
|
import static eu.dnetlib.dhp.application.ApplicationUtils.*;
|
2021-02-02 12:28:21 +01:00
|
|
|
|
2021-01-28 09:51:17 +01:00
|
|
|
import org.apache.commons.io.IOUtils;
|
|
|
|
import org.slf4j.Logger;
|
|
|
|
import org.slf4j.LoggerFactory;
|
|
|
|
|
|
|
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
|
|
|
|
2021-01-29 16:42:41 +01:00
|
|
|
import eu.dnetlib.data.mdstore.manager.common.model.MDStoreVersion;
|
2021-02-03 12:33:41 +01:00
|
|
|
import eu.dnetlib.dhp.aggregation.common.AggregationUtility;
|
2021-01-28 09:51:17 +01:00
|
|
|
import eu.dnetlib.dhp.application.ArgumentApplicationParser;
|
2021-02-03 12:33:41 +01:00
|
|
|
import eu.dnetlib.dhp.collection.worker.utils.CollectorPluginErrorLogList;
|
2021-01-28 09:51:17 +01:00
|
|
|
import eu.dnetlib.dhp.collection.worker.utils.CollectorPluginFactory;
|
2021-01-29 16:42:41 +01:00
|
|
|
import eu.dnetlib.dhp.collector.worker.model.ApiDescriptor;
|
2021-01-28 09:51:17 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* DnetCollectortWorkerApplication is the main class responsible to start the Dnet Collection into HDFS. This module
|
|
|
|
* will be executed on the hadoop cluster and taking in input some parameters that tells it which is the right collector
|
|
|
|
* plugin to use and where store the data into HDFS path
|
|
|
|
*
|
|
|
|
* @author Sandro La Bruzzo
|
|
|
|
*/
|
|
|
|
public class CollectorWorkerApplication {
|
|
|
|
|
|
|
|
private static final Logger log = LoggerFactory.getLogger(CollectorWorkerApplication.class);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* @param args
|
|
|
|
*/
|
|
|
|
public static void main(final String[] args) throws Exception {
|
|
|
|
|
|
|
|
final ArgumentApplicationParser argumentParser = new ArgumentApplicationParser(
|
|
|
|
IOUtils
|
|
|
|
.toString(
|
|
|
|
CollectorWorker.class
|
|
|
|
.getResourceAsStream(
|
|
|
|
"/eu/dnetlib/dhp/collection/collector_parameter.json")));
|
|
|
|
argumentParser.parseArgument(args);
|
|
|
|
|
|
|
|
final String hdfsuri = argumentParser.get("namenode");
|
|
|
|
log.info("hdfsURI is {}", hdfsuri);
|
2021-01-29 16:42:41 +01:00
|
|
|
|
2021-01-28 09:51:17 +01:00
|
|
|
final String apiDescriptor = argumentParser.get("apidescriptor");
|
2021-01-29 16:42:41 +01:00
|
|
|
log.info("apiDescriptor is {}", apiDescriptor);
|
|
|
|
|
|
|
|
final String mdStoreVersion = argumentParser.get("mdStoreVersion");
|
|
|
|
log.info("mdStoreVersion is {}", mdStoreVersion);
|
2021-01-28 09:51:17 +01:00
|
|
|
|
2021-02-03 12:33:41 +01:00
|
|
|
final MDStoreVersion currentVersion = MAPPER.readValue(mdStoreVersion, MDStoreVersion.class);
|
|
|
|
final String hdfsPath = currentVersion.getHdfsPath() + SEQUENCE_FILE_NAME;
|
|
|
|
log.info("hdfs path is {}", hdfsPath);
|
|
|
|
|
|
|
|
final ApiDescriptor api = MAPPER.readValue(apiDescriptor, ApiDescriptor.class);
|
2021-01-28 09:51:17 +01:00
|
|
|
|
2021-02-03 12:33:41 +01:00
|
|
|
final CollectorWorker worker = new CollectorWorker(api, hdfsuri, hdfsPath);
|
|
|
|
CollectorPluginErrorLogList errors = worker.collect();
|
2021-01-28 09:51:17 +01:00
|
|
|
|
2021-02-03 12:33:41 +01:00
|
|
|
populateOOZIEEnv("collectorErrors", errors.toString());
|
2021-01-29 16:42:41 +01:00
|
|
|
|
2021-01-28 09:51:17 +01:00
|
|
|
}
|
2021-01-29 16:42:41 +01:00
|
|
|
|
2021-01-28 09:51:17 +01:00
|
|
|
}
|