dnet-hadoop/dhp-workflows/dhp-aggregation/README.md

29 lines
1.2 KiB
Markdown
Raw Normal View History

2019-04-03 16:05:16 +02:00
Description of the Module
--------------------------
This module defines a **collector worker application** that runs on Hadoop.
2019-04-03 16:05:16 +02:00
2019-10-02 15:26:58 +02:00
It is responsible for harvesting metadata using different plugins.
2019-04-03 16:05:16 +02:00
The collector worker uses a message queue to inform the progress
of the harvesting action (using a message queue for sending **ONGOING** messages) furthermore,
It gives, at the end of the job, some information about the status
of the collection i.e Number of records collected(using a message queue for sending **REPORT** messages).
To work the collection worker need some parameter like:
2019-04-03 16:05:16 +02:00
* **hdfsPath**: the path where storing the sequential file
* **apidescriptor**: the JSON encoding of the API Descriptor
* **namenode**: the Name Node URI
* **userHDFS**: the user wich create the hdfs seq file
* **rabbitUser**: the user to connect with RabbitMq for messaging
* **rabbitPassWord**: the password to connect with RabbitMq for messaging
* **rabbitHost**: the host of the RabbitMq server
* **rabbitOngoingQueue**: the name of the ongoing queue
* **rabbitReportQueue**: the name of the report queue
* **workflowId**: the identifier of the dnet Workflow
2019-04-03 16:05:16 +02:00
##Plugins
* OAI Plugin
2019-04-03 16:05:16 +02:00
## Usage
TODO