forked from D-Net/dnet-hadoop
More progress on adding queries to the code. Initial database and table creation seems OK. Downloading logs from available piwik_ids
This commit is contained in:
parent
637e61bb0f
commit
cf7b9c6db3
|
@ -43,7 +43,7 @@ public abstract class ConnectDB {
|
|||
// Class.forName(properties.getProperty("Stats_db_Driver"));
|
||||
|
||||
dbURL = "jdbc:hive2://iis-cdh5-test-m3.ocean.icm.edu.pl:10000/;UseNativeQuery=1";
|
||||
usageStatsDBSchema = "usagestats";
|
||||
usageStatsDBSchema = "usagestats_20200902";
|
||||
statsDBSchema = "openaire_prod_stats_shadow_20200821";
|
||||
|
||||
Class.forName("org.apache.hive.jdbc.HiveDriver");
|
||||
|
|
|
@ -86,9 +86,9 @@ public class PiwikDownloadLogs {
|
|||
Statement statement = ConnectDB.getConnection().createStatement();
|
||||
|
||||
ResultSet rs = statement
|
||||
.executeQuery(
|
||||
"SELECT distinct piwik_id from " + ConnectDB.getStatsDBSchema()
|
||||
+ ".datasource where piwik_id is not null and piwik_id <> 0 order by piwik_id");
|
||||
.executeQuery(
|
||||
"SELECT distinct piwik_id from " + ConnectDB.getStatsDBSchema()
|
||||
+ ".datasource where piwik_id is not null and piwik_id <> 0 order by piwik_id");
|
||||
while (rs.next()) {
|
||||
int siteId = rs.getInt(1);
|
||||
SimpleDateFormat simpleDateFormat = new SimpleDateFormat("YYYY-MM");
|
||||
|
|
|
@ -39,6 +39,7 @@ public class PiwikStatsDB {
|
|||
public PiwikStatsDB(String logRepoPath, String logPortalPath) throws Exception {
|
||||
this.logRepoPath = logRepoPath;
|
||||
this.logPortalPath = logPortalPath;
|
||||
this.createDatabase();
|
||||
this.createTables();
|
||||
// The piwiklog table is not needed since it is built
|
||||
// on top of JSON files
|
||||
|
@ -69,36 +70,46 @@ public class PiwikStatsDB {
|
|||
this.CounterRobotsURL = CounterRobotsURL;
|
||||
}
|
||||
|
||||
private void createDatabase() throws Exception {
|
||||
try {
|
||||
stmt = ConnectDB.getConnection().createStatement();
|
||||
String createDatabase = "CREATE DATABASE IF NOT EXISTS " + ConnectDB.getUsageStatsDBSchema();
|
||||
stmt.executeUpdate(createDatabase);
|
||||
|
||||
} catch (Exception e) {
|
||||
log.error("Failed to create database: " + e);
|
||||
throw new Exception("Failed to create database: " + e.toString(), e);
|
||||
}
|
||||
}
|
||||
|
||||
private void createTables() throws Exception {
|
||||
try {
|
||||
stmt = ConnectDB.getConnection().createStatement();
|
||||
|
||||
// Create Piwiklog table - This table should exist
|
||||
String sqlCreateTablePiwikLog =
|
||||
"CREATE TABLE IF NOT EXISTS"
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".piwiklog(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, action, timestamp, entity_id) "
|
||||
+ "into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
String sqlCreateTablePiwikLog = "CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".piwiklog(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, action, timestamp, entity_id) "
|
||||
+ "into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
stmt.executeUpdate(sqlCreateTablePiwikLog);
|
||||
|
||||
|
||||
/////////////////////////////////////////
|
||||
// Rule for duplicate inserts @ piwiklog
|
||||
/////////////////////////////////////////
|
||||
|
||||
String sqlCreateTablePortalLog =
|
||||
"CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".process_portal_log(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, timestamp) into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
|
||||
String sqlCreateTablePortalLog = "CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".process_portal_log(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, timestamp) into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
stmt.executeUpdate(sqlCreateTablePortalLog);
|
||||
|
||||
|
||||
//////////////////////////////////////////////////
|
||||
// Rule for duplicate inserts @ process_portal_log
|
||||
//////////////////////////////////////////////////
|
||||
|
||||
|
||||
stmt.close();
|
||||
ConnectDB.getConnection().close();
|
||||
log.info("Usage Tables Created");
|
||||
|
@ -112,35 +123,31 @@ public class PiwikStatsDB {
|
|||
private void createTmpTables() throws Exception {
|
||||
try {
|
||||
Statement stmt = ConnectDB.getConnection().createStatement();
|
||||
String sqlCreateTmpTablePiwikLog =
|
||||
"CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".piwiklogtmp(source INT, id_visit STRING, country STRING, action STRING, url STRING, entity_id STRING, "
|
||||
+ "source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, action, timestamp, entity_id) into 100 buckets "
|
||||
+ "stored as orc tblproperties('transactional'='true');";
|
||||
String sqlCreateTmpTablePiwikLog = "CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".piwiklogtmp(source INT, id_visit STRING, country STRING, action STRING, url STRING, entity_id STRING, "
|
||||
+ "source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, action, timestamp, entity_id) into 100 buckets "
|
||||
+ "stored as orc tblproperties('transactional'='true');";
|
||||
stmt.executeUpdate(sqlCreateTmpTablePiwikLog);
|
||||
|
||||
//////////////////////////////////////////////////
|
||||
// Rule for duplicate inserts @ piwiklogtmp
|
||||
//////////////////////////////////////////////////
|
||||
|
||||
|
||||
//////////////////////////////////////////////////
|
||||
// Copy from public.piwiklog to piwiklog
|
||||
//////////////////////////////////////////////////
|
||||
//////////////////////////////////////////////////
|
||||
// String sqlCopyPublicPiwiklog="insert into piwiklog select * from public.piwiklog;";
|
||||
// stmt.executeUpdate(sqlCopyPublicPiwiklog);
|
||||
|
||||
|
||||
|
||||
String sqlCreateTmpTablePortalLog =
|
||||
"CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".process_portal_log_tmp(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, timestamp) into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
|
||||
String sqlCreateTmpTablePortalLog = "CREATE TABLE IF NOT EXISTS "
|
||||
+ ConnectDB.getUsageStatsDBSchema()
|
||||
+ ".process_portal_log_tmp(source INT, id_visit STRING, country STRING, action STRING, url STRING, "
|
||||
+ "entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) "
|
||||
+ "clustered by (source, id_visit, timestamp) into 100 buckets stored as orc tblproperties('transactional'='true')";
|
||||
stmt.executeUpdate(sqlCreateTmpTablePortalLog);
|
||||
|
||||
|
||||
//////////////////////////////////////////////////
|
||||
// Rule for duplicate inserts @ process_portal_log_tmp
|
||||
//////////////////////////////////////////////////
|
||||
|
@ -199,47 +206,43 @@ public class PiwikStatsDB {
|
|||
public void processRepositoryLog() throws Exception {
|
||||
Statement stmt = ConnectDB.getConnection().createStatement();
|
||||
ConnectDB.getConnection().setAutoCommit(false);
|
||||
|
||||
String stm_piwiklogtmp_json =
|
||||
"CREATE EXTERNAL TABLE IF NOT EXISTS " +
|
||||
ConnectDB.getUsageStatsDBSchema() +
|
||||
".piwiklogtmp_json(\n" +
|
||||
" `idSite` STRING,\n" +
|
||||
" `idVisit` STRING,\n" +
|
||||
" `country` STRING,\n" +
|
||||
" `referrerName` STRING,\n" +
|
||||
" `browser` STRING,\n" +
|
||||
" `actionDetails` ARRAY<\n" +
|
||||
" struct<\n" +
|
||||
" type: STRING,\n" +
|
||||
" url: STRING,\n" +
|
||||
" `customVariables`: struct<\n" +
|
||||
" `1`: struct<\n" +
|
||||
" `customVariablePageValue1`: STRING\n" +
|
||||
" >\n" +
|
||||
" >,\n" +
|
||||
" timestamp: String\n" +
|
||||
" >\n" +
|
||||
" >\n" +
|
||||
")\n" +
|
||||
"ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe'\n" +
|
||||
"LOCATION '/user/spyros/logs/usage_stats_logs/Repologs2/'\n" +
|
||||
"TBLPROPERTIES (\"transactional\"=\"false\");\n" +
|
||||
"";
|
||||
|
||||
String stm_piwiklogtmp_json = "CREATE EXTERNAL TABLE IF NOT EXISTS " +
|
||||
ConnectDB.getUsageStatsDBSchema() +
|
||||
".piwiklogtmp_json(\n" +
|
||||
" `idSite` STRING,\n" +
|
||||
" `idVisit` STRING,\n" +
|
||||
" `country` STRING,\n" +
|
||||
" `referrerName` STRING,\n" +
|
||||
" `browser` STRING,\n" +
|
||||
" `actionDetails` ARRAY<\n" +
|
||||
" struct<\n" +
|
||||
" type: STRING,\n" +
|
||||
" url: STRING,\n" +
|
||||
" `customVariables`: struct<\n" +
|
||||
" `1`: struct<\n" +
|
||||
" `customVariablePageValue1`: STRING\n" +
|
||||
" >\n" +
|
||||
" >,\n" +
|
||||
" timestamp: String\n" +
|
||||
" >\n" +
|
||||
" >\n" +
|
||||
")\n" +
|
||||
"ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe'\n" +
|
||||
"LOCATION '/user/spyros/logs/usage_stats_logs/Repologs2/'\n" +
|
||||
"TBLPROPERTIES (\"transactional\"=\"false\");\n" +
|
||||
"";
|
||||
stmt.executeUpdate(stm_piwiklogtmp_json);
|
||||
|
||||
|
||||
String stm_piwiklogtmp =
|
||||
"CREATE TABLE " +
|
||||
ConnectDB.getUsageStatsDBSchema() +
|
||||
".piwiklogtmp (source BIGINT, id_Visit STRING, country STRING, action STRING, url STRING, " +
|
||||
"entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) " +
|
||||
"clustered by (source) into 100 buckets stored as orc tblproperties('transactional'='true');";
|
||||
stmt.executeUpdate(processRepositoryLog);
|
||||
|
||||
|
||||
|
||||
String stm_piwiklogtmp = "CREATE TABLE " +
|
||||
ConnectDB.getUsageStatsDBSchema() +
|
||||
".piwiklogtmp (source BIGINT, id_Visit STRING, country STRING, action STRING, url STRING, " +
|
||||
"entity_id STRING, source_item_type STRING, timestamp STRING, referrer_name STRING, agent STRING) " +
|
||||
"clustered by (source) into 100 buckets stored as orc tblproperties('transactional'='true');";
|
||||
stmt.executeUpdate(stm_piwiklogtmp);
|
||||
|
||||
stmt.close();
|
||||
|
||||
|
||||
// ArrayList<String> jsonFiles = listHdfsDir(this.logRepoPath);
|
||||
//// File dir = new File(this.logRepoPath);
|
||||
//// File[] jsonFiles = dir.listFiles();
|
||||
|
|
|
@ -27,13 +27,13 @@ public class UsageStatsExporter {
|
|||
|
||||
String matomoAuthToken = "703bd17d845acdaf795e01bb1e0895b9";
|
||||
String matomoBaseURL = "analytics.openaire.eu";
|
||||
String repoLogPath = "/user/spyros/logs/usage_stats_logs/Repologs";
|
||||
String portalLogPath = "/user/spyros/logs/usage_stats_logs/Portallogs/";
|
||||
String repoLogPath = "/user/spyros/logs/usage_stats_logs2/Repologs";
|
||||
String portalLogPath = "/user/spyros/logs/usage_stats_logs2/Portallogs/";
|
||||
String portalMatomoID = "109";
|
||||
String irusUKBaseURL = "https://irus.jisc.ac.uk/api/sushilite/v1_7/";
|
||||
|
||||
String irusUKReportPath = "/user/spyros/logs/usage_stats_logs/irusUKReports";
|
||||
String sarcsReportPath = "/user/spyros/logs/usage_stats_logs/sarcReports";
|
||||
String irusUKReportPath = "/user/spyros/logs/usage_stats_logs2/irusUKReports";
|
||||
String sarcsReportPath = "/user/spyros/logs/usage_stats_logs2/sarcReports";
|
||||
|
||||
// connect to DB
|
||||
ConnectDB.init(properties);
|
||||
|
|
Loading…
Reference in New Issue