Merge pull request 'feature/21031' (#1) from feature/21031 into master
Reviewed-on: #1
This commit is contained in:
commit
81d792162d
|
@ -5,6 +5,7 @@ This project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.htm
|
|||
|
||||
## [v2.0.0-SNAPSHOT]
|
||||
|
||||
- Added Jupyter Harvester [#21031]
|
||||
- Switched accounting JSON management to gcube-jackson [#19115]
|
||||
- Switched smart-executor JSON management to gcube-jackson [#19647]
|
||||
|
||||
|
|
|
@ -21,6 +21,7 @@ import org.gcube.common.scope.impl.ScopeBean;
|
|||
import org.gcube.common.scope.impl.ScopeBean.Type;
|
||||
import org.gcube.dataharvest.harvester.CatalogueAccessesHarvester;
|
||||
import org.gcube.dataharvest.harvester.CoreServicesAccessesHarvester;
|
||||
import org.gcube.dataharvest.harvester.JupyterAccessesHarvester;
|
||||
import org.gcube.dataharvest.harvester.MethodInvocationHarvester;
|
||||
import org.gcube.dataharvest.harvester.SocialInteractionsHarvester;
|
||||
import org.gcube.dataharvest.harvester.VREAccessesHarvester;
|
||||
|
@ -53,10 +54,11 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
public static final String DRY_RUN_INPUT_PARAMETER = "dryRun";
|
||||
|
||||
/**
|
||||
* Allows partial harvesting of data of the current period.
|
||||
* This means that in MONTHLY aggregation type the current month is harvested instead of the previous month which
|
||||
* is done when the month is completed.
|
||||
* This allow the portlet to display monthly data in the current moth even the data is partial (till the current day).
|
||||
* Allows partial harvesting of data of the current period. This means that
|
||||
* in MONTHLY aggregation type the current month is harvested instead of the
|
||||
* previous month which is done when the month is completed. This allow the
|
||||
* portlet to display monthly data in the current moth even the data is
|
||||
* partial (till the current day).
|
||||
*/
|
||||
public static final String PARTIAL_HARVESTING = "partialHarvesting";
|
||||
|
||||
|
@ -85,7 +87,6 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
|
||||
};
|
||||
|
||||
|
||||
public static InheritableThreadLocal<Properties> getProperties() {
|
||||
return properties;
|
||||
}
|
||||
|
@ -107,12 +108,11 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
|
||||
};
|
||||
|
||||
|
||||
public static ScopeDescriptor getScopeDescriptor(String context) {
|
||||
return scopeDescriptors.get().get(context);
|
||||
}
|
||||
|
||||
protected static final InheritableThreadLocal<Map<String, ScopeDescriptor>> scopeDescriptors = new InheritableThreadLocal<Map<String, ScopeDescriptor>>() {
|
||||
public static final InheritableThreadLocal<Map<String, ScopeDescriptor>> scopeDescriptors = new InheritableThreadLocal<Map<String, ScopeDescriptor>>() {
|
||||
|
||||
@Override
|
||||
protected Map<String, ScopeDescriptor> initialValue() {
|
||||
|
@ -176,7 +176,7 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
boolean getVREUsers = true;
|
||||
if (inputs.containsKey(GET_VRE_USERS_INPUT_PARAMETER)) {
|
||||
try {
|
||||
reRun = (boolean) inputs.get(GET_VRE_USERS_INPUT_PARAMETER);
|
||||
getVREUsers = (boolean) inputs.get(GET_VRE_USERS_INPUT_PARAMETER);
|
||||
} catch (Exception e) {
|
||||
throw new IllegalArgumentException("'" + GET_VRE_USERS_INPUT_PARAMETER + "' must be a boolean");
|
||||
}
|
||||
|
@ -205,8 +205,8 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
|
||||
end = DateUtils.getEndDateFromStartDate(aggregationType, start, 1, partialHarvesting);
|
||||
|
||||
logger.debug("Harvesting from {} to {} (ReRun:{} - GetVREUsers:{} - DryRun:{})",
|
||||
DateUtils.format(start), DateUtils.format(end), reRun, getVREUsers, dryRun);
|
||||
logger.debug("Harvesting from {} to {} (ReRun:{} - GetVREUsers:{} - DryRun:{})", DateUtils.format(start),
|
||||
DateUtils.format(end), reRun, getVREUsers, dryRun);
|
||||
|
||||
Properties properties = getConfigParameters();
|
||||
getProperties().set(properties);
|
||||
|
@ -227,7 +227,6 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
}
|
||||
scopeDescriptors.set(scopeDescriptorMap);
|
||||
|
||||
|
||||
Set<Dimension> dimensionSet = dao.getDimensions();
|
||||
Map<String, Dimension> dimensionMap = new HashMap<>();
|
||||
for (Dimension dimension : dimensionSet) {
|
||||
|
@ -241,6 +240,7 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
String initialToken = SecurityTokenProvider.instance.get();
|
||||
|
||||
VREAccessesHarvester vreAccessesHarvester = null;
|
||||
JupyterAccessesHarvester jupyterAccessesHarvester = null;
|
||||
|
||||
for (String context : contexts) {
|
||||
// Setting the token for the context
|
||||
|
@ -267,17 +267,18 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
accountingRecords.addAll(records);
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("Error harvesting {} for {}", CatalogueAccessesHarvester.class.getSimpleName(), context, e);
|
||||
logger.error("Error harvesting {} for {}", CatalogueAccessesHarvester.class.getSimpleName(),
|
||||
context, e);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
if (vreAccessesHarvester == null) {
|
||||
|
||||
if (scopeBean.is(Type.INFRASTRUCTURE)) {
|
||||
vreAccessesHarvester = new VREAccessesHarvester(start, end);
|
||||
} else {
|
||||
// This code should be never used because the scopes are sorted by fullname
|
||||
// This code should be never used because the scopes are
|
||||
// sorted by fullname
|
||||
|
||||
ScopeBean parent = scopeBean.enclosingScope();
|
||||
while (!parent.is(Type.INFRASTRUCTURE)) {
|
||||
|
@ -295,6 +296,29 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
|
||||
}
|
||||
|
||||
if (jupyterAccessesHarvester == null) {
|
||||
|
||||
if (scopeBean.is(Type.INFRASTRUCTURE)) {
|
||||
jupyterAccessesHarvester = new JupyterAccessesHarvester(start, end);
|
||||
} else {
|
||||
// This code should be never used because the scopes are
|
||||
// sorted by fullname
|
||||
|
||||
ScopeBean parent = scopeBean.enclosingScope();
|
||||
while (!parent.is(Type.INFRASTRUCTURE)) {
|
||||
parent = scopeBean.enclosingScope();
|
||||
}
|
||||
|
||||
// Setting back token for the context
|
||||
Utils.setContext(contextAuthorization.getTokenForContext(parent.toString()));
|
||||
|
||||
jupyterAccessesHarvester = new JupyterAccessesHarvester(start, end);
|
||||
|
||||
// Setting back token for the context
|
||||
Utils.setContext(contextAuthorization.getTokenForContext(context));
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
if ((context.startsWith(SO_BIG_DATA_VO) || context.startsWith(SO_BIG_DATA_EU_VRE)
|
||||
|| context.startsWith(SO_BIG_DATA_IT_VRE))
|
||||
|
@ -311,8 +335,24 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = vreAccessesHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* vreAccessesHarvester.getData(); data.addAll(harvested);
|
||||
*/
|
||||
} catch (Exception e) {
|
||||
logger.error("Error harvesting VRE Accesses for {}", context, e);
|
||||
}
|
||||
|
||||
try {
|
||||
// Collecting Google Analytics Data for Jupyters Accesses
|
||||
logger.info("Going to harvest Jupyter Accesses for {}", context);
|
||||
|
||||
List<AccountingRecord> harvested = jupyterAccessesHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
* List<HarvestedData> harvested =
|
||||
* jupyterAccessesHarvester.getData();
|
||||
* data.addAll(harvested);
|
||||
*/
|
||||
} catch (Exception e) {
|
||||
logger.error("Error harvesting VRE Accesses for {}", context, e);
|
||||
|
@ -327,8 +367,8 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = socialHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* socialHarvester.getData(); data.addAll(harvested);
|
||||
*/
|
||||
} catch (Exception e) {
|
||||
logger.error("Error harvesting Social Interactions for {}", context, e);
|
||||
|
@ -337,9 +377,11 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
try {
|
||||
// Collecting info on VRE users
|
||||
if (getVREUsers) {
|
||||
// Harvesting Users only for VREs (not for VO and ROOT which is the sum of the children contexts)
|
||||
// Harvesting Users only for VREs (not for VO and ROOT
|
||||
// which is the sum of the children contexts)
|
||||
// The VREUsers can be only Harvested for the last month
|
||||
if(scopeBean.is(Type.VRE) && start.equals(DateUtils.getPreviousPeriod(aggregationType, partialHarvesting).getTime())) {
|
||||
if (scopeBean.is(Type.VRE) && start
|
||||
.equals(DateUtils.getPreviousPeriod(aggregationType, partialHarvesting).getTime())) {
|
||||
logger.info("Going to harvest Context Users for {}", context);
|
||||
VREUsersHarvester vreUsersHarvester = new VREUsersHarvester(start, end);
|
||||
|
||||
|
@ -347,8 +389,9 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = vreUsersHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* vreUsersHarvester.getData();
|
||||
* data.addAll(harvested);
|
||||
*/
|
||||
}
|
||||
}
|
||||
|
@ -359,17 +402,19 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
if (context.startsWith(SO_BIG_DATA_CATALOGUE_CONTEXT)) {
|
||||
|
||||
try {
|
||||
// Collecting info on Resource Catalogue (Dataset, Application, Deliverables, Methods)
|
||||
// Collecting info on Resource Catalogue (Dataset,
|
||||
// Application, Deliverables, Methods)
|
||||
logger.info("Going to harvest Resource Catalogue Information for {}", context);
|
||||
ResourceCatalogueHarvester resourceCatalogueHarvester = new ResourceCatalogueHarvester(start, end,
|
||||
contexts);
|
||||
ResourceCatalogueHarvester resourceCatalogueHarvester = new ResourceCatalogueHarvester(start,
|
||||
end, contexts);
|
||||
|
||||
List<AccountingRecord> harvested = resourceCatalogueHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = resourceCatalogueHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* resourceCatalogueHarvester.getData();
|
||||
* data.addAll(harvested);
|
||||
*/
|
||||
|
||||
} catch (Exception e) {
|
||||
|
@ -402,8 +447,9 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = tagMeMethodInvocationHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* tagMeMethodInvocationHarvester.getData();
|
||||
* data.addAll(harvested);
|
||||
*/
|
||||
|
||||
} catch (Exception e) {
|
||||
|
@ -415,13 +461,13 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
logger.info("Going to harvest Method Invocations for {}", context);
|
||||
MethodInvocationHarvester methodInvocationHarvester = new MethodInvocationHarvester(start, end);
|
||||
|
||||
|
||||
List<AccountingRecord> harvested = methodInvocationHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
|
||||
/*
|
||||
List<HarvestedData> harvested = methodInvocationHarvester.getData();
|
||||
data.addAll(harvested);
|
||||
* List<HarvestedData> harvested =
|
||||
* methodInvocationHarvester.getData();
|
||||
* data.addAll(harvested);
|
||||
*/
|
||||
} catch (Exception e) {
|
||||
logger.error("Error harvesting Method Invocations for {}", context, e);
|
||||
|
@ -432,10 +478,10 @@ public class AccountingDashboardHarvesterPlugin extends Plugin {
|
|||
|
||||
Utils.setContext(initialToken);
|
||||
|
||||
logger.debug("Harvest Measures from {} to {} are {}", DateUtils.format(start), DateUtils.format(end), accountingRecords);
|
||||
logger.debug("Harvest Measures from {} to {} are {}", DateUtils.format(start), DateUtils.format(end),
|
||||
accountingRecords);
|
||||
if (!dryRun) {
|
||||
dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
//dbaseManager.insertMonthlyData(start, end, data, reRun);
|
||||
} else {
|
||||
logger.debug("Harvested measures are {}", accountingRecords);
|
||||
}
|
||||
|
|
|
@ -15,6 +15,7 @@ public enum HarvestedDataKey {
|
|||
MESSAGES_ACCESSES("Messages Accesses"),
|
||||
NOTIFICATIONS_ACCESSES("Notifications Accesses"),
|
||||
PROFILE_ACCESSES("Profile Accesses"),
|
||||
JUPYTER_ACCESSES("Jupyter Accesses"),
|
||||
|
||||
CATALOGUE_ACCESSES("Catalogue Accesses"),
|
||||
CATALOGUE_DATASET_LIST_ACCESSES("Item List"),
|
||||
|
|
|
@ -0,0 +1,407 @@
|
|||
package org.gcube.dataharvest.harvester;
|
||||
|
||||
import static org.gcube.resources.discovery.icclient.ICFactory.clientFor;
|
||||
import static org.gcube.resources.discovery.icclient.ICFactory.queryFor;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.Reader;
|
||||
import java.io.StringReader;
|
||||
import java.security.GeneralSecurityException;
|
||||
import java.security.KeyFactory;
|
||||
import java.security.NoSuchAlgorithmException;
|
||||
import java.security.PrivateKey;
|
||||
import java.security.spec.InvalidKeySpecException;
|
||||
import java.security.spec.PKCS8EncodedKeySpec;
|
||||
import java.time.Instant;
|
||||
import java.time.LocalDate;
|
||||
import java.time.ZoneId;
|
||||
import java.time.format.DateTimeFormatter;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.Date;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
|
||||
import org.gcube.accounting.accounting.summary.access.model.ScopeDescriptor;
|
||||
import org.gcube.accounting.accounting.summary.access.model.update.AccountingRecord;
|
||||
import org.gcube.common.encryption.encrypter.StringEncrypter;
|
||||
import org.gcube.common.resources.gcore.ServiceEndpoint;
|
||||
import org.gcube.common.resources.gcore.ServiceEndpoint.AccessPoint;
|
||||
import org.gcube.common.resources.gcore.ServiceEndpoint.Property;
|
||||
import org.gcube.common.resources.gcore.utils.Group;
|
||||
import org.gcube.common.scope.api.ScopeProvider;
|
||||
import org.gcube.common.scope.impl.ScopeBean;
|
||||
import org.gcube.dataharvest.AccountingDashboardHarvesterPlugin;
|
||||
import org.gcube.dataharvest.datamodel.AnalyticsReportCredentials;
|
||||
import org.gcube.dataharvest.datamodel.HarvestedDataKey;
|
||||
import org.gcube.dataharvest.datamodel.VREAccessesReportRow;
|
||||
import org.gcube.resources.discovery.client.api.DiscoveryClient;
|
||||
import org.gcube.resources.discovery.client.queries.api.SimpleQuery;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import com.google.api.client.googleapis.auth.oauth2.GoogleCredential;
|
||||
import com.google.api.client.googleapis.auth.oauth2.GoogleCredential.Builder;
|
||||
import com.google.api.client.googleapis.javanet.GoogleNetHttpTransport;
|
||||
import com.google.api.client.googleapis.util.Utils;
|
||||
import com.google.api.client.http.HttpTransport;
|
||||
import com.google.api.client.json.JsonFactory;
|
||||
import com.google.api.client.json.gson.GsonFactory;
|
||||
import com.google.api.client.util.PemReader;
|
||||
import com.google.api.client.util.PemReader.Section;
|
||||
import com.google.api.client.util.SecurityUtils;
|
||||
import com.google.api.services.analyticsreporting.v4.AnalyticsReporting;
|
||||
import com.google.api.services.analyticsreporting.v4.AnalyticsReportingScopes;
|
||||
import com.google.api.services.analyticsreporting.v4.model.DateRange;
|
||||
import com.google.api.services.analyticsreporting.v4.model.DateRangeValues;
|
||||
import com.google.api.services.analyticsreporting.v4.model.GetReportsRequest;
|
||||
import com.google.api.services.analyticsreporting.v4.model.GetReportsResponse;
|
||||
import com.google.api.services.analyticsreporting.v4.model.Metric;
|
||||
import com.google.api.services.analyticsreporting.v4.model.Report;
|
||||
import com.google.api.services.analyticsreporting.v4.model.ReportRequest;
|
||||
import com.google.api.services.analyticsreporting.v4.model.ReportRow;
|
||||
|
||||
/**
|
||||
*
|
||||
* @author Giancarlo Panichi (ISTI CNR)
|
||||
*
|
||||
*/
|
||||
public class JupyterAccessesHarvester extends BasicHarvester {
|
||||
|
||||
private static Logger logger = LoggerFactory.getLogger(JupyterAccessesHarvester.class);
|
||||
|
||||
private static final JsonFactory JSON_FACTORY = GsonFactory.getDefaultInstance();
|
||||
|
||||
private static final String SERVICE_ENDPOINT_CATEGORY = "OnlineService";
|
||||
private static final String SERVICE_ENDPOINT_NAME = "BigGAnalyticsReportService";
|
||||
private static final String AP_VIEWS_PROPERTY = "views";
|
||||
private static final String AP_CLIENT_PROPERTY = "clientId";
|
||||
private static final String AP_PRIVATEKEY_PROPERTY = "privateKeyId";
|
||||
private static final String APPLICATION_NAME = "Analytics Reporting";
|
||||
|
||||
private List<VREAccessesReportRow> vreAccesses;
|
||||
|
||||
public JupyterAccessesHarvester(Date start, Date end) throws Exception {
|
||||
super(start, end);
|
||||
logger.debug("JupyerAccessHArvester: {}, {}", start, end);
|
||||
vreAccesses = getAllAccesses(start, end);
|
||||
}
|
||||
|
||||
@Override
|
||||
public List<AccountingRecord> getAccountingRecords() throws Exception {
|
||||
try {
|
||||
String context = org.gcube.dataharvest.utils.Utils.getCurrentContext();
|
||||
|
||||
ArrayList<AccountingRecord> accountingRecords = new ArrayList<AccountingRecord>();
|
||||
|
||||
int measure = 0;
|
||||
|
||||
ScopeBean scopeBean = new ScopeBean(context);
|
||||
String lowerCasedContext = scopeBean.name().toLowerCase();
|
||||
logger.debug("JupyerAccessHArvester lowerCasedContext: {}", lowerCasedContext);
|
||||
for (VREAccessesReportRow row : vreAccesses) {
|
||||
String pagePath = row.getPagePath().toLowerCase();
|
||||
if (pagePath != null && !pagePath.isEmpty()) {
|
||||
if (pagePath.contains(lowerCasedContext)) {
|
||||
if (!pagePath.contains("catalogue")) {
|
||||
if (pagePath.contains("jupyter") || pagePath.contains("jupiter")) {
|
||||
logger.trace("Matched jupyter or jupiter ({}) : {}", lowerCasedContext, pagePath);
|
||||
measure += row.getVisitNumber();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
ScopeDescriptor scopeDescriptor = AccountingDashboardHarvesterPlugin.getScopeDescriptor();
|
||||
|
||||
AccountingRecord ar = new AccountingRecord(scopeDescriptor, instant,
|
||||
getDimension(HarvestedDataKey.JUPYTER_ACCESSES), (long) measure);
|
||||
logger.debug("{} : {}", ar.getDimension().getId(), ar.getMeasure());
|
||||
accountingRecords.add(ar);
|
||||
|
||||
return accountingRecords;
|
||||
|
||||
} catch (Exception e) {
|
||||
throw e;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
* @return a list of {@link VREAccessesReportRow} objects containing the
|
||||
* pagePath and the visit number e.g. VREAccessesReportRow
|
||||
* [pagePath=/group/agroclimaticmodelling/add-new-users,
|
||||
* visitNumber=1] VREAccessesReportRow
|
||||
* [pagePath=/group/agroclimaticmodelling/administration,
|
||||
* visitNumber=2] VREAccessesReportRow
|
||||
* [pagePath=/group/agroclimaticmodelling/agroclimaticmodelling,
|
||||
* visitNumber=39]
|
||||
*/
|
||||
private static List<VREAccessesReportRow> getAllAccesses(Date start, Date end) throws Exception {
|
||||
DateRange dateRange = getDateRangeForAnalytics(start, end);
|
||||
logger.trace("Getting accesses in this time range {}", dateRange.toPrettyString());
|
||||
|
||||
AnalyticsReportCredentials credentialsFromD4S = getAuthorisedApplicationInfoFromIs();
|
||||
AnalyticsReporting service = initializeAnalyticsReporting(credentialsFromD4S);
|
||||
HashMap<String, List<GetReportsResponse>> responses = getReportResponses(service,
|
||||
credentialsFromD4S.getViewIds(), dateRange);
|
||||
List<VREAccessesReportRow> totalAccesses = new ArrayList<>();
|
||||
|
||||
for (String view : responses.keySet()) {
|
||||
List<VREAccessesReportRow> viewReport = parseResponse(view, responses.get(view));
|
||||
logger.trace("Got {} entries from view id={}", viewReport.size(), view);
|
||||
totalAccesses.addAll(viewReport);
|
||||
}
|
||||
logger.trace("Merged in {} total entries from all views", totalAccesses.size());
|
||||
return totalAccesses;
|
||||
}
|
||||
|
||||
/**
|
||||
* Initializes an Analytics Reporting API V4 service object.
|
||||
*
|
||||
* @return An authorized Analytics Reporting API V4 service object.
|
||||
* @throws IOException
|
||||
* @throws GeneralSecurityException
|
||||
*/
|
||||
private static AnalyticsReporting initializeAnalyticsReporting(AnalyticsReportCredentials cred)
|
||||
throws GeneralSecurityException, IOException {
|
||||
HttpTransport httpTransport = GoogleNetHttpTransport.newTrustedTransport();
|
||||
GoogleCredential credential = fromD4SServiceEndpoint(cred).createScoped(AnalyticsReportingScopes.all());
|
||||
|
||||
// Construct the Analytics Reporting service object.
|
||||
return new AnalyticsReporting.Builder(httpTransport, JSON_FACTORY, credential)
|
||||
.setApplicationName(APPLICATION_NAME).build();
|
||||
}
|
||||
|
||||
/**
|
||||
* Queries the Analytics Reporting API V4.
|
||||
*
|
||||
* @param service
|
||||
* An authorized Analytics Reporting API V4 service object.
|
||||
* @return GetReportResponse The Analytics Reporting API V4 response.
|
||||
* @throws IOException
|
||||
*/
|
||||
private static HashMap<String, List<GetReportsResponse>> getReportResponses(AnalyticsReporting service,
|
||||
List<String> viewIDs, DateRange dateRange) throws IOException {
|
||||
|
||||
HashMap<String, List<GetReportsResponse>> reports = new HashMap<>();
|
||||
|
||||
// Create the Metrics object.
|
||||
Metric sessions = new Metric().setExpression("ga:pageviews").setAlias("pages");
|
||||
com.google.api.services.analyticsreporting.v4.model.Dimension pageTitle = new com.google.api.services.analyticsreporting.v4.model.Dimension()
|
||||
.setName("ga:pagePath");
|
||||
|
||||
for (String view : viewIDs) {
|
||||
List<GetReportsResponse> gReportResponses = new ArrayList<>();
|
||||
logger.info("Getting data from Google Analytics for viewid: " + view);
|
||||
boolean iterateMorePages = true;
|
||||
String nextPageToken = null;
|
||||
while (iterateMorePages) {
|
||||
// Create the ReportRequest object.
|
||||
ReportRequest request = new ReportRequest().setViewId(view.trim())
|
||||
.setDateRanges(Arrays.asList(dateRange)).setMetrics(Arrays.asList(sessions))
|
||||
.setDimensions(Arrays.asList(pageTitle));
|
||||
request.setPageSize(1000);
|
||||
request.setPageToken(nextPageToken);
|
||||
ArrayList<ReportRequest> requests = new ArrayList<ReportRequest>();
|
||||
requests.add(request);
|
||||
// Create the GetReportsRequest object.
|
||||
GetReportsRequest getReport = new GetReportsRequest().setReportRequests(requests);
|
||||
// Call the batchGet method.
|
||||
GetReportsResponse response = service.reports().batchGet(getReport).execute();
|
||||
nextPageToken = response.getReports().get(0).getNextPageToken();
|
||||
iterateMorePages = (nextPageToken != null);
|
||||
logger.debug("got nextPageToken: " + nextPageToken);
|
||||
gReportResponses.add(response);
|
||||
}
|
||||
reports.put(view, gReportResponses);
|
||||
}
|
||||
// Return the response.
|
||||
return reports;
|
||||
}
|
||||
|
||||
/**
|
||||
* Parses and prints the Analytics Reporting API V4 response.
|
||||
*
|
||||
* @param response
|
||||
* An Analytics Reporting API V4 response.
|
||||
*/
|
||||
/**
|
||||
* Parses and prints the Analytics Reporting API V4 response.
|
||||
*
|
||||
* @param response
|
||||
* An Analytics Reporting API V4 response.
|
||||
*/
|
||||
private static List<VREAccessesReportRow> parseResponse(String viewId, List<GetReportsResponse> responses) {
|
||||
logger.debug("parsing Response for " + viewId);
|
||||
|
||||
List<VREAccessesReportRow> toReturn = new ArrayList<>();
|
||||
for (GetReportsResponse response : responses) {
|
||||
for (Report report : response.getReports()) {
|
||||
List<ReportRow> rows = report.getData().getRows();
|
||||
if (rows == null) {
|
||||
logger.warn("No data found for " + viewId);
|
||||
} else {
|
||||
for (ReportRow row : rows) {
|
||||
String dimension = row.getDimensions().get(0);
|
||||
DateRangeValues metric = row.getMetrics().get(0);
|
||||
VREAccessesReportRow var = new VREAccessesReportRow();
|
||||
boolean validEntry = false;
|
||||
String pagePath = dimension;
|
||||
if (pagePath.startsWith("/group") || pagePath.startsWith("/web")) {
|
||||
var.setPagePath(dimension);
|
||||
validEntry = true;
|
||||
}
|
||||
if (validEntry) {
|
||||
var.setVisitNumber(Integer.parseInt(metric.getValues().get(0)));
|
||||
toReturn.add(var);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
return toReturn;
|
||||
}
|
||||
|
||||
private static GoogleCredential fromD4SServiceEndpoint(AnalyticsReportCredentials cred) throws IOException {
|
||||
|
||||
String clientId = cred.getClientId();
|
||||
String clientEmail = cred.getClientEmail();
|
||||
String privateKeyPem = cred.getPrivateKeyPem();
|
||||
String privateKeyId = cred.getPrivateKeyId();
|
||||
String tokenUri = cred.getTokenUri();
|
||||
String projectId = cred.getProjectId();
|
||||
|
||||
if (clientId == null || clientEmail == null || privateKeyPem == null || privateKeyId == null) {
|
||||
throw new IOException("Error reading service account credential from stream, "
|
||||
+ "expecting 'client_id', 'client_email', 'private_key' and 'private_key_id'.");
|
||||
}
|
||||
|
||||
PrivateKey privateKey = privateKeyFromPkcs8(privateKeyPem);
|
||||
|
||||
Collection<String> emptyScopes = Collections.emptyList();
|
||||
|
||||
Builder credentialBuilder = new GoogleCredential.Builder().setTransport(Utils.getDefaultTransport())
|
||||
.setJsonFactory(Utils.getDefaultJsonFactory()).setServiceAccountId(clientEmail)
|
||||
.setServiceAccountScopes(emptyScopes).setServiceAccountPrivateKey(privateKey)
|
||||
.setServiceAccountPrivateKeyId(privateKeyId);
|
||||
|
||||
if (tokenUri != null) {
|
||||
credentialBuilder.setTokenServerEncodedUrl(tokenUri);
|
||||
}
|
||||
|
||||
if (projectId != null) {
|
||||
credentialBuilder.setServiceAccountProjectId(projectId);
|
||||
}
|
||||
|
||||
// Don't do a refresh at this point, as it will always fail before the
|
||||
// scopes are added.
|
||||
return credentialBuilder.build();
|
||||
}
|
||||
|
||||
private static PrivateKey privateKeyFromPkcs8(String privateKeyPem) throws IOException {
|
||||
Reader reader = new StringReader(privateKeyPem);
|
||||
Section section = PemReader.readFirstSectionAndClose(reader, "PRIVATE KEY");
|
||||
if (section == null) {
|
||||
throw new IOException("Invalid PKCS8 data.");
|
||||
}
|
||||
byte[] bytes = section.getBase64DecodedBytes();
|
||||
PKCS8EncodedKeySpec keySpec = new PKCS8EncodedKeySpec(bytes);
|
||||
Exception unexpectedException = null;
|
||||
try {
|
||||
KeyFactory keyFactory = SecurityUtils.getRsaKeyFactory();
|
||||
PrivateKey privateKey = keyFactory.generatePrivate(keySpec);
|
||||
return privateKey;
|
||||
} catch (NoSuchAlgorithmException exception) {
|
||||
unexpectedException = exception;
|
||||
} catch (InvalidKeySpecException exception) {
|
||||
unexpectedException = exception;
|
||||
}
|
||||
throw new IOException("Unexpected exception reading PKCS data", unexpectedException);
|
||||
}
|
||||
|
||||
private static List<ServiceEndpoint> getAnalyticsReportingConfigurationFromIS(String infrastructureScope)
|
||||
throws Exception {
|
||||
String scope = infrastructureScope;
|
||||
String currScope = ScopeProvider.instance.get();
|
||||
ScopeProvider.instance.set(scope);
|
||||
SimpleQuery query = queryFor(ServiceEndpoint.class);
|
||||
query.addCondition("$resource/Profile/Category/text() eq '" + SERVICE_ENDPOINT_CATEGORY + "'");
|
||||
query.addCondition("$resource/Profile/Name/text() eq '" + SERVICE_ENDPOINT_NAME + "'");
|
||||
DiscoveryClient<ServiceEndpoint> client = clientFor(ServiceEndpoint.class);
|
||||
List<ServiceEndpoint> toReturn = client.submit(query);
|
||||
ScopeProvider.instance.set(currScope);
|
||||
return toReturn;
|
||||
}
|
||||
|
||||
/**
|
||||
* l
|
||||
*
|
||||
* @throws Exception
|
||||
*/
|
||||
private static AnalyticsReportCredentials getAuthorisedApplicationInfoFromIs() throws Exception {
|
||||
AnalyticsReportCredentials reportCredentials = new AnalyticsReportCredentials();
|
||||
|
||||
String context = org.gcube.dataharvest.utils.Utils.getCurrentContext();
|
||||
try {
|
||||
List<ServiceEndpoint> list = getAnalyticsReportingConfigurationFromIS(context);
|
||||
if (list.size() > 1) {
|
||||
logger.error("Too many Service Endpoints having name " + SERVICE_ENDPOINT_NAME
|
||||
+ " in this scope having Category " + SERVICE_ENDPOINT_CATEGORY);
|
||||
} else if (list.size() == 0) {
|
||||
logger.warn("There is no Service Endpoint having name " + SERVICE_ENDPOINT_NAME + " and Category "
|
||||
+ SERVICE_ENDPOINT_CATEGORY + " in this context: " + context);
|
||||
} else {
|
||||
|
||||
for (ServiceEndpoint res : list) {
|
||||
reportCredentials.setTokenUri(res.profile().runtime().hostedOn());
|
||||
Group<AccessPoint> apGroup = res.profile().accessPoints();
|
||||
AccessPoint[] accessPoints = (AccessPoint[]) apGroup.toArray(new AccessPoint[apGroup.size()]);
|
||||
AccessPoint found = accessPoints[0];
|
||||
reportCredentials.setClientEmail(found.address());
|
||||
reportCredentials.setProjectId(found.username());
|
||||
reportCredentials.setPrivateKeyPem(StringEncrypter.getEncrypter().decrypt(found.password()));
|
||||
for (Property prop : found.properties()) {
|
||||
if (prop.name().compareTo(AP_VIEWS_PROPERTY) == 0) {
|
||||
String decryptedValue = StringEncrypter.getEncrypter().decrypt(prop.value());
|
||||
String[] views = decryptedValue.split(";");
|
||||
reportCredentials.setViewIds(Arrays.asList(views));
|
||||
}
|
||||
if (prop.name().compareTo(AP_CLIENT_PROPERTY) == 0) {
|
||||
String decryptedValue = StringEncrypter.getEncrypter().decrypt(prop.value());
|
||||
reportCredentials.setClientId(decryptedValue);
|
||||
}
|
||||
if (prop.name().compareTo(AP_PRIVATEKEY_PROPERTY) == 0) {
|
||||
String decryptedValue = StringEncrypter.getEncrypter().decrypt(prop.value());
|
||||
reportCredentials.setPrivateKeyId(decryptedValue);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (Exception e) {
|
||||
e.printStackTrace();
|
||||
return null;
|
||||
}
|
||||
return reportCredentials;
|
||||
}
|
||||
|
||||
private static LocalDate asLocalDate(Date date) {
|
||||
return Instant.ofEpochMilli(date.getTime()).atZone(ZoneId.systemDefault()).toLocalDate();
|
||||
}
|
||||
|
||||
private static DateRange getDateRangeForAnalytics(Date start, Date end) {
|
||||
DateTimeFormatter formatter = DateTimeFormatter.ofPattern("yyyy-MM-dd"); // required
|
||||
// by
|
||||
// Analytics
|
||||
String startDate = asLocalDate(start).format(formatter);
|
||||
String endDate = asLocalDate(end).format(formatter);
|
||||
DateRange dateRange = new DateRange();// date format `yyyy-MM-dd`
|
||||
dateRange.setStartDate(startDate);
|
||||
dateRange.setEndDate(endDate);
|
||||
return dateRange;
|
||||
}
|
||||
|
||||
}
|
|
@ -0,0 +1,134 @@
|
|||
package org.gcube.dataharvest;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Calendar;
|
||||
import java.util.Date;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.SortedSet;
|
||||
|
||||
import org.gcube.accounting.accounting.summary.access.AccountingDao;
|
||||
import org.gcube.accounting.accounting.summary.access.model.ScopeDescriptor;
|
||||
import org.gcube.accounting.accounting.summary.access.model.internal.Dimension;
|
||||
import org.gcube.accounting.accounting.summary.access.model.update.AccountingRecord;
|
||||
import org.gcube.common.authorization.client.exceptions.ObjectNotFound;
|
||||
import org.gcube.common.scope.impl.ScopeBean;
|
||||
import org.gcube.dataharvest.harvester.JupyterAccessesHarvester;
|
||||
import org.gcube.dataharvest.utils.AggregationType;
|
||||
import org.gcube.dataharvest.utils.ContextAuthorization;
|
||||
import org.gcube.dataharvest.utils.ContextTest;
|
||||
import org.gcube.dataharvest.utils.DateUtils;
|
||||
import org.junit.Test;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
*
|
||||
* @author Giancarlo Panichi (ISTI CNR)
|
||||
*
|
||||
*/
|
||||
public class AccountingDataHarvesterJupyterTest extends ContextTest {
|
||||
|
||||
private static Logger logger = LoggerFactory.getLogger(AccountingDataHarvesterJupyterTest.class);
|
||||
public static final String ROOT = "/d4science.research-infrastructures.eu";
|
||||
// private static final String SCOPE = "/d4science.research-infrastructures.eu/D4OS/Blue-CloudLab";
|
||||
|
||||
protected AccountingDao getAccountingDao() throws ObjectNotFound, Exception {
|
||||
AccountingDao dao = AccountingDao.get();
|
||||
|
||||
Set<ScopeDescriptor> scopeDescriptorSet = dao.getContexts();
|
||||
Map<String, ScopeDescriptor> scopeDescriptorMap = new HashMap<>();
|
||||
for (ScopeDescriptor scopeDescriptor : scopeDescriptorSet) {
|
||||
scopeDescriptorMap.put(scopeDescriptor.getId(), scopeDescriptor);
|
||||
}
|
||||
AccountingDashboardHarvesterPlugin.scopeDescriptors.set(scopeDescriptorMap);
|
||||
|
||||
Set<Dimension> dimensionSet = dao.getDimensions();
|
||||
Map<String, Dimension> dimensionMap = new HashMap<>();
|
||||
for (Dimension dimension : dimensionSet) {
|
||||
dimensionMap.put(dimension.getId(), dimension);
|
||||
}
|
||||
|
||||
AccountingDashboardHarvesterPlugin.dimensions.set(dimensionMap);
|
||||
|
||||
return dao;
|
||||
}
|
||||
|
||||
//@Ignore
|
||||
@Test
|
||||
public void testJupyterAccessesHarvester() throws Exception {
|
||||
try {
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
AccountingDao dao = getAccountingDao();
|
||||
|
||||
List<Date> starts = new ArrayList<>();
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.JANUARY, 1).getTime());
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.FEBRUARY, 1).getTime());
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.MARCH, 1).getTime());
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.APRIL, 1).getTime());
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.MAY, 1).getTime());
|
||||
|
||||
AggregationType measureType = AggregationType.MONTHLY;
|
||||
|
||||
ContextAuthorization contextAuthorization = new ContextAuthorization();
|
||||
|
||||
SortedSet<String> contexts = contextAuthorization.getContexts();
|
||||
/*
|
||||
SortedSet<String> contexts = new TreeSet<>();
|
||||
contexts.add("/d4science.research-infrastructures.eu/D4OS/Blue-CloudLab");
|
||||
contexts.add("/d4science.research-infrastructures.eu/D4OS/Zoo-Phytoplankton_EOV");
|
||||
contexts.add("/d4science.research-infrastructures.eu/D4OS/MarineEnvironmentalIndicators");
|
||||
*/
|
||||
|
||||
List<AccountingRecord> accountingRecords = new ArrayList<>();
|
||||
|
||||
Set<ScopeDescriptor> scopeDescriptorSet = dao.getContexts();
|
||||
Map<String, ScopeDescriptor> scopeDescriptorMap = new HashMap<>();
|
||||
for (ScopeDescriptor scopeDescriptor : scopeDescriptorSet) {
|
||||
scopeDescriptorMap.put(scopeDescriptor.getId(), scopeDescriptor);
|
||||
}
|
||||
AccountingDashboardHarvesterPlugin.scopeDescriptors.set(scopeDescriptorMap);
|
||||
|
||||
|
||||
for (Date start : starts) {
|
||||
Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
JupyterAccessesHarvester jupyterAccessesHarvester = new JupyterAccessesHarvester(start, end);
|
||||
|
||||
for(String context : contexts) {
|
||||
ContextTest.setContext(contextAuthorization.getTokenForContext(context));
|
||||
|
||||
ScopeBean scopeBean = new ScopeBean(context);
|
||||
|
||||
ScopeDescriptor actualScopeDescriptor = scopeDescriptorMap.get(context);
|
||||
if (actualScopeDescriptor == null) {
|
||||
actualScopeDescriptor = new ScopeDescriptor(scopeBean.name(), context);
|
||||
}
|
||||
|
||||
AccountingDashboardHarvesterPlugin.scopeDescriptor.set(actualScopeDescriptor);
|
||||
|
||||
List<AccountingRecord> harvested = jupyterAccessesHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
// logger.debug("{}", accountingRecords);
|
||||
|
||||
logger.debug("Going to insert {}", accountingRecords);
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
|
||||
} catch (Throwable e) {
|
||||
logger.error(e.getLocalizedMessage(), e);
|
||||
throw e;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
}
|
|
@ -41,6 +41,7 @@ import org.gcube.vremanagement.executor.api.types.Scheduling;
|
|||
import org.gcube.vremanagement.executor.client.SmartExecutorClient;
|
||||
import org.gcube.vremanagement.executor.client.SmartExecutorClientFactory;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
import org.quartz.CronExpression;
|
||||
import org.slf4j.Logger;
|
||||
|
@ -70,7 +71,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
return contexts;
|
||||
}
|
||||
|
||||
// @Test
|
||||
@Ignore
|
||||
@Test
|
||||
public void getDimensions() {
|
||||
try {
|
||||
|
||||
|
@ -90,6 +92,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
//@Ignore
|
||||
@Test
|
||||
public void launch() {
|
||||
try {
|
||||
|
@ -105,15 +108,13 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
inputs.put(AccountingDashboardHarvesterPlugin.MEASURE_TYPE_INPUT_PARAMETER, aggregationType.name());
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.GET_VRE_USERS_INPUT_PARAMETER, true);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.RERUN_INPUT_PARAMETER, true);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.DRY_RUN_INPUT_PARAMETER, false);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.PARTIAL_HARVESTING, false);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.DRY_RUN_INPUT_PARAMETER, true);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.PARTIAL_HARVESTING, true);
|
||||
|
||||
/*
|
||||
Calendar from = DateUtils.getStartCalendar(2020, Calendar.MAY, 1);
|
||||
Calendar from = DateUtils.getStartCalendar(2021, Calendar.JANUARY, 1);
|
||||
String fromDate = DateUtils.LAUNCH_DATE_FORMAT.format(from.getTime());
|
||||
logger.trace("{} is {}", AccountingDashboardHarvesterPlugin.START_DATE_INPUT_PARAMETER, fromDate);
|
||||
inputs.put(AccountingDashboardHarvesterPlugin.START_DATE_INPUT_PARAMETER, fromDate);
|
||||
*/
|
||||
|
||||
accountingDataHarvesterPlugin.launch(inputs);
|
||||
|
||||
|
@ -124,7 +125,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
// @Test
|
||||
@Ignore
|
||||
@Test
|
||||
public void launchPluginOnSmartExecutor() {
|
||||
try {
|
||||
|
||||
|
@ -146,14 +148,17 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
inputs.put(AccountingDashboardHarvesterPlugin.PARTIAL_HARVESTING, false);
|
||||
|
||||
/*
|
||||
Calendar from = DateUtils.getStartCalendar(2016, Calendar.SEPTEMBER, 1);
|
||||
String fromDate = DateUtils.LAUNCH_DATE_FORMAT.format(from.getTime());
|
||||
logger.trace("{} is {}", AccountingDataHarvesterPlugin.START_DATE_INPUT_PARAMETER, fromDate);
|
||||
inputs.put(AccountingDataHarvesterPlugin.START_DATE_INPUT_PARAMETER, fromDate);
|
||||
* Calendar from = DateUtils.getStartCalendar(2016, Calendar.SEPTEMBER, 1);
|
||||
* String fromDate = DateUtils.LAUNCH_DATE_FORMAT.format(from.getTime());
|
||||
* logger.trace("{} is {}",
|
||||
* AccountingDataHarvesterPlugin.START_DATE_INPUT_PARAMETER, fromDate);
|
||||
* inputs.put(AccountingDataHarvesterPlugin. START_DATE_INPUT_PARAMETER,
|
||||
* fromDate);
|
||||
*/
|
||||
|
||||
// 3rd of the month for MONTHLY Harvesting at 10:00
|
||||
// CronExpression cronExpression = new CronExpression("0 0 10 3 1/1 ? *");
|
||||
// CronExpression cronExpression = new CronExpression("0 0 10 3 1/1
|
||||
// ? *");
|
||||
|
||||
// Every day at 10:00 for partial harvesting
|
||||
CronExpression cronExpression = new CronExpression("0 0 10 3 1/1 ? *");
|
||||
|
@ -161,7 +166,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
Scheduling scheduling = new Scheduling(cronExpression);
|
||||
scheduling.setGlobal(false);
|
||||
LaunchParameter launchParameter = new LaunchParameter(pluginName, inputs, scheduling);
|
||||
//LaunchParameter launchParameter = new LaunchParameter(pluginName, inputs);
|
||||
// LaunchParameter launchParameter = new LaunchParameter(pluginName,
|
||||
// inputs);
|
||||
|
||||
smartExecutor.launch(launchParameter);
|
||||
|
||||
|
@ -172,7 +178,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
// @Test
|
||||
//@Ignore
|
||||
@Test
|
||||
public void launchOldData() {
|
||||
try {
|
||||
|
||||
|
@ -208,6 +215,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void launchOldDataVREAccessesHarvester() {
|
||||
try {
|
||||
|
@ -249,7 +257,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
if (scopeBean.is(Type.INFRASTRUCTURE)) {
|
||||
vreAccessesHarvester = new VREAccessesHarvester(start, end);
|
||||
} else {
|
||||
// This code should be never used because the scopes are sorted by fullname
|
||||
// This code should be never used because the scopes
|
||||
// are sorted by fullname
|
||||
|
||||
ScopeBean parent = scopeBean.enclosingScope();
|
||||
while (!parent.is(Type.INFRASTRUCTURE)) {
|
||||
|
@ -273,7 +282,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
logger.info("Not Harvesting VREs Accesses for {} from {} to {}", context,
|
||||
DateUtils.format(start), DateUtils.format(end));
|
||||
} else {
|
||||
// Collecting Google Analytics Data for VREs Accesses
|
||||
// Collecting Google Analytics Data for VREs
|
||||
// Accesses
|
||||
List<AccountingRecord> harvested = vreAccessesHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
|
||||
|
@ -287,7 +297,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
accountingRecords);
|
||||
ContextTest.setContextByName(ROOT);
|
||||
|
||||
// dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
// dao.insertRecords(accountingRecords.toArray(new
|
||||
// AccountingRecord[1]));
|
||||
|
||||
Thread.sleep(TimeUnit.SECONDS.toMillis(10));
|
||||
|
||||
|
@ -305,6 +316,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testScopeBean() throws Exception {
|
||||
ContextTest.setContextByName(ROOT);
|
||||
|
@ -320,11 +332,13 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
for (String context : contexts) {
|
||||
ScopeBean scopeBean = new ScopeBean(context);
|
||||
// logger.debug("FullName {} - Name {}", scopeBean.toString(), scopeBean.name());
|
||||
// logger.debug("FullName {} - Name {}", scopeBean.toString(),
|
||||
// scopeBean.name());
|
||||
|
||||
try {
|
||||
|
||||
if(scopeBean.is(Type.VRE) && start.equals(DateUtils.getPreviousPeriod(aggregationType, false).getTime())) {
|
||||
if (scopeBean.is(Type.VRE)
|
||||
&& start.equals(DateUtils.getPreviousPeriod(aggregationType, false).getTime())) {
|
||||
logger.info("Harvesting (VRE Users) for {} from {} to {}", context, DateUtils.format(start),
|
||||
DateUtils.format(end));
|
||||
} else {
|
||||
|
@ -351,7 +365,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
}
|
||||
|
||||
// @Test
|
||||
@Ignore
|
||||
@Test
|
||||
public void testVREAccessesHarvester() throws Exception {
|
||||
try {
|
||||
// AccountingDao dao = getAccountingDao();
|
||||
|
@ -393,7 +408,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
logger.debug("{}", accountingRecords);
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
// dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
// dao.insertRecords(accountingRecords.toArray(new
|
||||
// AccountingRecord[1]));
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
|
@ -401,6 +417,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testVREAccessesHarvesterAll() {
|
||||
try {
|
||||
|
@ -409,8 +426,10 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
AggregationType measureType = AggregationType.MONTHLY;
|
||||
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
|
||||
Date start = DateUtils.getPreviousPeriod(measureType, false).getTime();
|
||||
Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
|
@ -436,7 +455,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
if (scopeBean.is(Type.INFRASTRUCTURE)) {
|
||||
vreAccessesHarvester = new VREAccessesHarvester(start, end);
|
||||
} else {
|
||||
// This code should be never used because the scopes are sorted by fullname
|
||||
// This code should be never used because the scopes are
|
||||
// sorted by fullname
|
||||
|
||||
ScopeBean parent = scopeBean.enclosingScope();
|
||||
while (!parent.is(Type.INFRASTRUCTURE)) {
|
||||
|
@ -476,6 +496,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testSocialInteraction() {
|
||||
try {
|
||||
|
@ -524,7 +545,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
logger.debug("Harvest Measures from {} to {} are {}", DateUtils.format(start), DateUtils.format(end),
|
||||
accountingRecords);
|
||||
ContextTest.setContextByName(ROOT);
|
||||
// dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
// dao.insertRecords(accountingRecords.toArray(new
|
||||
// AccountingRecord[1]));
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
|
@ -534,6 +556,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testMethodInvocation() {
|
||||
try {
|
||||
|
@ -559,6 +582,51 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMethodInvocationOldData() {
|
||||
try {
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
AccountingDao dao = getAccountingDao();
|
||||
|
||||
List<Date> starts = new ArrayList<>();
|
||||
starts.add(DateUtils.getStartCalendar(2021, Calendar.APRIL, 1).getTime());
|
||||
|
||||
AggregationType measureType = AggregationType.MONTHLY;
|
||||
|
||||
ContextAuthorization contextAuthorization = new ContextAuthorization();
|
||||
SortedSet<String> contexts = contextAuthorization.getContexts();
|
||||
|
||||
List<AccountingRecord> accountingRecords = new ArrayList<>();
|
||||
|
||||
for (Date start : starts) {
|
||||
Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
|
||||
MethodInvocationHarvester methodInvocationHarvester = new MethodInvocationHarvester(start, end);
|
||||
|
||||
for (String context : contexts) {
|
||||
// Setting the token for the context
|
||||
ContextTest.setContext(contextAuthorization.getTokenForContext(context));
|
||||
|
||||
|
||||
List<AccountingRecord> harvested = methodInvocationHarvester.getAccountingRecords();
|
||||
accountingRecords.addAll(harvested);
|
||||
|
||||
logger.debug("{} - {}", context, harvested);
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
logger.debug("Going to insert {}", accountingRecords);
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
}
|
||||
}
|
||||
|
||||
protected AccountingDao getAccountingDao() throws ObjectNotFound, Exception {
|
||||
AccountingDao dao = AccountingDao.get();
|
||||
|
||||
|
@ -596,6 +664,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
AccountingDashboardHarvesterPlugin.scopeDescriptor.set(actualScopeDescriptor);
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testTagMeMethodInvocation() throws Exception {
|
||||
try {
|
||||
|
@ -625,7 +694,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
// dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
// dao.insertRecords(accountingRecords.toArray(new
|
||||
// AccountingRecord[1]));
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
|
@ -633,6 +703,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testGetVREUsersForSpecificVRE() {
|
||||
try {
|
||||
|
@ -659,7 +730,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
logger.info("Harvested Data from {} to {} : {}", DateUtils.format(start), DateUtils.format(end), harvested);
|
||||
|
||||
ContextTest.setContextByName(ROOT);
|
||||
// dao.insertRecords(accountingRecords.toArray(new AccountingRecord[1]));
|
||||
// dao.insertRecords(accountingRecords.toArray(new
|
||||
// AccountingRecord[1]));
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
|
@ -667,6 +739,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testFilteringGenericResource() {
|
||||
try {
|
||||
|
@ -694,6 +767,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testResourceCatalogueHarvester() {
|
||||
try {
|
||||
|
@ -703,8 +777,10 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
AggregationType measureType = AggregationType.MONTHLY;
|
||||
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
|
||||
Date start = DateUtils.getPreviousPeriod(measureType, false).getTime();
|
||||
Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
|
@ -725,6 +801,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Test
|
||||
public void testCoreServicesHarvester() {
|
||||
try {
|
||||
|
@ -755,13 +832,12 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
end = DateUtils.getEndDateFromStartDate(AggregationType.MONTHLY, start, 1, false);
|
||||
}
|
||||
|
||||
|
||||
} catch (Exception e) {
|
||||
logger.error("", e);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@Ignore
|
||||
@Test
|
||||
public void testCatalogueHarvester() {
|
||||
try {
|
||||
|
@ -775,8 +851,8 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
Date end = DateUtils.getStartCalendar(2020, Calendar.FEBRUARY, 1).getTime();
|
||||
|
||||
/*
|
||||
* Date start = DateUtils.getPreviousPeriod(measureType, false).getTime();
|
||||
* Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
* Date start = DateUtils.getPreviousPeriod(measureType, false).getTime(); Date
|
||||
* end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
*/
|
||||
|
||||
ScopeBean scopeBean = new ScopeBean(context);
|
||||
|
@ -796,7 +872,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
}
|
||||
}
|
||||
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testDataMethodDownloadHarvester() {
|
||||
try {
|
||||
|
@ -806,8 +882,10 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
AggregationType measureType = AggregationType.MONTHLY;
|
||||
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY, 1).getTime();
|
||||
// Date start = DateUtils.getStartCalendar(2015, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
// Date end = DateUtils.getStartCalendar(2019, Calendar.FEBRUARY,
|
||||
// 1).getTime();
|
||||
|
||||
Date start = DateUtils.getPreviousPeriod(measureType, false).getTime();
|
||||
Date end = DateUtils.getEndDateFromStartDate(measureType, start, 1, false);
|
||||
|
@ -844,6 +922,7 @@ public class AccountingDataHarvesterPluginTest extends ContextTest {
|
|||
|
||||
public static final String E_LEARNING_AREA_VRE = "/d4science.research-infrastructures.eu/SoBigData/E-Learning_Area";
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void addMissingVREAccesses() {
|
||||
try {
|
||||
|
|
|
@ -3,6 +3,7 @@ package org.gcube.dataharvest.harvester.sobigdata;
|
|||
import java.util.List;
|
||||
|
||||
import org.gcube.dataharvest.utils.ContextTest;
|
||||
import org.junit.Ignore;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
@ -10,6 +11,7 @@ public class SoBigDataHarvesterTest extends ContextTest {
|
|||
|
||||
private static Logger logger = LoggerFactory.getLogger(SoBigDataHarvesterTest.class);
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testGroupList() throws Exception {
|
||||
// ContextTest.setContextByName("/d4science.research-infrastructures.eu/D4Research/AGINFRAplusDev");
|
||||
|
|
|
@ -5,6 +5,7 @@ import java.io.InputStream;
|
|||
import java.util.Properties;
|
||||
|
||||
import org.gcube.dataharvest.AccountingDashboardHarvesterPlugin;
|
||||
import org.junit.Ignore;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
@ -26,6 +27,7 @@ public class ContextAuthorizationTest extends ContextTest {
|
|||
AccountingDashboardHarvesterPlugin.getProperties().set(properties);
|
||||
}
|
||||
|
||||
@Ignore
|
||||
// @Test
|
||||
public void testRetrieveContextsAndTokens() throws Exception {
|
||||
try {
|
||||
|
|
|
@ -1,3 +1,4 @@
|
|||
/*.gcubekey
|
||||
/*.key
|
||||
/*.properties
|
||||
/howto.txt
|
||||
|
|
|
@ -0,0 +1,4 @@
|
|||
groupId=org.gcube.accounting
|
||||
artifactId=accounting-dashboard-harvester-se-plugin
|
||||
version=2.0.0-SNAPSHOT
|
||||
description=Accounting Dashboard Harvester Smart Executor Plugin.
|
Loading…
Reference in New Issue