From f8d8b514fd280ffd9df142fa5f74b783823f3e1a Mon Sep 17 00:00:00 2001 From: Konstantinos Triantafyllou Date: Fri, 9 Oct 2020 16:37:06 +0000 Subject: [PATCH] [Graph | Trunk]: Add missing properties. Change logos on home slider. Fix margins and paddings to be the same in all pages git-svn-id: https://svn.driver.research-infrastructures.eu/driver/dnet40/modules/uoa-graph-portal/trunk@59546 d315682c-612b-4755-9ff5-7f18f6832af3 --- src/app/about/about.component.css | 4 +- src/app/about/about.component.html | 805 +++++++++++------- src/app/contact/contact.component.html | 80 +- src/app/contact/contact.component.ts | 2 +- src/app/home/home.component.css | 15 +- src/app/home/home.component.html | 586 ++++++------- src/app/resources/resources.component.html | 120 +-- .../graph-assets/about/infrastructure.png | Bin 415877 -> 423131 bytes src/environments/environment.beta.ts | 4 +- src/environments/environment.prod.ts | 4 +- src/environments/environment.ts | 4 +- 11 files changed, 917 insertions(+), 707 deletions(-) diff --git a/src/app/about/about.component.css b/src/app/about/about.component.css index 1da3971..4193478 100644 --- a/src/app/about/about.component.css +++ b/src/app/about/about.component.css @@ -98,6 +98,6 @@ align-self: start; } -.uk-align-right\@m .uk-list>li::before, .uk-align-right\@m .uk-list>li::after { +.align-list li:before, .align-list li:after { display: inline; -} \ No newline at end of file +} diff --git a/src/app/about/about.component.html b/src/app/about/about.component.html index 785e711..74c9858 100644 --- a/src/app/about/about.component.html +++ b/src/app/about/about.component.html @@ -1,128 +1,163 @@
-
+
+
+
+

About

+
+
+

+ Open Science is gradually becoming the modus operandi in research practices, affecting the way researchers + collaborate and publish, discover, and access scientific knowledge. + Scientists are increasingly publishing research results beyond the article, to share all scientific + products (metadata and files) generated during an experiment, such as datasets, software, experiments. + They publish in scholarly communication data sources (e.g. institutional repositories, data archives, + software repositories), rely where possible on persistent identifiers (e.g. DOI, ORCID, Grid.ac, PDBs), + specify semantic links to other research products (e.g. supplementedBy, citedBy, versionOf), and possibly + to projects and/or relative funders. + By following such practices, scientists are implicitly constructing the Global Open Science Graph, where + by "graph" we mean a collection of objects interlinked by semantic relationships. +

+ The OpenAIRE Research Graph includes metadata and links between scientific products (e.g. literature, + datasets, software, and "other research products"), organizations, funders, funding streams, projects, + communities, and (provenance) data sources - the details of the graph data model can be found + in Zenodo.org. +

+ The Graph is available and obtained as an aggregation of the metadata and links collected from ~70.000 + trusted sources, further enriched with metadata and links provided by:

+
    +
  • OpenAIRE end-users, e.g. researchers, project administrators, data curators + providing links from scientific products to projects, funders, communities, or other products; +
  • +
  • OpenAIRE Full-text mining algorithms over around ~10Mi Open Access article + full-texts; +
  • +
  • Research infrastructure scholarly services, bridged to the graph via OpenAIRE, exposing metadata of + products such as research workflows, experiments, research objects, software, etc.. +
  • +
+
+
+
+
- - -
-
-

About

+
+
+

Architecture

-
-

- Open Science is gradually becoming the modus operandi in research practices, affecting the way researchers collaborate and publish, discover, and access scientific knowledge. - Scientists are increasingly publishing research results beyond the article, to share all scientific products (metadata and files) generated during an experiment, such as datasets, software, experiments. - They publish in scholarly communication data sources (e.g. institutional repositories, data archives, software repositories), rely where possible on persistent identifiers (e.g. DOI, ORCID, Grid.ac, PDBs), specify semantic links to other research products (e.g. supplementedBy, citedBy, versionOf), and possibly to projects and/or relative funders. - By following such practices, scientists are implicitly constructing the Global Open Science Graph, where by "graph" we mean a collection of objects interlinked by semantic relationships. -

- The OpenAIRE Research Graph includes metadata and links between scientific products (e.g. literature, datasets, software, and "other research products"), organizations, funders, funding streams, projects, communities, and (provenance) data sources - the details of the graph data model can be found in Zenodo.org. -

- The Graph is available and obtained as an aggregation of the metadata and links collected from ~70.000 trusted sources, further enriched with metadata and links provided by:

-
    -
  • OpenAIRE end-users, e.g. researchers, project administrators, data curators providing links from scientific products to projects, funders, communities, or other products;
  • -
  • OpenAIRE Full-text mining algorithms over around ~10Mi Open Access article full-texts;
  • -
  • Research infrastructure scholarly services, bridged to the graph via OpenAIRE, exposing metadata of products such as research workflows, experiments, research objects, software, etc..
  • -
+
+

How we build it

+
+

+ OpenAIRE collects metadata records from more than 70K scholarly communication sources from all over the + world, including Open Access institutional repositories, data archives, journals. + All the metadata records (i.e. descriptions of research products) are put together in a data lake, + together + with records from Crossref, Unpaywall, ORCID, Grid.ac, and information about projects provided by national + and international funders. + Dedicated inference algorithms applied to metadata and to the full-texts of Open Access publications + enrich + the content of the data lake with links between research results and projects, author affiliations, + subject + classification, links to entries from domain-specific databases. + Duplicated organisations and results are identified and merged together to obtain an open, trusted, public + resource enabling explorations of the scholarly communication landscape like never before. +

+ +
-
-
+
+ - -
- -

Architecture

-
-
-

How we build it

-
-

- OpenAIRE collects metadata records from more than 70K scholarly communication sources from all over the world, including Open Access institutional repositories, data archives, journals. - All the metadata records (i.e. descriptions of research products) are put together in a data lake, together with records from Crossref, Unpaywall, ORCID, Grid.ac, and information about projects provided by national and international funders. - Dedicated inference algorithms applied to metadata and to the full-texts of Open Access publications enrich the content of the data lake with links between research results and projects, author affiliations, subject classification, links to entries from domain-specific databases. - Duplicated organisations and results are identified and merged together to obtain an open, trusted, public resource enabling explorations of the scholarly communication landscape like never before. -

- -
+ + action point aggregation + + + action point deduplication + + + action point enrichment + + + action point post cleaning + + + action point indexing + + + action point stats analysis +
-
- +
+ -
- - -
    -
  • - -
    - +
      +
    • + +
      + Aggregation -
      +
      OpenAIRE collects metadata records from a variety of content providers as described in https://www.openaire.eu/aggregation-and-content-provision-workflows.

      - OpenAIRE aggregates metadata records describing objects of the research life-cycle from content providers compliant to the + OpenAIRE aggregates metadata records describing objects of the research life-cycle from content + providers compliant to the OpenAIRE guidelines - and from entity registries (i.e. data sources offering authoritative lists of entities, like OpenDOAR, re3data, DOAJ, and funder databases). - After collection, metadata are transformed according to the OpenAIRE internal metadata model, which is used to generate the final OpenAIRE Research Graph that you can access from the OpenAIRE portal and the APIs. + and from entity registries (i.e. data sources offering authoritative lists of entities, like OpenDOAR, + re3data, DOAJ, and funder databases). + After collection, metadata are transformed according to the OpenAIRE internal metadata model, which is + used to generate the final OpenAIRE Research Graph that you can access from the OpenAIRE portal and + the + APIs.

      - The transformation process includes the application of cleaning functions whose goal is to ensure that values are harmonised according to a common format (e.g. dates as YYYY-MM-dd) and, whenever applicable, to a common controlled vocabulary. + The transformation process includes the application of cleaning functions whose goal is to ensure that + values are harmonised according to a common format (e.g. dates as YYYY-MM-dd) and, whenever + applicable, + to a common controlled vocabulary. The controlled vocabularies used for cleansing are accessible at http://api.openaire.eu/vocabularies. - Each vocabulary features a set of controlled terms, each with one code, one label, and a set of synonyms. + Each vocabulary features a set of controlled terms, each with one code, one label, and a set of + synonyms. If a synonym is found as field value, the value is updated with the corresponding term. - Also, the OpenAIRE Research Graph is extended with other relevant scholarly communication sources that are too big to be integrated via the “normal” aggregation mechanism: DOIBoost (which merges Crossref, ORCID, Microsoft Academic Graph, and Unpaywall), and ScholeXplorer, one of the Scholix hubs offering a large set of links between research literature and data. + Also, the OpenAIRE Research Graph is extended with other relevant scholarly communication sources that + are too big to be integrated via the “normal” aggregation mechanism: DOIBoost (which merges Crossref, + ORCID, Microsoft Academic Graph, and Unpaywall), and ScholeXplorer, one of the Scholix hubs offering a + large set of links between research literature and data.
      - - - - -
      -
    • -
    • -
      - + + + + +
      +
    • +
    • +
      +
      -
        +
        • Deduplication -
          +
          - Clustering is a common heuristics used to overcome the N x N complexity required to match all pairs of objects to identify the equivalent ones. - The challenge is to identify a clustering function that maximizes the chance of comparing only records that may lead to a match, while minimizing the number of records that will not be matched while being equivalent. - Since the equivalence function is to some level tolerant to minimal errors (e.g. switching of characters in the title, or minimal difference in letters), we need this function to be not too precise (e.g. a hash of the title), but also not too flexible (e.g. random ngrams of the title). - On the other hand, reality tells us that in some cases equality of two records can only be determined by their PIDs (e.g. DOI) as the metadata properties are very different across different versions and no clustering function will ever bring them into the same cluster. + Clustering is a common heuristics used to overcome the N x N complexity required to match all + pairs of objects to identify the equivalent ones. + The challenge is to identify a clustering function that maximizes the chance of comparing only + records that may lead to a match, while minimizing the number of records that will not be + matched while being equivalent. + Since the equivalence function is to some level tolerant to minimal errors (e.g. switching of + characters in the title, or minimal difference in letters), we need this function to be not + too + precise (e.g. a hash of the title), but also not too flexible (e.g. random ngrams of the + title). + On the other hand, reality tells us that in some cases equality of two records can only be + determined by their PIDs (e.g. DOI) as the metadata properties are very different across + different versions and no clustering function will ever bring them into the same cluster. To match these requirements OpenAIRE clustering for products works with two functions:
          • -
            DOI: the function generates the DOI when this is provided as part of the record properties;
            +
            DOI: the function generates the DOI when this is provided as part of the record + properties; +
          • Title-based function: the function generates a key that depends on (i) number of significant words in the title (normalized, stemming, etc.), (ii) module 10 of the number of characters of such words, and - (iii) a string obtained as an alternation of the function prefix(3) and suffix(3) (and vice versa) o the first 3 words (2 words if the title only has 2). For example, the title “Entity deduplication in big data graphs for scholarly communication” becomes “entity deduplication big data graphs scholarly communication” with two keys key “7.1entionbig” and “7.1itydedbig” (where 1 is module 10 of 54 characters of the normalized title. + (iii) a string obtained as an alternation of the function prefix(3) and suffix(3) (and + vice + versa) o the first 3 words (2 words if the title only has 2). For example, the title + “Entity + deduplication in big data graphs for scholarly communication” becomes “entity + deduplication + big data graphs scholarly communication” with two keys key “7.1entionbig” and + “7.1itydedbig” + (where 1 is module 10 of 54 characters of the normalized title.
          - To give an idea, this configuration generates around 77Mi blocks, which we limited to 200 records each (only 15K blocks are affected by the cut), and entails 260Bi matches. Matches in a block are performed using a “sliding window” set to 80 records. The records are sorted lexicographically on a normalized version of their titles. The 1st record is matched against all the 80 following ones, then the second, etc. for an NlogN complexity. + To give an idea, this configuration generates around 77Mi blocks, which we limited to 200 + records each (only 15K blocks are affected by the cut), and entails 260Bi matches. Matches in + a + block are performed using a “sliding window” set to 80 records. The records are sorted + lexicographically on a normalized version of their titles. The 1st record is matched against + all + the 80 following ones, then the second, etc. for an NlogN complexity.
          @@ -192,7 +253,8 @@
        • Deduplication -
          +
          Once the clusters have been built, the algorithm proceeds with the comparisons. @@ -201,7 +263,9 @@
          • - 1. Tries to capture equivalence via PIDs: if records share a PID then they are equivalent + 1. Tries to capture equivalence via PIDs: if records + share + a PID then they are equivalent
          • @@ -212,7 +276,8 @@
          • a. - If record titles contain different “numbers” then they are different (this rule is subject to different feelings, and should be fine-tuned); + If record titles contain different “numbers” then they are different (this rule is + subject to different feelings, and should be fine-tuned);
          • @@ -224,7 +289,9 @@
          • c. - Note that different PIDs do not imply different records, as different versions may have different PIDs. + Note that different PIDs do not imply different records, as different versions may + have + different PIDs.
          @@ -234,22 +301,35 @@
          • - The titles of the two records are normalised and compared for similarity by applying the Levenstein distance algorithm. - The algorithm returns a number in the range [0,1], where 0 means “very different” and 1 means “equal”. - If the distance is greater than or equal 0,99 the two records are identified as duplicates. + The titles of the two records are normalised and compared for similarity by applying + the + Levenstein distance algorithm. + The algorithm returns a number in the range [0,1], where 0 means “very different” and + 1 + means “equal”. + If the distance is greater than or equal 0,99 the two records are identified as + duplicates.
          • -
            Dates are not regarded for equivalence matching because different versions of the same records should be merged and may be published on different dates, e.g. pre-print and published version of an article.
            +
            Dates are not regarded for equivalence matching because different versions of the + same records should be merged and may be published on different dates, e.g. pre-print + and published version of an article. +
        - Once the equivalence relationships between pairs of records are set, the groups of equivalent records are obtained (transitive closure, i.e. “mesh”). - From such sets a new representative object is obtained, which inherits all properties from the merged records and keeps track of their provenance. - The ID of the record is obtained by appending the prefix “dedup_” to the MD5 of the first ID (given their lexicographical ordering). - A new, more stable function to generate the ID is under development, which exploits the DOI when one of the records to be merged includes a Crossref or a DataCite record. + Once the equivalence relationships between pairs of records are set, the groups of equivalent + records are obtained (transitive closure, i.e. “mesh”). + From such sets a new representative object is obtained, which inherits all properties from the + merged records and keeps track of their provenance. + The ID of the record is obtained by appending the prefix “dedup_” to the MD5 of the first ID + (given their lexicographical ordering). + A new, more stable function to generate the ID is under development, which exploits the DOI + when + one of the records to be merged includes a Crossref or a DataCite record.
      @@ -264,15 +344,15 @@
    - - - - -
- -
  • -
    - + + + + +
    +
  • +
  • +
    +
    • General
    • @@ -287,13 +367,20 @@ src="assets/graph-assets/about/architecture/enrichment.svg" alt="Enrichment">

      - The aggregation processes are continuously running and apply vocabularies as they are in a given moment of time. - It could be the case that a vocabulary changes after the aggregation of one data source has finished, + The aggregation processes are continuously running and apply vocabularies as they are in a given + moment of time. + It could be the case that a vocabulary changes after the aggregation of one data source has + finished, thus the aggregated content does not reflect the current status of the controlled vocabularies.

      - In addition, the integration of ScholeXplorer and DOIBooost and some enrichment processes applied on the raw - and on the de-duplicated graph may introduce values that do not comply with the current status of the OpenAIRE controlled vocabularies. - For these reasons, we included a final step of cleansing at the end of the workflow materialisation. + In addition, the integration of ScholeXplorer and DOIBooost and some enrichment processes + applied + on the raw + and on the de-duplicated graph may introduce values that do not comply with the current status + of + the OpenAIRE controlled vocabularies. + For these reasons, we included a final step of cleansing at the end of the workflow + materialisation. The output of the final cleansing step is the final version of the OpenAIRE Research Graph.

      @@ -301,60 +388,111 @@
    • Enrichment -
      +
      - The OpenAIRE Research Graph is enriched by links mined by OpenAIRE’s full-text mining algorithms that scan the plaintexts of publications for funding information, references to datasets, software URIs, accession numbers of bioetities, and EPO patent mentions. - Custom mining modules also link research objects to specific research communities, initiatives and infrastructures. - In addition, other inference modules provide content-based document classification, document similarity, citation matching, and author affiliation matching. + The OpenAIRE Research Graph is enriched by links mined by OpenAIRE’s full-text mining + algorithms + that scan the plaintexts of publications for funding information, references to datasets, + software URIs, accession numbers of bioetities, and EPO patent mentions. + Custom mining modules also link research objects to specific research communities, initiatives + and infrastructures. + In addition, other inference modules provide content-based document classification, document + similarity, citation matching, and author affiliation matching.

      Project mining - in OpenAIRE text mines the full-texts of publications in order to extract matches to funding project codes/IDs. + in OpenAIRE text mines the full-texts of publications in order to extract matches to funding + project codes/IDs. The mining algorithm works by utilising (i) the grant identifier, and (ii) the project acronym (if available) of each project. The mining algorithm: - (1) Preprocesses/normalizes the full-texts using several functions, which depend on the characteristics of each funder (i.e., the format of the grant identifiers), such as stopword and/or punctuation removal, tokenization, stemming, converting to lowercase; then - (2) String matching of grant identifiers against the normalized text is done using database techniques; and - (3) The results are validated and cleaned using the context near the match by looking at the context around the matched ID for relevant metadata and positive or negative words/phrases, in order to calculate a confidence value for each publication-->project link. - A confidence threshold is set to optimise high accuracy while minimising false positives, such as matches with page or report numbers, post/zip codes, parts of telephone numbers, DOIs or URLs, accession numbers. - The algorithm also applies rules for disambiguating results, as different funders can share identical project IDs; for example, grant number 633172 could refer to H2020 project EuroMix but also to Australian-funded NHMRC project “Brain activity (EEG) analysis and brain imaging techniques to measure the neurobiological effects of sleep apnea”. + (1) Preprocesses/normalizes the full-texts using several functions, which depend on the + characteristics of each funder (i.e., the format of the grant identifiers), such as stopword + and/or punctuation removal, tokenization, stemming, converting to lowercase; then + (2) String matching of grant identifiers against the normalized text is done using database + techniques; and + (3) The results are validated and cleaned using the context near the match by looking at the + context around the matched ID for relevant metadata and positive or negative words/phrases, in + order to calculate a confidence value for each publication-->project link. + A confidence threshold is set to optimise high accuracy while minimising false positives, such + as matches with page or report numbers, post/zip codes, parts of telephone numbers, DOIs or + URLs, accession numbers. + The algorithm also applies rules for disambiguating results, as different funders can share + identical project IDs; for example, grant number 633172 could refer to H2020 project EuroMix + but + also to Australian-funded NHMRC project “Brain activity (EEG) analysis and brain imaging + techniques to measure the neurobiological effects of sleep apnea”. Project mining works very well and was the first Text & Data Mining (TDM) service of OpenAIRE. - Performance results vary from funder to funder but precision is higher than 98% for all funders and 99.5% for EC projects. - Recall is higher than 95% (99% for EC projects), when projects are properly acknowledged using project/grant IDs. + Performance results vary from funder to funder but precision is higher than 98% for all + funders + and 99.5% for EC projects. + Recall is higher than 95% (99% for EC projects), when projects are properly acknowledged using + project/grant IDs.

      Dataset extraction - runs on publications full-texts as described in “High pass text-filtering for Citation matching”, TPDL 2017[1]. - In particular, we search for citations to datasets using their DOIs, titles and other metadata (i.e., dates, creator names, publishers, etc.). - We extract parts of the text which look like citations and search for datasets using database join and pattern matching techniques. - Based on the experiments described in the paper, precision of the dataset extraction module is 98.5% and recall is 97.4% but it is also probably overestimated since it does not take into account corruptions that may take place during pdf to text extraction. + runs on publications full-texts as described in “High pass text-filtering for Citation + matching”, TPDL 2017[1]. + In particular, we search for citations to datasets using their DOIs, titles and other metadata + (i.e., dates, creator names, publishers, etc.). + We extract parts of the text which look like citations and search for datasets using database + join and pattern matching techniques. + Based on the experiments described in the paper, precision of the dataset extraction module is + 98.5% and recall is 97.4% but it is also probably overestimated since it does not take into + account corruptions that may take place during pdf to text extraction. It is calculated on the extracted full-texts of small samples from PubMed and arXiv.

      Software extraction runs also on parts of the text which look like citations. - We search the citations for links to software in open software repositories, specifically github, sourceforge, bitbucket and the google code archive. - After that, we search for links that are included in Software Heritage (SH, https://www.softwareheritage.org) and return the permanent URL that SH provides for each software project. - We also enrich this content with user names, titles and descriptions of the software projects using web mining techniques. - Since software mining is based on URL matching, our precision is 100% (we return a software link only if we find it in the text and there is no need to disambiguate). + We search the citations for links to software in open software repositories, specifically + github, sourceforge, bitbucket and the google code archive. + After that, we search for links that are included in Software Heritage (SH, + https://www.softwareheritage.org) and return the permanent URL that SH provides for each + software project. + We also enrich this content with user names, titles and descriptions of the software projects + using web mining techniques. + Since software mining is based on URL matching, our precision is 100% (we return a software + link + only if we find it in the text and there is no need to disambiguate). As for recall rate, this is not calculable for this mining task. - Although we apply all the necessary normalizations to the URLs in order to overcome usual issues (e.g., http or https, existence of www or not, lower/upper case), we do not calculate cases where a software is mentioned using its name and not by a link from the supported software repositories. + Although we apply all the necessary normalizations to the URLs in order to overcome usual + issues + (e.g., http or https, existence of www or not, lower/upper case), we do not calculate cases + where a software is mentioned using its name and not by a link from the supported software + repositories.

      - For the extraction of bio-entities, we focus on Protein Data Bank (PDB) entries. + For the extraction of bio-entities, we focus on Protein Data + Bank (PDB) entries. We have downloaded the database with PDB codes and we update it regularly. We search through the whole publication’s full-text for references to PDB codes. - We apply disambiguation rules (e.g., there are PDB codes that are the same as antibody codes or other issues) so that we return valid results. + We apply disambiguation rules (e.g., there are PDB codes that are the same as antibody codes + or + other issues) so that we return valid results. Current precision is 98%. - Although it's risky to mention recall rates since these are usually overestimated, we have calculated a recall rate of 98% using small samples from pubmed publications. - Moreover, our technique is able to identify about 30% more links to proteins than the ones that are tagged in Pubmed xmls. + Although it's risky to mention recall rates since these are usually overestimated, we have + calculated a recall rate of 98% using small samples from pubmed publications. + Moreover, our technique is able to identify about 30% more links to proteins than the ones + that + are tagged in Pubmed xmls.

      - Other text-mining modules include mining for links to EPO patents, or custom mining modules for linking research objects to specific research communities, initiatives and infrastructures, e.g. COVID-19 mining module. - Apart from text-mining modules, OpenAIRE also provides a document classification service that employs analysis of free text stemming from the abstracts of the publications. - The purpose of applying a document classification module is to assign a scientific text one or more predefined content classes. - In OpenAIRE, the currently used taxonomies are arXiv, MeSH (Medical Subject Headings), ACM and DDC (Dewey Decimal Classification, or Dewey Decimal System). + Other text-mining modules include mining for links to EPO + patents, or custom mining modules for linking research objects to specific research + communities, + initiatives and infrastructures, e.g. COVID-19 mining module. + Apart from text-mining modules, OpenAIRE also provides a document classification service that + employs analysis of free text stemming from the abstracts of the publications. + The purpose of applying a document classification module is to assign a scientific text one or + more predefined content classes. + In OpenAIRE, the currently used taxonomies are arXiv, MeSH (Medical Subject Headings), ACM and + DDC (Dewey Decimal Classification, or Dewey Decimal System).


      - [1] Foufoulas, Y., Stamatogiannakis, L., Dimitropoulos, H., & Ioannidis, Y. (2017, September). High-Pass Text Filtering for Citation Matching. - In International Conference on Theory and Practice of Digital Libraries (pp. 355-366). Springer, Cham. + [1] Foufoulas, Y., Stamatogiannakis, L., Dimitropoulos, H., & Ioannidis, Y. (2017, September). + High-Pass Text Filtering for Citation Matching. + In International Conference on Theory and Practice of Digital Libraries (pp. 355-366). + Springer, + Cham.
      @@ -371,48 +509,69 @@ Enrichment
      - The Deduction process (also known as “bulk tagging”) enriches each record with new information that can be derived from the existing property values. + The Deduction process (also known as “bulk tagging”) enriches each record with new information + that + can be derived from the existing property values.

      - As of September 2020, three procedures are in place to relate a research product to a research initiative, infrastructure (RI) or community (RC) based on: + As of September 2020, three procedures are in place to relate a research product to a research + initiative, infrastructure (RI) or community (RC) based on:
      • subjects (2.7M results tagged)
      • Zenodo community (16K results tagged)
      • the data source it comes from (250K results tagged)
      - The list of subjects, Zenodo communities and data sources used to enrich the products are defined by the managers of the community gateway or infrastructure monitoring dashboard associated with the RC/RI. + The list of subjects, Zenodo communities and data sources used to enrich the products are defined + by + the managers of the community gateway or infrastructure monitoring dashboard associated with the + RC/RI.
    • Enrichment -
      +
      - This process “propagates” properties and links from one product to another if between the two there is a “strong” semantic relationship. + This process “propagates” properties and links from one product to another if between the two + there is a “strong” semantic relationship.

      As of September 2020, the following procedures are in place:
      • Propagation of the property “country” to results from institutional repositories: - e.g. publication collected from an institutional repository maintained by an italian university will be enriched with the property “country = IT”. + e.g. publication collected from an institutional repository maintained by an italian + university will be enriched with the property “country = IT”.
      • - Propagation of links to projects: e.g. publication linked to project P “is supplemented by” a dataset D. + Propagation of links to projects: e.g. publication linked to project P “is supplemented + by” + a dataset D. Dataset D will get the link to project P. The relationships considered for this procedure are “isSupplementedBy” and “supplements”.
      • - Propagation of related community/infrastructure/initiative from organizations to products via affiliation relationships: e.g. a publication with an author affiliated with organization O. - The manager of the community gateway C declared that the outputs of O are all relevant for his/her community C. + Propagation of related community/infrastructure/initiative from organizations to products + via affiliation relationships: e.g. a publication with an author affiliated with + organization O. + The manager of the community gateway C declared that the outputs of O are all relevant for + his/her community C. The publication is tagged as relevant for C.
      • - Propagation of related community/infrastructure/initiative to related products: e.g. publication associated to community C is supplemented by a dataset D. + Propagation of related community/infrastructure/initiative to related products: e.g. + publication associated to community C is supplemented by a dataset D. Dataset D will get the association to C. The relationships considered for this procedure are “isSupplementedBy” and “supplements”.
      • - Propagation of ORCID identifiers to related products, if the products have the same authors: e.g. publication has ORCID for its authors and is supplemented by a dataset D. Dataset D has the same authors as the publication. Authors of D are enriched with the ORCIDs available in the publication. + Propagation of ORCID identifiers to related products, if the products have the same + authors: + e.g. publication has ORCID for its authors and is supplemented by a dataset D. Dataset D + has + the same authors as the publication. Authors of D are enriched with the ORCIDs available + in + the publication. The relationships considered for this procedure are “isSupplementedBy” and “supplements”.
      @@ -430,154 +589,172 @@
    - - - - -
    -
  • -
  • -
    - - Post Cleaning -
    -

    - Lorem ipsum... -

    + + + +
    - - - - -
    -
  • -
  • -
    - - Indexing -
    -

    - The final version of the OpenAIRE Research Graph is indexed on a Solr server that is used by the OpenAIRE portals (EXPLORE, CONNECT, PROVIDE) and APIs, the latter adopted by several third-party applications and organizations, such as: -

    -
      -
    • - EOSC - --The OpenAIRE Research Graph APIs and Portals will offer to the EOSC an Open Science Resource Catalogue, keeping an up to date map of all research results (publications, datasets, software), services, organizations, projects, funders in Europe and beyond. -
    • -
    • - DSpace & EPrints - repositories can install the OpenAIRE plugin to expose OpenAIRE compliant metadata records via their OAI-PMH endpoint and offer to researchers the possibility to link their depositions to the funding project, by selecting it from the list of project provided by OpenAIRE -
    • -
    • - EC participant portal (Sygma - System for Grant Management) - uses the OpenAIRE API in the “Continuous Reporting” section. - Sygma automatically fetches from the OpenAIRE Search API the list of publications and datasets in the OpenAIRE Research Graph that are linked to the project. - The user can select the research products from the list and easily compile the continuous reporting data of the project. -
    • -
    +
  • +
  • +
    + + Post Cleaning +
    +

    + Lorem ipsum... +

    +
    + + + +
    - - - - -
  • - -
  • -
    - - Stats Analysis -
    -

    - The OpenAIRE Research Graph is also processed by a pipeline for extracting the statistics and producing the charts for funders, research initiative, infrastructures, and policy makers that you can see on MONITOR. - Based on the information available on the graph, OpenAIRE provides a set of indicators for monitoring the funding and research impact and the uptake of Open Science publishing practices, - such as Open Access publishing of publications and datasets, availability of interlinks between research products, availability of post-print versions in institutional or thematic Open Access repositories, etc. -

    +
  • +
  • +
    + + Indexing +
    +

    + The final version of the OpenAIRE Research Graph is indexed on a Solr server that is used by the + OpenAIRE portals (EXPLORE, CONNECT, PROVIDE) and APIs, the latter adopted by several third-party + applications and organizations, such as: +

    +
      +
    • + EOSC + --The OpenAIRE Research Graph APIs and Portals will offer to the EOSC an Open Science Resource + Catalogue, keeping an up to date map of all research results (publications, datasets, software), + services, organizations, projects, funders in Europe and beyond. +
    • +
    • + DSpace & EPrints + repositories can install the OpenAIRE plugin to expose OpenAIRE compliant metadata records via their + OAI-PMH endpoint and offer to researchers the possibility to link their depositions to the funding + project, by selecting it from the list of project provided by OpenAIRE +
    • +
    • + EC participant portal (Sygma - System for Grant Management) + uses the OpenAIRE API in the “Continuous Reporting” section. + Sygma automatically fetches from the OpenAIRE Search API the list of publications and datasets in + the + OpenAIRE Research Graph that are linked to the project. + The user can select the research products from the list and easily compile the continuous reporting + data of the project. +
    • +
    +
    + + + +
    - - - - -
  • - - -
    - -
    -
    References
    - + +
  • +
    + + Stats Analysis +
    +

    + The OpenAIRE Research Graph is also processed by a pipeline for extracting the statistics and + producing + the charts for funders, research initiative, infrastructures, and policy makers that you can see on + MONITOR. + Based on the information available on the graph, OpenAIRE provides a set of indicators for monitoring + the funding and research impact and the uptake of Open Science publishing practices, + such as Open Access publishing of publications and datasets, availability of interlinks between + research + products, availability of post-print versions in institutional or thematic Open Access repositories, + etc. +

    +
    + + + + +
    +
  • + +
    +
    +
    References
    + +
    -
    -
    +

    Data & Metrics

    Coming soon...

    - - - - - - - - + + + + + + + +
    -
    -

    Infrastructure

    -
    -
    -

    - The OpenAIRE graph operates based on a vast variety of hardware and software. As of December 2019, the hardware infrastructure is the following: -

    +
    +

    Infrastructure

    +
    +
    +

    + The OpenAIRE graph operates based on a vast variety of hardware and software. As of December 2019, the + hardware infrastructure is the following: +

    +
    +
    -
    -
    -

    Team

    -
    -
    -
    -
    - -
    +
    +

    Team

    +
    +
    +
    - -
    -
    diff --git a/src/app/contact/contact.component.html b/src/app/contact/contact.component.html index a10aab9..6fb7bf2 100644 --- a/src/app/contact/contact.component.html +++ b/src/app/contact/contact.component.html @@ -1,38 +1,38 @@ -
    -
    -

    Support

    -
    -

    Do you have a question? Contact us.

    +
    +
    +

    Do you have a question? Contact us.

    -
    - -
    -
    - +
    + +
    +
    + +
    -
    -
    -
    +
    +