This commit is contained in:
Miriam Baglioni 2023-02-16 15:54:59 +01:00
commit 016337a0f9
13 changed files with 1652 additions and 1585 deletions

View File

@ -13,6 +13,8 @@ import java.util.stream.Collectors;
import java.util.stream.Stream; import java.util.stream.Stream;
import org.apache.commons.lang3.StringUtils; import org.apache.commons.lang3.StringUtils;
import org.apache.spark.api.java.function.MapFunction;
import org.apache.spark.sql.Encoders;
import com.github.sisyphsu.dateparser.DateParserUtils; import com.github.sisyphsu.dateparser.DateParserUtils;
import com.google.common.collect.Lists; import com.google.common.collect.Lists;
@ -23,8 +25,6 @@ import eu.dnetlib.dhp.schema.common.ModelConstants;
import eu.dnetlib.dhp.schema.common.ModelSupport; import eu.dnetlib.dhp.schema.common.ModelSupport;
import eu.dnetlib.dhp.schema.oaf.*; import eu.dnetlib.dhp.schema.oaf.*;
import me.xuender.unidecode.Unidecode; import me.xuender.unidecode.Unidecode;
import org.apache.spark.api.java.function.MapFunction;
import org.apache.spark.sql.Encoders;
public class GraphCleaningFunctions extends CleaningFunctions { public class GraphCleaningFunctions extends CleaningFunctions {

View File

@ -27,7 +27,8 @@ object SparkCreateBaselineDataFrame {
def requestBaseLineUpdatePage(maxFile: String): List[(String, String)] = { def requestBaseLineUpdatePage(maxFile: String): List[(String, String)] = {
val data = requestPage("https://ftp.ncbi.nlm.nih.gov/pubmed/updatefiles/") val data = requestPage("https://ftp.ncbi.nlm.nih.gov/pubmed/updatefiles/")
val result = data.linesWithSeparators.map(l =>l.stripLineEnd) val result = data.linesWithSeparators
.map(l => l.stripLineEnd)
.filter(l => l.startsWith("<a href=")) .filter(l => l.startsWith("<a href="))
.map { l => .map { l =>
val end = l.lastIndexOf("\">") val end = l.lastIndexOf("\">")

View File

@ -63,7 +63,9 @@ class BioScholixTest extends AbstractVocabularyTest {
val records: String = Source val records: String = Source
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed_dump")) .fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed_dump"))
.mkString .mkString
val r: List[Oaf] = records.linesWithSeparators.map(l =>l.stripLineEnd).toList val r: List[Oaf] = records.linesWithSeparators
.map(l => l.stripLineEnd)
.toList
.map(s => mapper.readValue(s, classOf[PMArticle])) .map(s => mapper.readValue(s, classOf[PMArticle]))
.map(a => PubMedToOaf.convert(a, vocabularies)) .map(a => PubMedToOaf.convert(a, vocabularies))
assertEquals(10, r.size) assertEquals(10, r.size)
@ -175,7 +177,8 @@ class BioScholixTest extends AbstractVocabularyTest {
.mkString .mkString
records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty)) records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty))
val result: List[Oaf] = records.linesWithSeparators.map(l =>l.stripLineEnd).toList.flatMap(o => BioDBToOAF.pdbTOOaf(o)) val result: List[Oaf] =
records.linesWithSeparators.map(l => l.stripLineEnd).toList.flatMap(o => BioDBToOAF.pdbTOOaf(o))
assertTrue(result.nonEmpty) assertTrue(result.nonEmpty)
result.foreach(r => assertNotNull(r)) result.foreach(r => assertNotNull(r))
@ -196,7 +199,8 @@ class BioScholixTest extends AbstractVocabularyTest {
.mkString .mkString
records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty)) records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty))
val result: List[Oaf] = records.linesWithSeparators.map(l =>l.stripLineEnd).toList.flatMap(o => BioDBToOAF.uniprotToOAF(o)) val result: List[Oaf] =
records.linesWithSeparators.map(l => l.stripLineEnd).toList.flatMap(o => BioDBToOAF.uniprotToOAF(o))
assertTrue(result.nonEmpty) assertTrue(result.nonEmpty)
result.foreach(r => assertNotNull(r)) result.foreach(r => assertNotNull(r))
@ -241,7 +245,8 @@ class BioScholixTest extends AbstractVocabularyTest {
.mkString .mkString
records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty)) records.linesWithSeparators.map(l => l.stripLineEnd).foreach(s => assertTrue(s.nonEmpty))
val result: List[Oaf] = records.linesWithSeparators.map(l =>l.stripLineEnd).map(s => BioDBToOAF.crossrefLinksToOaf(s)).toList val result: List[Oaf] =
records.linesWithSeparators.map(l => l.stripLineEnd).map(s => BioDBToOAF.crossrefLinksToOaf(s)).toList
assertNotNull(result) assertNotNull(result)
assertTrue(result.nonEmpty) assertTrue(result.nonEmpty)
@ -280,10 +285,13 @@ class BioScholixTest extends AbstractVocabularyTest {
implicit lazy val formats: DefaultFormats.type = org.json4s.DefaultFormats implicit lazy val formats: DefaultFormats.type = org.json4s.DefaultFormats
val l: List[ScholixResolved] = records.linesWithSeparators.map(l =>l.stripLineEnd).map { input => val l: List[ScholixResolved] = records.linesWithSeparators
.map(l => l.stripLineEnd)
.map { input =>
lazy val json = parse(input) lazy val json = parse(input)
json.extract[ScholixResolved] json.extract[ScholixResolved]
}.toList }
.toList
val result: List[Oaf] = l.map(s => BioDBToOAF.scholixResolvedToOAF(s)) val result: List[Oaf] = l.map(s => BioDBToOAF.scholixResolvedToOAF(s))

View File

@ -37,12 +37,24 @@ public class SubscriptionUtils {
} }
public static boolean verifyDateRange(final long date, final String min, final String max) { public static boolean verifyDateRange(final long date, final String min, final String max) {
long from = 0;
long to = Long.MAX_VALUE;
try { try {
return date >= DateUtils.parseDate(min, "yyyy-MM-dd").getTime() from = min != null ? DateUtils.parseDate(min, "yyyy-MM-dd").getTime() : 0;
&& date < DateUtils.parseDate(max, "yyyy-MM-dd").getTime() + ONE_DAY;
} catch (final ParseException e) { } catch (final ParseException e) {
return false; from = 0;
} }
try {
to = max != null ? DateUtils.parseDate(max, "yyyy-MM-dd").getTime() + ONE_DAY : Long.MAX_VALUE;
} catch (final ParseException e) {
to = Long.MAX_VALUE;
}
return date >= from && date < to;
} }
public static boolean verifyExact(final String s1, final String s2) { public static boolean verifyExact(final String s1, final String s2) {

View File

@ -41,6 +41,18 @@ public class SubscriptionUtilsTest {
assertTrue(SubscriptionUtils.verifyDateRange(date, "2010-01-01", "2011-01-01")); assertTrue(SubscriptionUtils.verifyDateRange(date, "2010-01-01", "2011-01-01"));
assertFalse(SubscriptionUtils.verifyDateRange(date, "2020-01-01", "2021-01-01")); assertFalse(SubscriptionUtils.verifyDateRange(date, "2020-01-01", "2021-01-01"));
assertTrue(SubscriptionUtils.verifyDateRange(date, "2010-01-01", "NULL"));
assertTrue(SubscriptionUtils.verifyDateRange(date, "2010-01-01", null));
assertTrue(SubscriptionUtils.verifyDateRange(date, "NULL", "2011-01-01"));
assertTrue(SubscriptionUtils.verifyDateRange(date, null, "2011-01-01"));
assertTrue(SubscriptionUtils.verifyDateRange(date, "NULL", "NULL"));
assertTrue(SubscriptionUtils.verifyDateRange(date, null, null));
assertFalse(SubscriptionUtils.verifyDateRange(date, "2020-01-01", null));
assertFalse(SubscriptionUtils.verifyDateRange(date, "2020-01-01", "NULL"));
assertFalse(SubscriptionUtils.verifyDateRange(date, null, "2005-01-01"));
assertFalse(SubscriptionUtils.verifyDateRange(date, "NULL", "2005-01-01"));
} }
@Test @Test

View File

@ -309,6 +309,8 @@ case object Crossref2Oaf {
result result
} }
def generateAuhtor(given: String, family: String, orcid: String, index: Int): Author = { def generateAuhtor(given: String, family: String, orcid: String, index: Int): Author = {
val a = new Author val a = new Author
a.setName(given) a.setName(given)
@ -370,10 +372,57 @@ case object Crossref2Oaf {
case dataset: Dataset => convertDataset(dataset) case dataset: Dataset => convertDataset(dataset)
} }
val doisReference:List[String] = for {
JObject(reference_json) <- json \ "reference"
JField("DOI", JString(doi_json)) <- reference_json
} yield doi_json
if (doisReference!= null && doisReference.nonEmpty) {
val citation_relations:List[Relation] = generateCitationRelations(doisReference, result)
resultList = resultList ::: citation_relations
}
resultList = resultList ::: List(result) resultList = resultList ::: List(result)
resultList resultList
} }
private def createCiteRelation(source:Result, targetPid:String, targetPidType:String) :List[Relation] = {
val targetId = IdentifierFactory.idFromPid("50",targetPidType, targetPid, true)
val from = new Relation
from.setSource(source.getId)
from.setTarget(targetId)
from.setRelType(ModelConstants.RESULT_RESULT)
from.setRelClass(ModelConstants.CITES)
from.setSubRelType(ModelConstants.CITATION)
from.setCollectedfrom(source.getCollectedfrom)
from.setDataInfo(source.getDataInfo)
from.setLastupdatetimestamp(source.getLastupdatetimestamp)
val to = new Relation
to.setTarget(source.getId)
to.setSource(targetId)
to.setRelType(ModelConstants.RESULT_RESULT)
to.setRelClass(ModelConstants.IS_CITED_BY)
to.setSubRelType(ModelConstants.CITATION)
to.setCollectedfrom(source.getCollectedfrom)
to.setDataInfo(source.getDataInfo)
to.setLastupdatetimestamp(source.getLastupdatetimestamp)
List(from,to)
}
def generateCitationRelations(dois:List[String], result:Result):List[Relation] = {
dois.flatMap(d => createCiteRelation(result, d, "doi"))
}
def mappingFunderToRelations( def mappingFunderToRelations(
funders: List[mappingFunder], funders: List[mappingFunder],
sourceId: String, sourceId: String,

View File

@ -94,7 +94,6 @@
"family": "Stein", "family": "Stein",
"sequence": "first", "sequence": "first",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -104,7 +103,6 @@
"family": "Velzen", "family": "Velzen",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -114,7 +112,6 @@
"family": "Kowalski", "family": "Kowalski",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -122,7 +119,6 @@
"family": "Franckowiak", "family": "Franckowiak",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -132,7 +128,6 @@
"family": "Gezari", "family": "Gezari",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -142,7 +137,6 @@
"family": "Miller-Jones", "family": "Miller-Jones",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -150,7 +144,6 @@
"family": "Frederick", "family": "Frederick",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -160,7 +153,6 @@
"family": "Sfaradi", "family": "Sfaradi",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -168,7 +160,6 @@
"family": "Bietenholz", "family": "Bietenholz",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -178,7 +169,6 @@
"family": "Horesh", "family": "Horesh",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -186,7 +176,6 @@
"family": "Fender", "family": "Fender",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -196,7 +185,6 @@
"family": "Garrappa", "family": "Garrappa",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -206,7 +194,6 @@
"family": "Ahumada", "family": "Ahumada",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -214,7 +201,6 @@
"family": "Andreoni", "family": "Andreoni",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -222,7 +208,6 @@
"family": "Belicki", "family": "Belicki",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -232,7 +217,6 @@
"family": "Bellm", "family": "Bellm",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -240,7 +224,6 @@
"family": "Böttcher", "family": "Böttcher",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -248,7 +231,6 @@
"family": "Brinnel", "family": "Brinnel",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -256,7 +238,6 @@
"family": "Burruss", "family": "Burruss",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -266,7 +247,6 @@
"family": "Cenko", "family": "Cenko",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -276,7 +256,6 @@
"family": "Coughlin", "family": "Coughlin",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -286,7 +265,6 @@
"family": "Cunningham", "family": "Cunningham",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -294,7 +272,6 @@
"family": "Drake", "family": "Drake",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -302,7 +279,6 @@
"family": "Farrar", "family": "Farrar",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -310,7 +286,6 @@
"family": "Feeney", "family": "Feeney",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -318,7 +293,6 @@
"family": "Foley", "family": "Foley",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -328,7 +302,6 @@
"family": "Gal-Yam", "family": "Gal-Yam",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -336,7 +309,6 @@
"family": "Golkhou", "family": "Golkhou",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -346,7 +318,6 @@
"family": "Goobar", "family": "Goobar",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -356,7 +327,6 @@
"family": "Graham", "family": "Graham",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -364,7 +334,6 @@
"family": "Hammerstein", "family": "Hammerstein",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -374,7 +343,6 @@
"family": "Helou", "family": "Helou",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -384,7 +352,6 @@
"family": "Hung", "family": "Hung",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -392,7 +359,6 @@
"family": "Kasliwal", "family": "Kasliwal",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -402,7 +368,6 @@
"family": "Kilpatrick", "family": "Kilpatrick",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -412,7 +377,6 @@
"family": "Kong", "family": "Kong",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -422,7 +386,6 @@
"family": "Kupfer", "family": "Kupfer",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -432,7 +395,6 @@
"family": "Laher", "family": "Laher",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -442,7 +404,6 @@
"family": "Mahabal", "family": "Mahabal",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -452,7 +413,6 @@
"family": "Masci", "family": "Masci",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -462,7 +422,6 @@
"family": "Necker", "family": "Necker",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -472,7 +431,6 @@
"family": "Nordin", "family": "Nordin",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -480,7 +438,6 @@
"family": "Perley", "family": "Perley",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -490,7 +447,6 @@
"family": "Rigault", "family": "Rigault",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -500,7 +456,6 @@
"family": "Reusch", "family": "Reusch",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -508,7 +463,6 @@
"family": "Rodriguez", "family": "Rodriguez",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -518,7 +472,6 @@
"family": "Rojas-Bravo", "family": "Rojas-Bravo",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -528,7 +481,6 @@
"family": "Rusholme", "family": "Rusholme",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -538,7 +490,6 @@
"family": "Shupe", "family": "Shupe",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -548,7 +499,6 @@
"family": "Singer", "family": "Singer",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -558,7 +508,6 @@
"family": "Sollerman", "family": "Sollerman",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -566,7 +515,6 @@
"family": "Soumagnac", "family": "Soumagnac",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -574,7 +522,6 @@
"family": "Stern", "family": "Stern",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -582,7 +529,6 @@
"family": "Taggart", "family": "Taggart",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -590,7 +536,6 @@
"family": "van Santen", "family": "van Santen",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -598,7 +543,6 @@
"family": "Ward", "family": "Ward",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -606,7 +550,6 @@
"family": "Woudt", "family": "Woudt",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
}, },
{ {
@ -616,7 +559,6 @@
"family": "Yao", "family": "Yao",
"sequence": "additional", "sequence": "additional",
"affiliation": [ "affiliation": [
] ]
} }
], ],
@ -1412,7 +1354,6 @@
"Nature Astronomy" "Nature Astronomy"
], ],
"original-title": [ "original-title": [
], ],
"language": "en", "language": "en",
"link": [ "link": [
@ -1448,10 +1389,8 @@
}, },
"score": 1.0, "score": 1.0,
"subtitle": [ "subtitle": [
], ],
"short-title": [ "short-title": [
], ],
"issued": { "issued": {
"date-parts": [ "date-parts": [
@ -1480,7 +1419,6 @@
"URL": "http://dx.doi.org/10.1038/s41550-020-01295-8", "URL": "http://dx.doi.org/10.1038/s41550-020-01295-8",
"relation": { "relation": {
"cites": [ "cites": [
] ]
}, },
"ISSN": [ "ISSN": [

View File

@ -1,9 +1,14 @@
package eu.dnetlib.dhp.doiboost.crossref package eu.dnetlib.dhp.doiboost.crossref
import eu.dnetlib.dhp.schema.common.ModelConstants
import eu.dnetlib.dhp.schema.oaf._ import eu.dnetlib.dhp.schema.oaf._
import eu.dnetlib.dhp.utils.DHPUtils import eu.dnetlib.dhp.utils.DHPUtils
import eu.dnetlib.doiboost.crossref.Crossref2Oaf import eu.dnetlib.doiboost.crossref.Crossref2Oaf
import org.codehaus.jackson.map.{ObjectMapper, SerializationConfig} import org.codehaus.jackson.map.{ObjectMapper, SerializationConfig}
import org.json4s
import org.json4s.JsonAST.{JField, JObject, JString}
import org.json4s.{DefaultFormats, JValue}
import org.json4s.jackson.JsonMethods
import org.junit.jupiter.api.Assertions._ import org.junit.jupiter.api.Assertions._
import org.junit.jupiter.api.Test import org.junit.jupiter.api.Test
import org.slf4j.{Logger, LoggerFactory} import org.slf4j.{Logger, LoggerFactory}
@ -109,6 +114,44 @@ class CrossrefMappingTest {
} }
private def parseJson(input:String):JValue = {
implicit lazy val formats: DefaultFormats.type = org.json4s.DefaultFormats
lazy val json: json4s.JValue = JsonMethods.parse(input)
json
}
@Test
def testCitationRelations():Unit = {
val json = Source.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/doiboost/crossref/publication_license_embargo.json")).mkString
assertNotNull(json)
assertFalse(json.isEmpty)
val result:List[Oaf] = Crossref2Oaf.convert(json)
assertTrue(result.nonEmpty)
val j = parseJson(json)
val doisReference: List[String] = for {
JObject(reference_json) <- j \ "reference"
JField("DOI", JString(doi_json)) <- reference_json
} yield doi_json
val relationList:List[Relation] = result.filter(s => s.isInstanceOf[Relation]).map(r=> r.asInstanceOf[Relation]).filter(r => r.getSubRelType.equalsIgnoreCase(ModelConstants.CITATION))
assertNotNull(relationList)
assertFalse(relationList.isEmpty)
assertEquals(doisReference.size*2, relationList.size)
}
@Test @Test
def testEmptyTitle(): Unit = { def testEmptyTitle(): Unit = {
val json = Source val json = Source

View File

@ -25,7 +25,9 @@ class MappingORCIDToOAFTest {
.mkString .mkString
assertNotNull(json) assertNotNull(json)
assertFalse(json.isEmpty) assertFalse(json.isEmpty)
json.linesWithSeparators.map(l =>l.stripLineEnd).foreach(s => { json.linesWithSeparators
.map(l => l.stripLineEnd)
.foreach(s => {
assertNotNull(ORCIDToOAF.extractValueFromInputString(s)) assertNotNull(ORCIDToOAF.extractValueFromInputString(s))
}) })
} }

View File

@ -5,7 +5,6 @@ import java.io.IOException;
import java.nio.file.Files; import java.nio.file.Files;
import java.nio.file.Path; import java.nio.file.Path;
import eu.dnetlib.dhp.schema.oaf.Dataset;
import org.apache.commons.io.FileUtils; import org.apache.commons.io.FileUtils;
import org.apache.spark.SparkConf; import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaRDD;
@ -27,6 +26,7 @@ import org.slf4j.LoggerFactory;
import com.fasterxml.jackson.databind.ObjectMapper; import com.fasterxml.jackson.databind.ObjectMapper;
import eu.dnetlib.dhp.oa.graph.clean.country.CleanCountrySparkJob; import eu.dnetlib.dhp.oa.graph.clean.country.CleanCountrySparkJob;
import eu.dnetlib.dhp.schema.oaf.Dataset;
import eu.dnetlib.dhp.schema.oaf.Publication; import eu.dnetlib.dhp.schema.oaf.Publication;
public class CleanCountryTest { public class CleanCountryTest {
@ -185,7 +185,6 @@ public class CleanCountryTest {
Assertions.assertEquals(0, tmp.first().getCountry().size()); Assertions.assertEquals(0, tmp.first().getCountry().size());
} }
} }

View File

@ -53,7 +53,8 @@ class ResolveEntitiesTest extends Serializable {
def generateUpdates(spark: SparkSession): Unit = { def generateUpdates(spark: SparkSession): Unit = {
val template = Source.fromInputStream(this.getClass.getResourceAsStream("updates")).mkString val template = Source.fromInputStream(this.getClass.getResourceAsStream("updates")).mkString
val pids: List[String] = template.linesWithSeparators.map(l =>l.stripLineEnd) val pids: List[String] = template.linesWithSeparators
.map(l => l.stripLineEnd)
.map { id => .map { id =>
val r = new Result val r = new Result
r.setId(id.toLowerCase.trim) r.setId(id.toLowerCase.trim)
@ -264,7 +265,8 @@ class ResolveEntitiesTest extends Serializable {
Source Source
.fromInputStream(this.getClass.getResourceAsStream(s"publication")) .fromInputStream(this.getClass.getResourceAsStream(s"publication"))
.mkString .mkString
.linesWithSeparators.map(l =>l.stripLineEnd) .linesWithSeparators
.map(l => l.stripLineEnd)
.next(), .next(),
classOf[Publication] classOf[Publication]
) )

View File

@ -69,7 +69,8 @@ class ScholixGraphTest extends AbstractVocabularyTest {
getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/merge_result_scholix") getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/merge_result_scholix")
) )
.mkString .mkString
val result: List[(Relation, ScholixSummary)] = inputRelations.linesWithSeparators.map(l =>l.stripLineEnd) val result: List[(Relation, ScholixSummary)] = inputRelations.linesWithSeparators
.map(l => l.stripLineEnd)
.sliding(2) .sliding(2)
.map(s => (s.head, s(1))) .map(s => (s.head, s(1)))
.map(p => (mapper.readValue(p._1, classOf[Relation]), mapper.readValue(p._2, classOf[ScholixSummary]))) .map(p => (mapper.readValue(p._1, classOf[Relation]), mapper.readValue(p._2, classOf[ScholixSummary])))