forked from D-Net/dnet-hadoop
272 lines
8.3 KiB
Scala
272 lines
8.3 KiB
Scala
package eu.dnetlib.dhp.sx.bio
|
|
|
|
import com.fasterxml.jackson.databind.{DeserializationFeature, ObjectMapper, SerializationFeature}
|
|
import eu.dnetlib.dhp.aggregation.AbstractVocabularyTest
|
|
import eu.dnetlib.dhp.schema.oaf.utils.PidType
|
|
import eu.dnetlib.dhp.schema.oaf.{Oaf, Publication, Relation, Result}
|
|
import eu.dnetlib.dhp.sx.bio.BioDBToOAF.ScholixResolved
|
|
import eu.dnetlib.dhp.sx.bio.pubmed.{PMArticle, PMParser, PubMedToOaf}
|
|
import org.json4s.DefaultFormats
|
|
import org.json4s.JsonAST.{JField, JObject, JString}
|
|
import org.json4s.jackson.JsonMethods.parse
|
|
import org.junit.jupiter.api.Assertions._
|
|
import org.junit.jupiter.api.extension.ExtendWith
|
|
import org.junit.jupiter.api.{BeforeEach, Test}
|
|
import org.mockito.junit.jupiter.MockitoExtension
|
|
|
|
import java.io.{BufferedReader, InputStream, InputStreamReader}
|
|
import java.util.zip.GZIPInputStream
|
|
import scala.collection.JavaConverters._
|
|
import scala.collection.mutable.ListBuffer
|
|
import scala.io.Source
|
|
import scala.xml.pull.XMLEventReader
|
|
|
|
@ExtendWith(Array(classOf[MockitoExtension]))
|
|
class BioScholixTest extends AbstractVocabularyTest {
|
|
|
|
val mapper = new ObjectMapper().enable(SerializationFeature.INDENT_OUTPUT)
|
|
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false)
|
|
|
|
@BeforeEach
|
|
def setUp(): Unit = {
|
|
|
|
super.setUpVocabulary()
|
|
}
|
|
|
|
class BufferedReaderIterator(reader: BufferedReader) extends Iterator[String] {
|
|
override def hasNext() = reader.ready
|
|
override def next() = reader.readLine()
|
|
}
|
|
|
|
object GzFileIterator {
|
|
|
|
def apply(is: InputStream, encoding: String) = {
|
|
new BufferedReaderIterator(
|
|
new BufferedReader(new InputStreamReader(new GZIPInputStream(is), encoding))
|
|
)
|
|
}
|
|
}
|
|
|
|
@Test
|
|
def testEBIData() = {
|
|
val inputXML = Source
|
|
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed.xml"))
|
|
.mkString
|
|
val xml = new XMLEventReader(Source.fromBytes(inputXML.getBytes()))
|
|
new PMParser(xml).foreach(s => println(mapper.writeValueAsString(s)))
|
|
}
|
|
|
|
@Test
|
|
def testPubmedToOaf(): Unit = {
|
|
assertNotNull(vocabularies)
|
|
assertTrue(vocabularies.vocabularyExists("dnet:publication_resource"))
|
|
val records: String = Source
|
|
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed_dump"))
|
|
.mkString
|
|
val r: List[Oaf] = records.lines.toList
|
|
.map(s => mapper.readValue(s, classOf[PMArticle]))
|
|
.map(a => PubMedToOaf.convert(a, vocabularies))
|
|
assertEquals(10, r.size)
|
|
assertTrue(
|
|
r.map(p => p.asInstanceOf[Result])
|
|
.flatMap(p => p.getInstance().asScala.map(i => i.getInstancetype.getClassid))
|
|
.exists(p => "0037".equalsIgnoreCase(p))
|
|
)
|
|
println(mapper.writeValueAsString(r.head))
|
|
|
|
}
|
|
|
|
|
|
private def checkPMArticle(article:PMArticle): Unit = {
|
|
assertNotNull(article.getPmid)
|
|
assertNotNull(article.getTitle)
|
|
assertNotNull(article.getAuthors)
|
|
article.getAuthors.asScala.foreach{a =>
|
|
assertNotNull(a)
|
|
assertNotNull(a.getFullName)
|
|
}
|
|
|
|
}
|
|
|
|
@Test
|
|
def testParsingPubmedXML():Unit = {
|
|
val xml = new XMLEventReader(Source.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed.xml")))
|
|
val parser = new PMParser(xml)
|
|
parser.foreach(checkPMArticle)
|
|
}
|
|
|
|
|
|
private def checkPubmedPublication(o:Oaf): Unit = {
|
|
assertTrue(o.isInstanceOf[Publication])
|
|
val p:Publication = o.asInstanceOf[Publication]
|
|
assertNotNull(p.getId)
|
|
assertNotNull(p.getTitle)
|
|
p.getTitle.asScala.foreach(t =>assertNotNull(t.getValue))
|
|
p.getAuthor.asScala.foreach(a =>assertNotNull(a.getFullname))
|
|
assertNotNull(p.getInstance())
|
|
p.getInstance().asScala.foreach { i =>
|
|
assertNotNull(i.getCollectedfrom)
|
|
assertNotNull(i.getPid)
|
|
assertNotNull(i.getInstancetype)
|
|
}
|
|
assertNotNull(p.getOriginalId)
|
|
p.getOriginalId.asScala.foreach(oId => assertNotNull(oId))
|
|
|
|
|
|
val hasPMC = p.getInstance().asScala.exists(i => i.getPid.asScala.exists(pid => pid.getQualifier.getClassid.equalsIgnoreCase(PidType.pmc.toString)))
|
|
|
|
|
|
|
|
if (hasPMC) {
|
|
assertTrue(p.getOriginalId.asScala.exists(oId => oId.startsWith("od_______267::")))
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
@Test
|
|
def testPubmedMapping() :Unit = {
|
|
|
|
val xml = new XMLEventReader(Source.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pubmed.xml")))
|
|
val parser = new PMParser(xml)
|
|
val results = ListBuffer[Oaf]()
|
|
parser.foreach(x => results += PubMedToOaf.convert(x, vocabularies))
|
|
|
|
|
|
|
|
|
|
results.foreach(checkPubmedPublication)
|
|
|
|
|
|
|
|
}
|
|
|
|
@Test
|
|
def testPDBToOAF(): Unit = {
|
|
|
|
assertNotNull(vocabularies)
|
|
assertTrue(vocabularies.vocabularyExists("dnet:publication_resource"))
|
|
val records: String = Source
|
|
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/pdb_dump"))
|
|
.mkString
|
|
records.lines.foreach(s => assertTrue(s.nonEmpty))
|
|
|
|
val result: List[Oaf] = records.lines.toList.flatMap(o => BioDBToOAF.pdbTOOaf(o))
|
|
|
|
assertTrue(result.nonEmpty)
|
|
result.foreach(r => assertNotNull(r))
|
|
|
|
println(result.count(o => o.isInstanceOf[Relation]))
|
|
println(mapper.writeValueAsString(result.head))
|
|
|
|
}
|
|
|
|
@Test
|
|
def testUNIprotToOAF(): Unit = {
|
|
|
|
assertNotNull(vocabularies)
|
|
assertTrue(vocabularies.vocabularyExists("dnet:publication_resource"))
|
|
|
|
val records: String = Source
|
|
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/uniprot_dump"))
|
|
.mkString
|
|
records.lines.foreach(s => assertTrue(s.nonEmpty))
|
|
|
|
val result: List[Oaf] = records.lines.toList.flatMap(o => BioDBToOAF.uniprotToOAF(o))
|
|
|
|
assertTrue(result.nonEmpty)
|
|
result.foreach(r => assertNotNull(r))
|
|
|
|
println(result.count(o => o.isInstanceOf[Relation]))
|
|
println(mapper.writeValueAsString(result.head))
|
|
|
|
}
|
|
|
|
case class EBILinks(
|
|
relType: String,
|
|
date: String,
|
|
title: String,
|
|
pmid: String,
|
|
targetPid: String,
|
|
targetPidType: String
|
|
) {}
|
|
|
|
def parse_ebi_links(input: String): List[EBILinks] = {
|
|
implicit lazy val formats: DefaultFormats.type = org.json4s.DefaultFormats
|
|
lazy val json = parse(input)
|
|
val pmid = (json \ "publication" \ "pmid").extract[String]
|
|
for {
|
|
JObject(link) <- json \\ "Link"
|
|
JField("Target", JObject(target)) <- link
|
|
JField("RelationshipType", JObject(relType)) <- link
|
|
JField("Name", JString(relation)) <- relType
|
|
JField("PublicationDate", JString(publicationDate)) <- link
|
|
JField("Title", JString(title)) <- target
|
|
JField("Identifier", JObject(identifier)) <- target
|
|
JField("IDScheme", JString(idScheme)) <- identifier
|
|
JField("ID", JString(id)) <- identifier
|
|
|
|
} yield EBILinks(relation, publicationDate, title, pmid, id, idScheme)
|
|
}
|
|
|
|
@Test
|
|
def testCrossrefLinksToOAF(): Unit = {
|
|
|
|
val records: String = Source
|
|
.fromInputStream(getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/crossref_links"))
|
|
.mkString
|
|
records.lines.foreach(s => assertTrue(s.nonEmpty))
|
|
|
|
val result: List[Oaf] = records.lines.map(s => BioDBToOAF.crossrefLinksToOaf(s)).toList
|
|
|
|
assertNotNull(result)
|
|
assertTrue(result.nonEmpty)
|
|
|
|
println(mapper.writeValueAsString(result.head))
|
|
|
|
}
|
|
|
|
@Test
|
|
def testEBILinksToOAF(): Unit = {
|
|
val iterator = GzFileIterator(
|
|
getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/ebi_links.gz"),
|
|
"UTF-8"
|
|
)
|
|
val data = iterator.next()
|
|
|
|
val res = BioDBToOAF
|
|
.parse_ebi_links(BioDBToOAF.extractEBILinksFromDump(data).links)
|
|
.filter(BioDBToOAF.EBITargetLinksFilter)
|
|
.flatMap(BioDBToOAF.convertEBILinksToOaf)
|
|
print(res.length)
|
|
|
|
println(mapper.writeValueAsString(res.head))
|
|
|
|
}
|
|
|
|
@Test
|
|
def scholixResolvedToOAF(): Unit = {
|
|
|
|
val records: String = Source
|
|
.fromInputStream(
|
|
getClass.getResourceAsStream("/eu/dnetlib/dhp/sx/graph/bio/scholix_resolved")
|
|
)
|
|
.mkString
|
|
records.lines.foreach(s => assertTrue(s.nonEmpty))
|
|
|
|
implicit lazy val formats: DefaultFormats.type = org.json4s.DefaultFormats
|
|
|
|
val l: List[ScholixResolved] = records.lines.map { input =>
|
|
lazy val json = parse(input)
|
|
json.extract[ScholixResolved]
|
|
}.toList
|
|
|
|
val result: List[Oaf] = l.map(s => BioDBToOAF.scholixResolvedToOAF(s))
|
|
|
|
assertTrue(result.nonEmpty)
|
|
}
|
|
|
|
}
|