package eu.dnetlib.dedup; import java.io.IOException; import java.io.StringWriter; import java.nio.charset.StandardCharsets; import java.security.MessageDigest; import java.text.Normalizer; import java.util.*; import java.util.stream.Collectors; import org.apache.commons.codec.binary.Hex; import org.apache.commons.io.IOUtils; import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.spark.SparkContext; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.util.LongAccumulator; import com.google.common.collect.Sets; import com.wcohen.ss.JaroWinkler; import eu.dnetlib.dhp.schema.oaf.Author; import eu.dnetlib.dhp.schema.oaf.StructuredProperty; import eu.dnetlib.pace.clustering.BlacklistAwareClusteringCombiner; import eu.dnetlib.pace.config.DedupConfig; import eu.dnetlib.pace.model.MapDocument; import eu.dnetlib.pace.model.Person; import scala.Tuple2; public class DedupUtility { private static final Double THRESHOLD = 0.95; public static Map constructAccumulator( final DedupConfig dedupConf, final SparkContext context) { Map accumulators = new HashMap<>(); String acc1 = String.format("%s::%s", dedupConf.getWf().getEntityType(), "records per hash key = 1"); accumulators.put(acc1, context.longAccumulator(acc1)); String acc2 = String .format( "%s::%s", dedupConf.getWf().getEntityType(), "missing " + dedupConf.getWf().getOrderField()); accumulators.put(acc2, context.longAccumulator(acc2)); String acc3 = String .format( "%s::%s", dedupConf.getWf().getEntityType(), String .format( "Skipped records for count(%s) >= %s", dedupConf.getWf().getOrderField(), dedupConf.getWf().getGroupMaxSize())); accumulators.put(acc3, context.longAccumulator(acc3)); String acc4 = String.format("%s::%s", dedupConf.getWf().getEntityType(), "skip list"); accumulators.put(acc4, context.longAccumulator(acc4)); String acc5 = String.format("%s::%s", dedupConf.getWf().getEntityType(), "dedupSimilarity (x2)"); accumulators.put(acc5, context.longAccumulator(acc5)); String acc6 = String .format( "%s::%s", dedupConf.getWf().getEntityType(), "d < " + dedupConf.getWf().getThreshold()); accumulators.put(acc6, context.longAccumulator(acc6)); return accumulators; } public static JavaRDD loadDataFromHDFS(String path, JavaSparkContext context) { return context.textFile(path); } public static void deleteIfExists(String path) throws IOException { Configuration conf = new Configuration(); FileSystem fileSystem = FileSystem.get(conf); if (fileSystem.exists(new Path(path))) { fileSystem.delete(new Path(path), true); } } public static DedupConfig loadConfigFromHDFS(String path) throws IOException { Configuration conf = new Configuration(); FileSystem fileSystem = FileSystem.get(conf); FSDataInputStream inputStream = new FSDataInputStream(fileSystem.open(new Path(path))); return DedupConfig.load(IOUtils.toString(inputStream, StandardCharsets.UTF_8.name())); } static String readFromClasspath(final String filename, final Class clazz) { final StringWriter sw = new StringWriter(); try { IOUtils.copy(clazz.getResourceAsStream(filename), sw); return sw.toString(); } catch (final IOException e) { throw new RuntimeException("cannot load resource from classpath: " + filename); } } static Set getGroupingKeys(DedupConfig conf, MapDocument doc) { return Sets.newHashSet(BlacklistAwareClusteringCombiner.filterAndCombine(doc, conf)); } public static String md5(final String s) { try { final MessageDigest md = MessageDigest.getInstance("MD5"); md.update(s.getBytes(StandardCharsets.UTF_8)); return new String(Hex.encodeHex(md.digest())); } catch (final Exception e) { System.err.println("Error creating id"); return null; } } public static List mergeAuthor(final List a, final List b) { int pa = countAuthorsPids(a); int pb = countAuthorsPids(b); List base, enrich; int sa = authorsSize(a); int sb = authorsSize(b); if (pa == pb) { base = sa > sb ? a : b; enrich = sa > sb ? b : a; } else { base = pa > pb ? a : b; enrich = pa > pb ? b : a; } enrichPidFromList(base, enrich); return base; } private static void enrichPidFromList(List base, List enrich) { if (base == null || enrich == null) return; final Map basePidAuthorMap = base .stream() .filter(a -> a.getPid() != null && a.getPid().size() > 0) .flatMap(a -> a.getPid().stream().map(p -> new Tuple2<>(p.toComparableString(), a))) .collect(Collectors.toMap(Tuple2::_1, Tuple2::_2, (x1, x2) -> x1)); final List> pidToEnrich = enrich .stream() .filter(a -> a.getPid() != null && a.getPid().size() > 0) .flatMap( a -> a .getPid() .stream() .filter(p -> !basePidAuthorMap.containsKey(p.toComparableString())) .map(p -> new Tuple2<>(p, a))) .collect(Collectors.toList()); pidToEnrich .forEach( a -> { Optional> simAuhtor = base .stream() .map(ba -> new Tuple2<>(sim(ba, a._2()), ba)) .max(Comparator.comparing(Tuple2::_1)); if (simAuhtor.isPresent() && simAuhtor.get()._1() > THRESHOLD) { Author r = simAuhtor.get()._2(); r.getPid().add(a._1()); } }); } public static String createEntityPath(final String basePath, final String entityType) { return String.format("%s/%s", basePath, entityType); } public static String createSimRelPath(final String basePath, final String entityType) { return String.format("%s/%s/simRel", basePath, entityType); } public static String createMergeRelPath(final String basePath, final String entityType) { return String.format("%s/%s/mergeRel", basePath, entityType); } private static Double sim(Author a, Author b) { final Person pa = parse(a); final Person pb = parse(b); if (pa.isAccurate() & pb.isAccurate()) { return new JaroWinkler() .score(normalize(pa.getSurnameString()), normalize(pb.getSurnameString())); } else { return new JaroWinkler() .score(normalize(pa.getNormalisedFullname()), normalize(pb.getNormalisedFullname())); } } private static String normalize(final String s) { return nfd(s) .toLowerCase() // do not compact the regexes in a single expression, would cause StackOverflowError // in case // of large input strings .replaceAll("(\\W)+", " ") .replaceAll("(\\p{InCombiningDiacriticalMarks})+", " ") .replaceAll("(\\p{Punct})+", " ") .replaceAll("(\\d)+", " ") .replaceAll("(\\n)+", " ") .trim(); } private static String nfd(final String s) { return Normalizer.normalize(s, Normalizer.Form.NFD); } private static Person parse(Author author) { if (StringUtils.isNotBlank(author.getSurname())) { return new Person(author.getSurname() + ", " + author.getName(), false); } else { return new Person(author.getFullname(), false); } } private static int countAuthorsPids(List authors) { if (authors == null) return 0; return (int) authors.stream().filter(DedupUtility::hasPid).count(); } private static int authorsSize(List authors) { if (authors == null) return 0; return authors.size(); } private static boolean hasPid(Author a) { if (a == null || a.getPid() == null || a.getPid().size() == 0) return false; return a.getPid().stream().anyMatch(p -> p != null && StringUtils.isNotBlank(p.getValue())); } }