Logo des Repositoriums
 

it - Information Technology 62(2) - April 2020

Autor*innen mit den meisten Dokumenten  

Auflistung nach:

Neueste Veröffentlichungen

1 - 7 von 7
  • Zeitschriftenartikel
    Fast paraphrase extraction in Ancient Greek literature
    (it - Information Technology: Vol. 62, No. 2, 2020) Pöckelmann, Marcus; Dähne, Janis; Ritter, Jörg; Molitor, Paul
    In this paper, A shorter version of the paper appeared in German in the final report of the Digital Plato project which was funded by the Volkswagen Foundation from 2016 to 2019. [35], [28]. we present a method for paraphrase extraction in Ancient Greek that can be applied to huge text corpora in interactive humanities applications. Since lexical databases and POS tagging are either unavailable or do not achieve sufficient accuracy for ancient languages, our approach is based on pure word embeddings and the word mover’s distance (WMD) [20]. We show how to adapt the WMD approach to paraphrase searching such that the expensive WMD computation has to be computed for a small fraction of the text segments contained in the corpus, only. Formally, the time complexity will be reduced from O(N·K3·logK)\mathcal{O}(N\cdot {K^{3}}\cdot \log K) to O(N+K3·logK)\mathcal{O}(N+{K^{3}}\cdot \log K), compared to the brute-force approach which computes the WMD between each text segment of the corpus and the search query. N is the length of the corpus and K the size of its vocabulary. The method, which searches not only for paraphrases of the same length as the search query but also for paraphrases of varying lengths, was evaluated on the Thesaurus Linguae Graecae ® (TLG ® ) [25]. The TLG consists of about 75·10675\cdot {10^{6}} Greek words. We searched the whole TLG for paraphrases for given passages of Plato. The experimental results show that our method and the brute-force approach, with only very few exceptions, propose the same text passages in the TLG as possible paraphrases. The computation times of our method are in a range that allows its application in interactive systems and let the humanities scholars work productively and smoothly.
  • Zeitschriftenartikel
    From giant despair to a new heaven: The early years of automatic collation
    (it - Information Technology: Vol. 62, No. 2, 2020) Nury, Elisa; Spadini, Elena
    This article presents a commented history of automatic collation, from the 1940s until the end of the twentieth century. We look at how the collation was progressively mechanized and automatized with algorithms, and how the issues raised throughout this period carry on into today’s scholarship. In particular, we examine the inner workings of early collation algorithms and their different steps in relation to the formalization of the Gothenburg Model. The scholars working with automatic collation also offer fascinating insights to study the collaborations between Humanists and Computer Scientists, and the reception of computers by philologists.
  • Zeitschriftenartikel
    Digital methods for intertextuality studies
    (it - Information Technology: Vol. 62, No. 2, 2020) Molitor, Paul; Ritter, Jörg
    Article Digital methods for intertextuality studies was published on April 1, 2020 in the journal it - Information Technology (volume 62, issue 2).
  • Zeitschriftenartikel
    Intertextuality and Digital Humanities
    (it - Information Technology: Vol. 62, No. 2, 2020) Schubert, Charlotte
    Proceeding from the debate on intertextuality, some considerations are presented here for Literary and Historical Studies that suggest a theory-driven approach applying algorithm-based procedures. It will be shown that methodical tensions between qualitative and quantitative approaches can be solved simultaneously in this way. On this basis, the approach combines intertextuality theory with an algorithm-based procedure (here a search based on Word Mover’s Distance).
  • Zeitschriftenartikel
    Frontmatter
    (it - Information Technology: Vol. 62, No. 2, 2020) Frontmatter
    Article Frontmatter was published on April 1, 2020 in the journal it - Information Technology (volume 62, issue 2).
  • Zeitschriftenartikel
    The CITE architecture (CTS/CITE) for analysis and alignment
    (it - Information Technology: Vol. 62, No. 2, 2020) Blackwell, Christopher W.; Smith, Neel
    Documenting text-reuse (when one text includes a quotation or paraphrase of, or even allusion to another text) is one example of the problem of analysis and alignment . The most clever analytical tools will be of no avail unless their results can be cited , as scholarly evidence has been cited for centuries. This is where the CITE Architecture can help. CITE solves several problems at once. The first problem is the endless possible number of analyses (by which we mean “desirable ways of splitting up a text”): do we choose to “read” a text passage-by-passage, clause-by-clause, word-by-word, or syllable-by-syllable? The second, related to the first, is that of overlapping hierarchies: The first two words of the Iliad are “μῆνιν ἄειδε,” but the first metrical foot of the poem is “μηνιν α”; the first noun-phrase is “μῆνιν οὐλομένενην”, the first word of the first line, and the first word of the second line, and nothing inin between . All of these issues are present when documenting text-reuse, and especially when documenting different (and perhaps contradictory) scholarly assertions of text-reuse. In our experience, over 25 years of computational textual analysis, no other technological standard can address this problem as easily.
  • Zeitschriftenartikel
    On divergence-based author obfuscation: An attack on the state of the art in statistical authorship verification
    (it - Information Technology: Vol. 62, No. 2, 2020) Bevendorff, Janek; Wenzel, Tobias; Potthast, Martin; Hagen, Matthias; Stein, Benno
    Authorship verification is the task of determining whether two texts were written by the same author based on a writing style analysis. Author obfuscation is the adversarial task of preventing a successful verification by altering a text’s style so that it does not resemble that of its original author anymore. This paper introduces new algorithms for both tasks and reports on a comprehensive evaluation to ascertain the merits of the state of the art in authorship verification to withstand obfuscation. After introducing a new generalization of the well-known unmasking algorithm for short texts, thus completing our collection of state-of-the-art algorithms for verification, we introduce an approach that (1) models writing style difference as the Jensen-Shannon distance between the character n-gram distributions of texts, and (2) manipulates an author’s writing style in a sophisticated manner using heuristic search. For obfuscation, we explore the huge space of textual variants in order to find a paraphrased version of the to-be-obfuscated text that has a sufficiently high Jensen-Shannon distance at minimal costs in terms of text quality loss. We analyze, quantify, and illustrate the rationale of this approach, define paraphrasing operators, derive text length-invariant thresholds for termination, and develop an effective obfuscation framework. Our authorship obfuscation approach defeats the presented state-of-the-art verification approaches, while keeping text changes at a minimum. As a final contribution, we discuss and experimentally evaluate a reverse obfuscation attack against our obfuscation approach as well as possible remedies.