<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<!--<?xml-stylesheet type="text/xsl" href="article.xsl"?>-->
<article article-type="research-article" dtd-version="1.2" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="issn">2940-1348</journal-id>
<journal-title-group>
<journal-title>Journal of Computational Literary Studies</journal-title>
</journal-title-group>
<issn pub-type="epub">2940-1348</issn>
<publisher>
<publisher-name>Technische Universit&#228;t Darmstadt</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.48694/jcls.3590</article-id>
<article-categories>
<subj-group>
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Novel Approach for Identification and Linking of Short Quotations in Scholarly Texts and Literary Works</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-0417-4054</contrib-id>
<name>
<surname>Arnold</surname>
<given-names>Frederik</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0003-3271-9653</contrib-id>
<name>
<surname>J&#228;schke</surname>
<given-names>Robert</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
</contrib-group>
<aff id="aff-1"><label>1</label>Berlin School for Library and Information Science, Humboldt-Universit&#228;t zu Berlin <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://ror.org/01hcx6992">ROR</ext-link>, Berlin, Germany</aff>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2023-12-20">
<day>20</day>
<month>12</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>2</volume>
<issue>1</issue>
<fpage>1</fpage>
<lpage>19</lpage>
<history>
<date date-type="received" iso-8601-date="2023-01-31">
<day>31</day>
<month>01</month>
<year>2023</year>
</date>
<date date-type="accepted" iso-8601-date="2023-10-27">
<day>27</day>
<month>10</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright: &#x00A9; 2023 The Author(s)</copyright-statement>
<copyright-year>2023</copyright-year>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>The text of this work is released under the Creative Commons license CC BY 4.0 International. You can find the contract text of the license at <uri xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</uri>. The illustrations are excluded from this license, here the copyright lies with the respective rights holder.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://jcls.io/article/10.48694/jcls.3590/"/>
<abstract>
<p>We present two approaches for the identification and linking of short quotations between scholarly works and literary works: <italic>ProQuo</italic>, a specialized pipeline, and <italic>ProQuoLM</italic>, a more general language model based approach. Our evaluation shows that both approaches outperform a strong baseline and the overall performance is on the same level. We compare the performance of ProQuoLM on texts with and without (page) reference information and find that reference information is not used. Based on our findings, we propose the following steps for future improvements: Further analysis of the influence of a bigger context window for better handling of long distance references and the introduction of positional information of the literary work so that reference information can be (better) utilized.</p>
</abstract>
<kwd-group>
<kwd>quotation linking</kwd>
<kwd>literary works</kwd>
<kwd>scholarly works</kwd>
<kwd>machine learning</kwd>
<kwd>language models</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="S1">
<title>1. Introduction</title>
<p>Scholarly and literary texts do not exists in a vacuum but rather interact in various ways: Literary scholars quote literary works, scholarly works and other sources, to support the reasoning of their interpretations or to build on earlier publications. Although (literary) interpretations usually only concretely refer to certain passages of a (literary) text, they often claim to interpret the entire work. We know little about the inner workings of this skillful selection, the corresponding attentional behavior, as well as the canonization of passages, that, for various reasons, lend themselves to support the interpretation. The intertextual relationships between interpretations and objects of interpretation vary in nature, ranging from relatively vague references to renderings of clearly identifiable passages of text in the interpreter's own words to direct quotations.</p>
<p>Long quotations, that is, quotations of a length of five words or more, can be identified using text reuse detection methods (<xref ref-type="bibr" rid="B3">Arnold and J&#228;schke 2021</xref>). Shorter quotations are a major challenge for reasons we will explain in a moment. They are important, however, either because they apply to particularly weighty words or because they are indicative of references to passages. Other uses include intertextuality research, for example, in the analysis of quotations from Hamlet (<xref ref-type="bibr" rid="B15">Hohl Trillini and Quassdorf 2010</xref>) or Shakespeare in general (<xref ref-type="bibr" rid="B17">Molz 2020</xref>), argument mining in scholarly texts where the context in the literary work is relevant to understand how texts are analysed (<xref ref-type="bibr" rid="B9">Descher and Petraschka 2018</xref>, <xref ref-type="bibr" rid="B28">Winko and Jannidis 2015</xref>), or the identification of key passages, that is, passages that are particularly important to expert readers (<xref ref-type="bibr" rid="B4">Arnold and Fiechter 2022</xref>).</p>
<p>As already mentioned, quotations can be of varying length from single words to whole paragraphs. Bibliographic references, often in footnotes or a dedicated reference section, identify the work a quotation is taken from. Page references, either in footnotes or in parentheses in the running text, are often used to indicate specific pages.<xref ref-type="fn" rid="n1">1</xref> Despite this information, identifying the exact source location of a quotation is a hard task.</p>
<p>Existing tools for the identification of quotations, for example, Copyfind (<xref ref-type="bibr" rid="B5">Bloomfield 2016</xref>), Passim (<xref ref-type="bibr" rid="B24">Smith et al. 2014</xref>), TextMatcher (<xref ref-type="bibr" rid="B21">Reeve 2020</xref>) or Quid (<xref ref-type="bibr" rid="B3">Arnold and J&#228;schke 2021</xref>), are not suitable for unambiguously identifying instances which are shorter than at least a couple of words, as they often rely on text reuse detection methods. For these shorter quotations, especially for quotations consisting of just one word, a number of challenges arise which the tools just mentioned cannot solve. Firstly, short quotations are much more likely to have multiple possible sources in the literary work, which makes it more difficult to link a quotation to its source. Secondly, quotations from other sources, for example, other scholarly works or quotations from the Bible, are much more likely to also occur in the literary work just by chance.</p>
<p>In this paper, we present and compare two tools for the identification and linking of short quotations between scholarly works and literary works: ProQuo and ProQuoLM. Quotations, long and short, are often accompanied by citation information, for example, page or line numbers, either in the running text in parentheses or in footnotes (<xref ref-type="fig" rid="F1">Figure 1</xref>). Our main idea behind ProQuo is to use the references corresponding to long quotations as examples to distinguish references corresponding to short quotations from other text in parentheses and other references, for example, Bible references or references to other literary works. We then extract relations between short quotations and references and use that information and the position of long quotations as anchors to link short quotations to the literary work.</p>
<fig id="F1">
<caption>
<p><bold>Figure 1:</bold> Example shows an excerpt of a scholarly work (<xref ref-type="bibr" rid="B22">Schaum 2004</xref>) which quotes from a literary work (excerpt from <xref ref-type="bibr" rid="B11">Droste-H&#252;lshoff 1979</xref>). A single word quotation is shown in <styled-content style="color: #4bc479;">green</styled-content>, a long quote in <styled-content style="color: #0303fc;">dark blue</styled-content> and a (page) reference in <styled-content style="color: #03c8fc;">light blue</styled-content>.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g1.png"/>
</fig>
<p>We compare this specialized pipeline with its explainable steps to a more general, state-of-the-art neural language model approach which we named ProQuoLM. For this second approach, we first extract candidates for short quotations and then use a fine-tuned language model to filter the candidates. The comparison allows us to investigate and illustrate the advantages and disadvantages of a pipeline with explainable steps and a blackbox neural language model approach. This is especially relevant in light of recent discussions about computational approaches in digital humanities (<xref ref-type="bibr" rid="B8">Da 2019</xref>).</p>
<p>This paper is organized as follows: In <xref ref-type="sec" rid="S2">section 2</xref>, we provide an overview on related work. In <xref ref-type="sec" rid="S3">section 3</xref> we describe our approaches. In <xref ref-type="sec" rid="S4">section 4</xref> we present our dataset and experimental setup, followed by <xref ref-type="sec" rid="S5">section 5</xref> where we present the results.</p>
</sec>
<sec id="S2">
<title>2. Related Work</title>
<p>Our task is related to reference extraction and segmentation, quotation detection and quotation attribution.</p>
<p>Existing tools for reference extraction and segmentation (<xref ref-type="bibr" rid="B13">GROBID 2008&#8211;2022</xref>; <xref ref-type="bibr" rid="B20">Prasad et al. 2018</xref>) focus on STEM fields (science, technology, engineering, medicine) where references appear in a dedicated reference section and are referenced in the running text in some form, for example, author-year mentions. The focus is on the identification of these reference sections, linking author-year mentions in the running text to entries in the reference section and the segmentation of references into individual fields, for example, author, title, year etc.</p>
<p>The next related task, quotation detection, aims to identify reported speech, thought and writing in text (<xref ref-type="bibr" rid="B18">Papay and Pad&#243; 2019</xref>; <xref ref-type="bibr" rid="B19">Pareti et al. 2013</xref>; <xref ref-type="bibr" rid="B23">Scheible et al. 2016</xref>). This task is normally constrained to individual texts and a focus on speech. For our task on the other hand, we are interested in the detection of quotations as a type of scholarly citation.</p>
<p>Quotation attribution is the task of identifying the source of a quotation (<xref ref-type="bibr" rid="B1">Almeida et al. 2014</xref>; <xref ref-type="bibr" rid="B12">Elson and McKeown 2010</xref>). Existing approaches are often focused on speaker attribution in fiction or news paper articles. For the task at hand, our goal is different. We want to distinguish between quotations from a given primary literary work and other sources and identify a specific occurrence in the case of multiple occurrences. We aim to combine aspects of these three tasks into the new task of identifying quotations in one text and linking those quotations to their source in another text by using page references.</p>
<p>Arnold and J&#228;schke (<xref ref-type="bibr" rid="B3">2021</xref>) presented Quid: a tool for the identification of text re-use with a focus on quotations with a length of at least five words between literary and scholarly works. Five words is not a hard limit but they determined that shorter quotations generate too many ambiguous matches without more advanced methods. Quid outperformed other approaches which led us to the decision to use it in this work. We will also use Quid for the extraction of candidates for quotations shorter than five words which we then filter further.</p>
</sec>
<sec id="S3">
<title>3. Methods</title>
<p>In this section, we first define the task and then present two approaches to solve it. The first approach is a specialized pipeline and the second approach is a more general approach based on a neural network language model.</p>
<sec id="S3.1">
<title>3.1 Task</title>
<p>Our overall goal is to identify short quotations in the scholarly work and link the quotations to their source text in the literary work. For this task, we make the following assumptions. Firstly, we work with a corpus of scholarly works for which we know that their main focus is on the primary literary work which we are interested in. Secondly, we assume that all quotations appear in quotation marks and that the texts do not contain errors, for instance, due to OCR. Handling texts with such issues is out of scope of this work and there are other efforts solving this task (<xref ref-type="bibr" rid="B7">Brunner et al. 2020</xref>).</p>
<p>We focus on scholarly works with references in parentheses in the running text. This decision was made based on the number of scholarly works in our corpus with references in the running text (<xref ref-type="sec" rid="S4.1">subsection 4.1</xref>) and due to the high variance in structure of references in footnotes.</p>
</sec>
<sec id="S3.2">
<title>3.2 ProQuo</title>
<p><xref ref-type="fig" rid="F2">Figure 2</xref> shows the building blocks of our first approach which we named ProQuo. This approach is divided into three steps: <italic>Reference Identification, Relation Extraction</italic>, and <italic>Quotation Linking</italic>. In the first step, we use long quotations (<styled-content style="color: #0303fc;">dark blue</styled-content>), extracted using Quid, and their references (<styled-content style="color: #964f9f;">pink</styled-content>) as anchors. We use these known references as examples to identify other references to the literary work (<styled-content style="color: #03c8fc;">light blue</styled-content>) and distinguish them from other text in parentheses (strikethrough). In the second step, we then link the identified references to their corresponding short quotations (<styled-content style="color: #4bc479;">green</styled-content>). In the final step, the identified short quotations are linked to their source in the literary work (black dashed arrows).</p>
<fig id="F2">
<caption>
<p><bold>Figure 2:</bold> Visualization of quotation identification and linking in three steps.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g2.png"/>
</fig>
<sec id="S3.2.1">
<title>3.2.1 Step 1: Reference Identification</title>
<p>The goal of this step is to distinguish between true references to the literary work (<xref ref-type="fig" rid="F2">Figure 2</xref>, <styled-content style="color: #03c8fc;">light blue</styled-content>) and other text in parentheses. References are written in a number of ways, as <xref ref-type="table" rid="T1">Table 1</xref> shows.<xref ref-type="fn" rid="n2">2</xref> Often references only contain a page number (Ex. 1, 3) but they can also contain line numbers (Ex. 2) or information on the cited edition (Ex. 4). In this work, we are only interested in page numbers and ignore the other information. To extract the page number from a reference string, we perform the following searches until we get a match:</p>
<list list-type="bullet">
<list-item><p>A number which immediately follows the string &#8220;S.&#8221;;</p></list-item>
<list-item><p>A number which is not preceded by a letter.</p></list-item>
</list>
<table-wrap id="T1">
<caption>
<p><bold>Table 1:</bold> Examples for references.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="top">No.</td>
<td align="left" valign="top">Example</td>
<td align="left" valign="top">Citation Target</td>
</tr>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">(12)</td>
<td align="left" valign="middle" rowspan="7">Literary work</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">(12, 12-14)</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">(S. 12)</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">(HKA V,1. S.12)</td>
</tr>
<tr>
<td align="left" valign="top">5</td>
<td align="left" valign="top">(I, S.12)</td>
</tr>
<tr>
<td align="left" valign="top">6</td>
<td align="left" valign="top">(Jb, 12)</td>
</tr>
<tr>
<td align="left" valign="top">7</td>
<td align="left" valign="top">(SW9 II, 12)</td>
</tr>
<tr>
<td align="left" valign="top">8</td>
<td align="left" valign="top">(1987)</td>
<td align="left" valign="middle" rowspan="3">Other</td>
</tr>
<tr>
<td align="left" valign="top">9</td>
<td align="left" valign="top">(Johannes 8, 11)</td>
</tr>
<tr>
<td align="left" valign="top">10</td>
<td align="left" valign="top">(other text)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>A scholarly work can use any of the variants from <xref ref-type="table" rid="T1">Table 1</xref> to reference the literary work and at the same time use some other variant to point to other (literary) works or even use a similar looking variant to reference the same source in case of citations from collected works. At the same time, we need to distinguish true references from other text which appears in parentheses. This includes dates (Ex. 8), other citations, for example, Bible citations (Ex. 9), and text in general.</p>
<p>To overcome these challenges, we use the following approach. We first identify the best example for a reference to the literary work in the scholarly work. We use quotations longer than five words (<xref ref-type="fig" rid="F2">Figure 2</xref>, <styled-content style="color: #0303fc;">dark blue</styled-content>) to extract up to <italic>n<sub>ref</sub></italic> examples of the type of reference (<xref ref-type="fig" rid="F2">Figure 2</xref>, <styled-content style="color: #964f9f;">pink</styled-content>) for a specific scholarly work. The examples are extracted starting with the longest quotation with a maximum distance of <italic>d<sub>ref</sub></italic> characters between reference and quotation and a maximum reference length of <italic>l<sub>ref</sub></italic>. If less than three examples could be found, we use the one from the longest quotation. Otherwise, all examples are clustered with spectral clustering into two clusters. We use the probability that two references are similar (the model to determine similarity is described below) as the similarity in the affinity matrix for the clustering. From the bigger cluster, we then select the reference example which belongs to the longest quotation. This clustering procedure is necessary to reduce the probability of selecting an incorrect reference example, which could happen in cases where Quid made a mistake or when the long quotation is not followed by a reference but by some other text in parentheses.</p>
<p>To classify whether two references are similar, we trained a twin network (<xref ref-type="bibr" rid="B6">Bromley et al. 1993</xref>) for binary classification. The network is made up of two sub-networks, each a character-level BiLSTM (<xref ref-type="bibr" rid="B14">Hochreiter and Schmidhuber 1997</xref>) on top of an embedding layer. The outputs of the sub-networks are compared using Manhattan distance. Two references are classified as similar if the probability given by the model is over a threshold <italic>t<sub>ref</sub></italic>. Using this model, all text in parentheses is compared against the selected example to distinguish between true references and other text occurring in parentheses.</p>
</sec>
<sec id="S3.2.2">
<title>3.2.2 Step 2: Relation Extraction</title>
<p>The goal of this step is to identify relations between quotations, that is, text in quotation marks, and the references identified in the previous step. First, we extract all quotations and create all possible combinations of quotations and references where the quotation and reference are within a distance of <italic>d<sub>rel</sub></italic> tokens. We determine tokens by white space tokenization. We surround the quotation which we are interested in with a start and end tag, replace the reference text with a special tag and also replace all other references with another special tag. Then, we use a machine learning model to classify each pair as belonging together (<xref ref-type="fig" rid="F2">Figure 2</xref>, solid black arrows) or not (for example, <italic>Quote/Quote</italic> 2 and <italic>p.9</italic> (solid red arrows)). We classify a quotation and reference as belonging together if the probability given by the model is over the threshold <italic>t<sub>rel</sub></italic>. For quotations with multiple reference candidates, we take the relation with the highest probability.</p>
<p>For the classification we compare two machine learning models: a token-level BiLSTM with a classification layer with sigmoid activation and a fine-tuned German uncased BERT model<xref ref-type="fn" rid="n3">3</xref> (<xref ref-type="bibr" rid="B10">Devlin et al. 2019</xref>) with a linear layer on top of the pooled output.</p>
</sec>
<sec id="S3.2.3">
<title>3.2.3 Step 3: Quotation Linking</title>
<p>The goal of this step is to link quotations from the scholarly work to their source in the literary work (<xref ref-type="fig" rid="F2">Figure 2</xref>, dashed black arrows) and exclude other possible candidates (dashed red arrows). The main idea is to use long quotations with known links and references as anchors. We then link short quotations relative to these known positions.</p>
<p>Scholarly works cite different editions of the literary work. Since automatic identification of the cited edition is out of the scope of this work, we decided to map all citations to one edition. To achieve this, we estimate a <italic>virtual page</italic> size by using the references from the long quotations:</p>
<disp-formula id="FD1">
<label>(1)</label>
<alternatives>
<mml:math id="Eq001-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>z</mml:mi><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>q</mml:mi><mml:mi>u</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>r</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>q</mml:mi><mml:mi>u</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>r</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M1">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
page\_size = \frac{{last\_quote\_end - first\_quote\_start}}{{last\_page - first\_page}}
\]
\end{document}
</tex-math>
<graphic xlink:href="jcls-3590_arnold-e1.gif"/>
</alternatives>
</disp-formula>
<p>Here <italic>page_size</italic> is the estimated page length (the number of characters) of the literary work, <italic>first_quote_start</italic> and <italic>last_quote_end</italic> are character positions of the first and last quotation in the scholarly work, respectively, and <italic>first_page</italic> and <italic>last_page</italic> are the corresponding page numbers in the literary work, respectively.</p>
<p>Using this virtual page size we can approximate the character position of short quotations in the literary work. It should be noted that short quotations can appear without a reference. We distinguish between short quotations with and without a reference and the approach differs:</p>
<disp-formula id="FD2">
<label>(2)</label>
<alternatives>
<mml:math id="Eq002-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mo>=</mml:mo><mml:mi>q</mml:mi><mml:mi>u</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>r</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi></mml:mrow>
</mml:math>
<tex-math id="M2">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
page\_diff = quote\_page - first\_page
\]
\end{document}
</tex-math>
<graphic xlink:href="jcls-3590_arnold-e2.gif"/>
</alternatives>
</disp-formula>
<disp-formula id="FD3">
<label>(3)</label>
<alternatives>
<mml:math id="Eq003-mml">
<mml:mrow><mml:mi>q</mml:mi><mml:mi>u</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>r</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>q</mml:mi><mml:mi>u</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>p</mml:mi><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mo>&#x005F;</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>z</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow>
</mml:math>
<tex-math id="M3">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
quote\_pos = first\_quote\_start + (page\_diff \times page\_size)
\]
\end{document}
</tex-math>
<graphic xlink:href="jcls-3590_arnold-e3.gif"/>
</alternatives>
</disp-formula>
<p>For <italic>quotations with a reference</italic>, we approximate the character position of the quotation in the literary work by using <xref ref-type="disp-formula" rid="FD2">Equations 2</xref> and <xref ref-type="disp-formula" rid="FD3">3</xref>, where <italic>quote_page</italic> is the page number of the quotation we want to link and <italic>page_diff</italic> is the distance in number of pages between the quotation and the first known page number.</p>
<p>For <italic>quotations without a reference</italic>, we first try to find the closest quotation in the scholarly work from the already linked quotations within a certain distance <italic>d<sub>link</sub></italic>. We use the midpoint of that quotation as the approximate position.</p>
<p>If an approximate position could be determined, we use this position to define a search range <italic>r<sub>link</sub></italic>. For single word quotations, we then first perform exact string matching and if that does not lead to any matches, we perform fuzzy matching. In case of multiple matches, we take the match closest to the approximated quote position.</p>
<p>For longer quotations, we first try to find an exact match in the determined range. If that leads to exactly one match, that match is used. If there are no matches, the whole text is searched. If that does not lead to a single exact match, we use the matches from Quid as candidates. If there is a single candidate in the given range with an overlap of at least <italic>o<sub>link</sub></italic>%, that candidate is used. If there are no matches, the whole text is searched for a single unambiguous result.</p>
<p>If no approximate position could be determined, the whole text is searched for a single exact match and if there are no matches, we perform fuzzy matching and only use a single unambiguous result.</p>
<p>In our corpus, 11 scholarly works cite an edition of <italic>Michael Kohlhaas</italic> in parallel print. These texts were manually identified and this information is passed to the algorithm to adjust the calculations to only count every other page.<xref ref-type="fn" rid="n4">4</xref></p>
</sec>
</sec>
<sec id="S3.3">
<title>3.3 ProQuoLM</title>
<p>Having seen and appreciated the complexity of the aforementioned bespoke approach, we want to analyze how state-of-the-art neural language models can solve the task when it is formulated in a very simple way such that they can be fine-tuned and applied.</p>
<p>For the second approach, we first extract all text in quotation marks and for each quotation we determine all candidates in the literary work. For determining the candidates we use the same (fuzzy) matching approach as in <xref ref-type="sec" rid="S3.2.3">subsubsection 3.2.3</xref>. We then fine-tune the same German uncased BERT model as before for binary classification between a quotation and a candidate, both with a context window. Both text fragments, that is, quotation with context and candidate with context, have a maximum length of <italic>l<sub>lm</sub></italic> tokens each. We also surround the quotation which we are interested in with a start and end tag in both fragments. From all candidates, we select the one with the highest probability over a threshold <italic>t<sub>lm</sub></italic>.</p>
</sec>
</sec>
<sec id="S4">
<title>4. Experiments</title>
<p>In this section, we first give an overview of the dataset and our annotations. We then present the experiments to evaluate both approaches on texts with references in the running text. Finally, we evaluate ProQuoLM on texts with all reference information removed.</p>
<sec id="S4.1">
<title>4.1 Dataset and Annotation</title>
<p>We assess our methods by analyzing two literary texts, <italic>Die Judenbuche</italic> by Annette von Droste-H&#252;lshoff (<xref ref-type="bibr" rid="B11">1979</xref>) and <italic>Michael Kohlhaas</italic> by Heinrich von Kleist (<xref ref-type="bibr" rid="B16">1978</xref>). For each text, our corpus contains 44 and 49 interpretive scholarly articles, respectively, which were previously annotated in the ArguLIT project (<xref ref-type="bibr" rid="B27">Winko 2017&#8211;2020</xref>) using TEI/XML (<xref ref-type="bibr" rid="B25">TEI Consortium, eds. 2022</xref>).<xref ref-type="fn" rid="n5">5</xref> The annotations include quotations of different types, such as those from the primary literary work, other literary works or scholarly works. The original annotations were limited to clearly marked quotations, that is, with quotation marks. In this evaluation, we only focus on <italic>quotations coming from the primary literary work</italic>. The 93 scholarly works use references either in parentheses in the running text or in footnotes. For this work, we focus on scholarly works with references in the running text and ignore footnotes in all experiments, including quotations in footnotes. This decision was made mainly due to the varying structure of quotations in footnotes and to keep the focus on quotations with references in the running text. For <italic>Die Judenbuche</italic>, 24 scholarly works and for <italic>Michael Kohlhaas</italic>, 33 scholarly works have references in the running text.</p>
<p>We extended the original annotations of these scholarly works in two annotation tasks. In the <italic>reference annotation task</italic>, three persons annotated reference strings (<xref ref-type="table" rid="T1">Table 1</xref>) and links between reference strings and quotations. Five of the texts were annotated by all three annotators with F<sub>1</sub>-score inter-annotator agreements between pairs of annotators of 0.88, 0.93, and 0.90. <xref ref-type="table" rid="T2">Table 2</xref> shows statistics for the number of (short) quotations from the primary literary work with and without references. We also show the number of quotations in footnotes which only account for around 10 % of short quotations.</p>
<table-wrap id="T2">
<caption>
<p><bold>Table 2:</bold> Statistics for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic>.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="top">Literary work</td>
<td align="right">Die Judenbuche</td>
<td align="right">Michael Kohlhaas</td>
</tr>
<tr>
<td align="left" valign="top">All quotations (primary work)</td>
<td align="right">1736</td>
<td align="right">1788</td>
</tr>
<tr>
<td align="left" valign="top">Quotations with a reference</td>
<td align="right">1467</td>
<td align="right">1547</td>
</tr>
<tr>
<td align="left" valign="top">Short quotations</td>
<td align="right">817</td>
<td align="right">862</td>
</tr>
<tr>
<td align="left" valign="top">Short quotations with a reference</td>
<td align="right">672</td>
<td align="right">736</td>
</tr>
<tr>
<td align="left" valign="top">Quotations in footnotes</td>
<td align="right">94</td>
<td align="right">80</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In a <italic>linking annotation task</italic>, two persons annotated the origin of quotations from scholarly texts in the literary text. In this task, not only the literary works with references in the running text were annotated, but for <italic>Die Judenbuche</italic>, all 44 scholarly works were annotated. The additional annotated texts contain 270 short quotations. For consistency, we also ignore footnotes in the additionally annotated texts where references appear in footnotes, effectively resulting in texts without any reference information. This data is used to evaluate the performance of ProQuoLM, which does not rely on explicit reference information, on texts without references. To evaluate inter-annotator agreement, again, the same five texts as before were annotated by both annotators which resulted in an F<sub>1</sub>-score inter-annotator agreement of 0.90.</p>
</sec>
<sec id="S4.2">
<title>4.2 References in Running Text</title>
<p>For the experiments in this section, we perform 5-fold cross validation. We calculate precision and recall following Arnold and J&#228;schke (<xref ref-type="bibr" rid="B3">2021</xref>). We optimized the hyperparameters once on the validation data from the first split of our cross validation and use the hyperparameters for all evaluations.</p>
<sec id="S4.2.1">
<title>4.2.1 Reference Identification</title>
<p>To evaluate the performance of our model, we compare it against a baseline that classifies texts in parentheses as a reference if there is at least one number contained and the text is not longer than the maximum reference length <italic>l<sub>ref</sub></italic>, which we set to 25 characters. This value was chosen as it is in the 99 percentile of lengths in our corpus</p>
<p>The output dimension of the embedding layer and the BiLSTM hidden state are both 32. A dropout of 0.2 is applied. The batch size was set to 512 and the network was trained for 10 epochs with binary crossentropy loss and Adam optimizer with a learning rate of 0.001. The number of examples <italic>n<sub>ref</sub></italic> is set to 5. This worked well in our tests and leaves some room for incorrect examples. For the maximum distance <italic>d<sub>ref</sub></italic>, we determined 20 characters to work well. The inputs are padded/truncated to the maximum reference length. The classification threshold <italic>t<sub>ref</sub></italic> is set to 0.7.</p>
</sec>
<sec id="S4.2.2">
<title>4.2.2 Relation Extraction</title>
<p>To evaluate the performance of our two models, we compare them against three baselines. The first baseline (<italic>Ref After</italic>) always takes the closest reference after the quotation. The second baseline (<italic>Ref Before</italic>) works the same way but takes the closest references before the quotation and the last baseline (<italic>Ref Closest</italic>) takes the closest reference before or after the quotation.</p>
<p>For the BiLSTM model, the output dimension of the embedding layer is 64, the hidden state is 64, and a dropout of 0.3 is applied. The batch size was set to 128 and the network was trained for 5 epochs with binary crossentropy loss and Adam optimizer with a learning rate of 0.01. We use WordPiece embeddings (<xref ref-type="bibr" rid="B29">Wu et al. 2016</xref>) with a 8000 token vocabulary. The classification threshold <italic>t<sub>rel</sub></italic> is set to 0.4. The BERT model was fine-tuned for 3 epochs with a batch size of 12 and a learning rate of 10<sup>-5</sup>. The classification threshold <italic>t<sub>rel</sub></italic> is set to 0.5. The maximum distance <italic>d<sub>rel</sub></italic> between a quotation and reference to still be considered is 100 tokens which is in the 93 percentile of distances in our corpus. We tried to increase the maximum distance but got overall worse results as false positives increased. The input is padded/truncated to a length of 200 tokens.</p>
</sec>
<sec id="S4.2.3">
<title>4.2.3 Quotation Linking</title>
<p>To evaluate the performance of our algorithm, we compare it against a baseline which always links a quotation to the first matching instance.</p>
<p>We determined a search range <italic>r<sub>link</sub></italic> of one page before and after the approximate position to work best. For quotations without a reference, the maximum distance <italic>d<sub>link</sub></italic> is 500 tokens. The minimum candidate overlap <italic>o<sub>link</sub></italic> is 70 %.</p>
</sec>
<sec id="S4.2.4">
<title>4.2.4 The Complete Pipeline and Language Model Approach</title>
<p>In this experiment, we perform two evaluations of our two approaches and compare the results against the same baseline as for the quotation linking task. We first perform the same 5-fold cross validation as before and then a second evaluation where we split the scholarly works by the literary work they interpret and train on the texts from one literary work and evaluate on the other. This is relevant as it indicates how well the approaches can generalize and perform on a completely new literary work.</p>
<p>For ProQuoLM, the model was fine-tuned for 3 epochs with a batch size of 4 and a learning rate of 10<sup>-5</sup>. The classification threshold <italic>t<sub>lm</sub></italic> is set to 0.5 and the maximum length <italic>l<sub>lm</sub></italic> to 200 tokens.</p>
</sec>
</sec>
<sec id="S4.3">
<title>4.3 References in Footnotes</title>
<p>Our second approach ProQuoLM does not rely on explicit reference information for quotations. With this experiment, we investigate whether reference information is needed at all or if our second approach can also handle texts with references in footnotes. We do this by evaluating how well ProQuoLM performs on texts where all reference information is removed including footnotes.</p>
</sec>
</sec>
<sec id="S5">
<title>5. Results</title>
<p>We first present the results for the experiments of the individual steps of ProQuo (<xref ref-type="sec" rid="S5.1">subsection 5.1</xref> to <xref ref-type="sec" rid="S5.3">subsection 5.3</xref>), followed by the results of the complete pipeline ProQuo compared to ProQuoLM (<xref ref-type="sec" rid="S5.4">subsection 5.4</xref>). Finally, we present how ProQuoLM performs on texts without any reference information (<xref ref-type="sec" rid="S5.4">subsection 5.4</xref>).</p>
<sec id="S5.1">
<title>5.1 Reference Extraction</title>
<p><xref ref-type="table" rid="T3">Table 3</xref> shows the results for our baseline and model for reference extraction. Our model outperforms a strong baseline for both literary works. The baseline only misses cases where the reference is not in parentheses or does not contain a number, for example, <italic>ebd.</italic> [ibid.] False positives include dates, Bible quotations, or quotations from other scholarly texts. Our model misses less <italic>ebd.</italic> references but all cases not in parentheses and some other special cases. This includes instances where the reference style differs from all other references, for example, references to a specific verse (<italic>V. 8</italic>) and not a page. Other false negatives include references that consist of two references (<italic>S. 47 und S. 50</italic>) and references which differ from the rest as they are followed by additional information (<italic>Jb, 35, Herv. durch Autor</italic> [author's emphasis]). False positives include instances where numbers appear in parentheses with the same style as true references but are used to structure the text (e. g., in enumerations) or reference other scholarly works.</p>
<table-wrap id="T3">
<caption>
<p><bold>Table 3:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> for reference classification.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="middle" rowspan="2">Approach</td>
<td align="center" colspan="3">Die Judenbuche</td>
<td align="center" colspan="3">Michael Kohlhaas</td>
</tr>
<tr>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
</tr>
<tr>
<td align="left" valign="top">Baseline</td>
<td align="right">0.86</td>
<td align="right">0.95</td>
<td align="right">0.90</td>
<td align="right">0.80</td>
<td align="right"><bold>0.95</bold></td>
<td align="right">0.87</td>
</tr>
<tr>
<td align="left" valign="top">Model</td>
<td align="right"><bold>0.95</bold></td>
<td align="right"><bold>0.96</bold></td>
<td align="right"><bold>0.95</bold></td>
<td align="right"><bold>0.97</bold></td>
<td align="right">0.90</td>
<td align="right"><bold>0.93</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="fig" rid="F3">Figure 3a</xref> and <xref ref-type="fig" rid="F3">Figure 3b</xref> enable a more fine grained analysis.<xref ref-type="fn" rid="n6">6</xref> For <italic>Die Judenbuche</italic>, we can see that our model outperforms or is on par with the baseline for all texts except three. We get similar results for <italic>Michael Kohlhaas</italic>, except that for seven texts the baseline performs better than the model. The results illustrate the importance of our model. Texts for which the baseline struggles often have a high number of quotations with references from sources other than the primary literary work.</p>
<fig id="F3">
<caption>
<p><bold>Figure 3:</bold> F<sub>1</sub>-score comparison for reference extraction.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g3.png"/>
</fig>
</sec>
<sec id="S5.2">
<title>5.2 Relation Extraction</title>
<p><xref ref-type="table" rid="T4">Table 4</xref> shows the results of our two models and three baselines. <italic>Ref Closest</italic> is the best performing baseline with an F<sub>1</sub>-score of 0.65 (<italic>Die Judenbuche</italic>) and 0.75 (<italic>Michael Kohlhaas</italic>). <italic>Ref Closest</italic> has the highest recall but lacks precision. This is to be expected as the baseline does not distinguish between quotations from the primary literary work and quotations from other sources. The poor performance of <italic>Ref Before</italic> confirms that references typically follow a citation.</p>
<p>The LSTM-based model outperforms all three baselines. The BERT model performs best overall but worse for <italic>Die Judenbuche</italic> than for <italic>Michael Kohlhaas</italic>.</p>
<table-wrap id="T4">
<caption>
<p><bold>Table 4:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> for relation extraction.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="middle" rowspan="2">Approach</td>
<td align="center" colspan="3">Die Judenbuche</td>
<td align="center" colspan="3">Michael Kohlhaas</td>
</tr>
<tr>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
</tr>
<tr>
<td align="left" valign="top">Ref After</td>
<td align="right">0.59</td>
<td align="right">0.63</td>
<td align="right">0.61</td>
<td align="right">0.72</td>
<td align="right">0.78</td>
<td align="right">0.75</td>
</tr>
<tr>
<td align="left" valign="top">Ref Before</td>
<td align="right">0.25</td>
<td align="right">0.21</td>
<td align="right">0.23</td>
<td align="right">0.14</td>
<td align="right">0.12</td>
<td align="right">0.13</td>
</tr>
<tr>
<td align="left" valign="top">Ref Closest</td>
<td align="right">0.57</td>
<td align="right"><bold>0.76</bold></td>
<td align="right">0.65</td>
<td align="right">0.66</td>
<td align="right"><bold>0.86</bold></td>
<td align="right">0.75</td>
</tr>
<tr>
<td align="left" valign="top">LSTM</td>
<td align="right">0.83</td>
<td align="right">0.59</td>
<td align="right">0.69</td>
<td align="right">0.85</td>
<td align="right">0.69</td>
<td align="right">0.76</td>
</tr>
<tr>
<td align="left" valign="top">BERT</td>
<td align="right"><bold>0.83</bold></td>
<td align="right">0.68</td>
<td align="right"><bold>0.74</bold></td>
<td align="right"><bold>0.93</bold></td>
<td align="right">0.81</td>
<td align="right"><bold>0.86</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>For <italic>Die Judenbuche</italic>, there are 213 false negatives; 98 of those are the result of long distances, that is, the distance between quotation and references is larger than 100 tokens. Another 67 are instances where the reference appears before the quotation. We get a similar result for <italic>Michael Kohlhaas</italic> with 178 false negatives, 69 long distance and 76 reference before quotation. The instances where the reference appears before the quotation are problematic due to the fact that a reference before a quotation is a lot less likely and our training data is limited in that regard.</p>
<fig id="F4">
<caption>
<p><bold>Figure 4:</bold> F<sub>1</sub>-score comparison for relation extraction.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g4.png"/>
</fig>
<p><xref ref-type="fig" rid="F4">Figure 4a</xref> and <xref ref-type="fig" rid="F4">Figure 4b</xref> show a comparison of the best baseline and the BERT model. These results illustrate the importance of the model for the difficult texts where the difference in performance between the baseline and model is largest. But they also show that the model struggles with some texts. In the case of <italic>Yi 2000</italic>, for example, all false positives are instances where the reference appears before the quotation.</p>
</sec>
<sec id="S5.3">
<title>5.3 Quotation Linking</title>
<p><xref ref-type="table" rid="T5">Table 5</xref> shows the results for the quotation linking step. We compare our algorithm against one baseline (see also <xref ref-type="fig" rid="F5">Figure 5a</xref> and <xref ref-type="fig" rid="F5">Figure 5b</xref>).</p>
<table-wrap id="T5">
<caption>
<p><bold>Table 5:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> for quotation linking.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="middle" rowspan="2">Approach</td>
<td align="center" colspan="3">Die Judenbuche</td>
<td align="center" colspan="3">Michael Kohlhaas</td>
</tr>
<tr>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
</tr>
<tr>
<td align="left" valign="top">Baseline</td>
<td align="right">0.65</td>
<td align="right"><bold>0.77</bold></td>
<td align="right">0.70</td>
<td align="right">0.59</td>
<td align="right"><bold>0.74</bold></td>
<td align="right">0.66</td>
</tr>
<tr>
<td align="left" valign="top">Algorithm</td>
<td align="right"><bold>0.85</bold></td>
<td align="right"><bold>0.77</bold></td>
<td align="right"><bold>0.81</bold></td>
<td align="right"><bold>0.86</bold></td>
<td align="right">0.69</td>
<td align="right"><bold>0.76</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The algorithm outperforms the baseline for both literary works and achieves a high precision. The baseline struggles with texts with a low percentage of quotations from the primary literary work which still appear in the literary work.</p>
<p>Our algorithm generates 158 false negatives for <italic>Die Judenbuche</italic>. 102 of those are single word quotations and 111 have a reference in our annotations. For <italic>Michael Kohlhaas</italic>, we get 271 false negatives of which 180 are single word quotations and 215 have a reference. These results indicate that for further improvements better handling of single word quotations is necessary. The results for <italic>Die Judenbuche</italic> would also indicate that an improvement in the relation extraction step should improve the overall results of the pipeline. At first glance, the overall worse results for <italic>Michael Kohlhaas</italic> in combination with the better results in the relation extraction step do not support this theory. But <italic>Michael Kohlhaas</italic> is roughly twice as long as <italic>Die Judenbuche</italic>, which makes the linking step considerably harder and which could counteract the better relation extraction performance.</p>
<fig id="F5">
<caption>
<p><bold>Figure 5:</bold> F<sub>1</sub>-score comparison for quotation linking.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g5.png"/>
</fig>
</sec>
<sec id="S5.4">
<title>5.4 The Complete Pipeline and Language Model Approach</title>
<p>The results in <xref ref-type="table" rid="T6">Table 6</xref> demonstrate that both approaches &#8211; ProQuo and ProQuoLM &#8211; perform on the same level. Compared to the baseline, the pipeline is a big improvement in precision, but recall is lower than the baseline for both literary works. Overall, ProQuoLM works best, with improvements in recall over ProQuo. ProQuoLM produces 169 false negatives for <italic>Die Judenbuche</italic>, 104 are single word quotations and 133 have a reference in our annotations. Similarly for <italic>Michael Kohlhaas</italic>, the results contain 267 false negatives, 177 are single word quotations and 222 have a reference.</p>
<table-wrap id="T6">
<caption>
<p><bold>Table 6:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> for the full pipeline. For each F<sub>1</sub>-score, the upper and lower bound of the 95 % confidence interval is reported.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="middle" rowspan="2">Approach</td>
<td align="center" colspan="3">Die Judenbuche</td>
<td align="center" colspan="3">Michael Kohlhaas</td>
</tr>
<tr>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
</tr>
<tr>
<td align="left" valign="top">Baseline</td>
<td align="right">0.65</td>
<td align="right"><bold>0.77</bold></td>
<td align="right">0.70 [0.60,0.78]</td>
<td align="right">0.59</td>
<td align="right"><bold>0.74</bold></td>
<td align="right">0.66 [0.56,0.69]</td>
</tr>
<tr>
<td align="left" valign="top">ProQuo</td>
<td align="right">0.87</td>
<td align="right">0.72</td>
<td align="right">0.79 [0.73,0.82]</td>
<td align="right"><bold>0.87</bold></td>
<td align="right">0.66</td>
<td align="right">0.75 [0.69,0.78]</td>
</tr>
<tr>
<td align="left" valign="top">ProQuoLM</td>
<td align="right"><bold>0.88</bold></td>
<td align="right">0.74</td>
<td align="right"><bold>0.80</bold> [0.74,0.86]</td>
<td align="right">0.86</td>
<td align="right">0.69</td>
<td align="right"><bold>0.77</bold> [0.70,0.81]</td>
</tr>
<tr>
<td align="center" colspan="7">Split by literary work</td>
</tr>
<tr>
<td align="left" valign="top">ProQuo</td>
<td align="right">0.87</td>
<td align="right">0.71</td>
<td align="right">0.78 [0.72,0.82]</td>
<td align="right">0.85</td>
<td align="right">0.63</td>
<td align="right">0.72 [0.66,0.76]</td>
</tr>
<tr>
<td align="left" valign="top">ProQuoLM</td>
<td align="right">0.82</td>
<td align="right">0.73</td>
<td align="right">0.77 [0.70,0.82]</td>
<td align="right">0.75</td>
<td align="right">0.70</td>
<td align="right">0.72 [0.65,0.77]</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The second evaluation shows the performance of ProQuo and ProQuoLM for training and evaluation split by literary work. This is relevant as it indicates what the performance will be on a completely new literary work. We can see that the difference in performance is larger when the scholarly works from <italic>Die Judenbuche</italic> are used as training data. This is not surprising as there are less scholarly works for <italic>Die Judenbuche</italic> and therefore less training data in that case.</p>
<p>In <xref ref-type="fig" rid="F6">Figure 6a</xref>&#8211;<xref ref-type="fig" rid="F6">Figure 6d</xref>, we report results broken down by quotation length in words. For both tools, ProQuo (top) and ProQuoLM (bottom), and both literary works, <italic>Die Judenbuche</italic> (left) and <italic>Michael Kohlhaas</italic> (right), single word quotations are the most difficult to identify and link. Similarly, both tools achieve better results for quotations of length three and four. Interestingly, for <italic>Die Judenbuche</italic> and two word quotations, there is a substantial difference in precision between the two approaches. We found that this is due to the term <italic>Die Judenbuche</italic> which ProQuo incorrectly identifies as a quotation in a number of cases. If we exclude these false positives, the precision rises from 0.83 to 0.93.</p>
<fig id="F6">
<caption>
<p><bold>Figure 6:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> by length (in words) of quotations.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="jcls-3590_arnold-g6.png"/>
</fig>
<p>The results show that the performance of both tools is on the same level, but from a usability perspective, ProQuoLM is superior to ProQuo. The approach is less complex, the creation of training data is a lot less time consuming and there is no need for specific handling of parallel print editions.</p>
</sec>
<sec id="S5.5">
<title>5.5 References in Footnotes</title>
<p>In this final experiment, we evaluate the performance of ProQuoLM trained solely on scholarly texts from <italic>Michael Kohlhaas</italic> and tested on scholarly texts from <italic>Die Judenbuche</italic> with references in footnotes. But, as before, we exclude footnotes, effectively resulting in scholarly works without any reference information. We compare ProQuoLM against the same baseline as before.</p>
<p><xref ref-type="table" rid="T7">Table 7</xref> shows that the performance is similar to the other results. This means that even without reference information, ProQuoLM performs on the same level as ProQuo which further highlights its advantages, as it is more versatile. It also leads us to the conclusion that ProQuoLM currently cannot make use of the information contained in references. Considering that there is no information available to the model from where in the literary work a candidate is taken, this is not surprising. Another reason could be that BERT is struggling with capturing numeracy (<xref ref-type="bibr" rid="B26">Wallace et al. 2019</xref>).</p>
<table-wrap id="T7">
<caption>
<p><bold>Table 7:</bold> Precision, recall, and F<sub>1</sub>-score for <italic>Die Judenbuche</italic> for texts with references in footnotes.</p>
</caption>
<table>
<tbody>
<tr>
<td align="left" valign="top">Approach</td>
<td align="right">Precision</td>
<td align="right">Recall</td>
<td align="right">F<sub>1</sub></td>
</tr>
<tr>
<td align="left" valign="top">Baseline</td>
<td align="right">0.52</td>
<td align="right"><bold>0.86</bold></td>
<td align="right">0.65</td>
</tr>
<tr>
<td align="left" valign="top">ProQuoLM</td>
<td align="right"><bold>0.80</bold></td>
<td align="right">0.83</td>
<td align="right"><bold>0.81</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="S6">
<title>6. Discussion</title>
<p>We presented two approaches for the identification and linking of short quotations between scholarly works and literary works. ProQuo is a pipeline consisting of three steps. We evaluated each step individually as well as the complete pipeline. ProQuo outperforms a strong baseline, which lacks precision, especially in cases with quotations from different sources. Our results illustrate that the simple approach of just performing text matching is not sufficient for the task at hand.</p>
<p>The second approach, ProQuoLM, performs on the same level as the pipeline but is superior from a usability perspective as it is less complex, more versatile and the creation of training data is less time consuming. We therefore consider ProQuoLM to be a better starting point for future improvements. However, it should be noted that, depending on the overall goal, ProQuo has the advantage that the idea behind the overall approach and the individual steps can be explained which makes it easier to identify specific issues. The following observations might not have been made without the pipeline and the possibility to investigate individual steps. The development of two approaches is more time and resource consuming but can be beneficial.</p>
<p>From our experiments, we can observe a number of things. Firstly, the distance between a quotation and corresponding reference information can be quite large but our context window is limited due to limitations of current language models. Secondly, the quotation linking step struggles with single word quotations even if they come with a reference. Lastly, ProQuoLM performs on the same level with and without reference information. Based on these observations alone it is not possible to determine the exact source of the remaining issues without further experiments. As a first step, we propose to test ProQuoLM with positional information from where a candidate is taken in the literary work to see if ProQuoLM can make use of reference information at all in the current version. Additionally, it might also be the case that more training would already improve this approach. Also, explicit usage of reference information from the first step of the pipeline in combination with ProQuoLM could be promising but, again, is limited by the fact that reference information can be scattered throughout the text.</p>
<p>Other areas for improvement include the resolution of references which point to other references, for example <italic>ebd.</italic>, and references with multiple page numbers, page ranges or line numbers which are currently not properly handled. We also do not handle quotations with multiple occurrences in the literary work. In the current approach, quotations are never linked to more than one occurrence.</p>
<p>For the presented approaches, we assume a corpus of scholarly works for which we know that the main source of quotations is a certain literary work. Arnold and J&#228;schke (<xref ref-type="bibr" rid="B4">2022</xref>) have found that existing approaches for automatic extraction of bibliographic information do not work for scholarly works in literary studies. This led us to conclude that advances in the extraction of literature references are needed before we can make use of bibliographic information to automatically match scholarly works with the main literary work in focus. Advances in this area would also allow for proper handling of citations from different editions of the literary work.</p>
<p>Another assumption we made for this work is that all quotations appear in quotation marks and that the texts do not contain errors, for instance, due to OCR or mistakes made by the authors. We did not analyze how such errors influence the results as it is beyond the scope of this work. Based on our findings, it seems likely that these errors would have a bigger impact on ProQuo compared to ProQuoLM considering that the former relies more on the availability of specific information. But a deeper analysis is needed to come up with quantifiable results.</p>
</sec>
<sec id="S7">
<title>7. Data Availability</title>
<p>The annotated scholarly works can currently not be made available due to copyright restrictions. All data that can be made available can be found here: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://hu.berlin/proquo-resources">https://hu.berlin/proquo-resources</ext-link> (DOI: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://doi.org/10.5281/zenodo.8232596">https://doi.org/10.5281/zenodo.8232596</ext-link>).</p>
</sec>
<sec id="S8">
<title>8. Software Availability</title>
<p>Software can be found here: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://hu.berlin/proquo">https://hu.berlin/proquo</ext-link> (DOI: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://doi.org/10.5281/zenodo.8221381">https://doi.org/10.5281/zenodo.8221381</ext-link>)</p>
</sec>
</body>
<back>
<sec id="S9">
<title>9. Acknowledgements</title>
<p>Parts of this research were funded by the German Research Foundation (DFG) priority programme (SPP) 2207 <italic>Computational Literary Studies</italic> project <italic>What matters? Key passages in literary works</italic> (grant no. 424207720). We would like to thank the project's student assistants Gregor Sanzenbacher and Nathalie Burkowski and our colleague Benjamin Fiechter for their annotation work as well as Steffen Martus for giving feedback on the manuscript.</p>
</sec>
<sec id="S10">
<title>10. Author Contributions</title>
<p><bold>Frederik Arnold:</bold> Software, Experiments, Conceptualization, Writing &#8211; original draft</p>
<p><bold>Robert J&#228;schke:</bold> Conceptualization, Writing &#8211; original draft</p>
</sec>
<fn-group>
<fn id="n1"><p>In this work, whenever we talk about <italic>references</italic>, we refer to the second type of reference, the one used to indicate specific pages.</p></fn>
<fn id="n2"><p>Examples taken from real texts are shown in the original language. Translations: &#8220;S.&#8221; &#8594; &#8220;p.&#8221;, &#8220;Johannes&#8221; &#8594; &#8220;John&#8221;. Other translations are given in the text in brackets.</p></fn>
<fn id="n3"><p>See: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://huggingface.co/dbmdz/bert-base-german-uncased">https://huggingface.co/dbmdz/bert-base-german-uncased</ext-link>.</p></fn>
<fn id="n4"><p>This is just a very rough approximation. The topic of parallel editions is much more complex and beyond the scope of this work.</p></fn>
<fn id="n5"><p>For the sake of brevity, we will reference <italic>Die Judenbuche</italic> and <italic>Michael Kohlhaas</italic> with J and K, respectively.</p></fn>
<fn id="n6"><p>The horizontal axes are labeled with the first (up to four) letters of the first author&#8217;s name followed by the year of publication. The labels can be used to identify the texts on: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://hu.berlin/quidex">https://hu.berlin/quidex</ext-link>.</p></fn>
</fn-group>
<ref-list>
<ref id="B1"><mixed-citation publication-type="book"><string-name><surname>Almeida</surname>, <given-names>Mariana S. C.</given-names></string-name>, <string-name><given-names>Miguel B.</given-names> <surname>Almeida</surname></string-name>, and <string-name><given-names>Andr&#233; F. T.</given-names> <surname>Martins</surname></string-name> (<year>2014</year>). <chapter-title>&#8220;A Joint Model for Quotation Attribution and Coreference Resolution&#8221;</chapter-title>. In: <source>Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics</source>. <publisher-name>Association for Computational Linguistics</publisher-name>, <fpage>39</fpage>&#8211;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.3115/v1/E14-1005</pub-id>.</mixed-citation></ref>
<ref id="B2"><mixed-citation publication-type="book"><chapter-title>&#8220;Lesen, was wirklich wichtig ist - Die Identifikation von Schl&#252;sselstellen durch ein neues Instrument zur Zitatanalyse&#8221;</chapter-title> In: <source>DHd2022 Kulturen des digitalen Gedächtnisses. 8. Tagung des Verbands Digital Humanities im deutschsprachigen Raum</source> (<year>2022</year>). <pub-id pub-id-type="doi">10.5281/zenodo.6327917</pub-id>.</mixed-citation></ref>
<ref id="B3"><mixed-citation publication-type="webpage"><string-name><surname>Arnold</surname>, <given-names>Frederik</given-names></string-name> and <string-name><given-names>Robert</given-names> <surname>J&#228;schke</surname></string-name> (<year>2021</year>). <article-title>&#8220;Lotte and Annette: A Framework for Finding and Exploring Key Passages in Literary Works&#8221;</article-title>. In: <source>Proceedings of the Workshop on Natural Language Processing for Digital Humanities. NLP Association of India (NLPAI)</source>, <fpage>55</fpage>&#8211;<lpage>63</lpage>. <uri>https://aclanthology.org/2021.nlp4dh-1.7</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B4"><mixed-citation publication-type="webpage"><string-name><surname>Arnold</surname>, <given-names>Frederik</given-names></string-name> and <string-name><given-names>Robert</given-names> <surname>J&#228;schke</surname></string-name> (<year>2022</year>). <chapter-title>&#8220;A Game with Complex Rules: Literature References in Literary Studies&#8221;</chapter-title>. In: <source>Proceedings of the Workshop on Understanding LIterature references in academic full TExt</source>. <publisher-name>CEUR Workshop Proceedings</publisher-name>, <fpage>7</fpage>&#8211;<lpage>15</lpage>. <uri>https://ceur-ws.org/Vol-3220/paper1.pdf</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B5"><mixed-citation publication-type="webpage"><string-name><surname>Bloomfield</surname>, <given-names>Lou</given-names></string-name> (<year>2016</year>). <source>Copyfind</source>. <uri>https://plagiarism.bloomfieldmedia.com/software/copyfind/</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B6"><mixed-citation publication-type="journal"><string-name><surname>Bromley</surname>, <given-names>Jane</given-names></string-name>, <string-name><given-names>Isabelle</given-names> <surname>Guyon</surname></string-name>, <string-name><given-names>Yann</given-names> <surname>LeCun</surname></string-name>, <string-name><given-names>Eduard</given-names> <surname>S&#228;ckinger</surname></string-name>, and <string-name><given-names>Roopak</given-names> <surname>Shah</surname></string-name> (<year>1993</year>). <article-title>&#8220;Signature Verification Using a Siamese Time Delay Neural Network&#8221;</article-title>. In: <source>Advances in Neural Information Processing Systems</source>. Vol. <volume>6</volume>. <uri>https://dl.acm.org/doi/10.5555/2987189.2987282</uri>.</mixed-citation></ref>
<ref id="B7"><mixed-citation publication-type="journal"><string-name><surname>Brunner</surname>, <given-names>Annelen</given-names></string-name>, <string-name><given-names>Ngoc</given-names> <surname>Duyen</surname></string-name>, <string-name><given-names>Tanja</given-names> <surname>Tu</surname></string-name>, <string-name><given-names>Lukas</given-names> <surname>Weimer</surname></string-name>, and <string-name><given-names>Fotis</given-names> <surname>Jannidis</surname></string-name> (<year>2020</year>). <article-title>&#8220;To BERT or not to BERT-Comparing Contextual Embeddings in a Deep Learning Architecture for the Automatic Recognition of four Types of Speech, Thought and Writing Representation.&#8221;</article-title> In: <source>SwissText/KONVENS</source>. <uri>https://ceur-ws.org/Vol-2624/paper5.pdf</uri> (visited on 11/02/2023)</mixed-citation></ref>
<ref id="B8"><mixed-citation publication-type="journal"><string-name><surname>Da</surname>, <given-names>Nan Z</given-names></string-name>. (<year>2019</year>). <article-title>&#8220;The Computational Case against Computational Literary Studies&#8221;</article-title>. In: <source>Critical Inquiry</source> <volume>45</volume> (<issue>3</issue>), <fpage>601</fpage>&#8211;<lpage>639</lpage>. <pub-id pub-id-type="doi">10.1086/702594</pub-id>.</mixed-citation></ref>
<ref id="B9"><mixed-citation publication-type="journal"><string-name><surname>Descher</surname>, <given-names>Stefan</given-names></string-name> and <string-name><given-names>Thomas</given-names> <surname>Petraschka</surname></string-name> (<year>2018</year>). <article-title>&#8220;Die Explizierung des Impliziten&#8221;</article-title>. In: <source>Scientia Poetica</source> <volume>22</volume> (<issue>1</issue>), <fpage>180</fpage>&#8211;<lpage>208</lpage>. <pub-id pub-id-type="doi">10.1515/scipo-2018-007</pub-id>.</mixed-citation></ref>
<ref id="B10"><mixed-citation publication-type="book"><string-name><surname>Devlin</surname>, <given-names>Jacob</given-names></string-name>, <string-name><given-names>Ming-Wei</given-names> <surname>Chang</surname></string-name>, <string-name><given-names>Kenton</given-names> <surname>Lee</surname></string-name>, and <string-name><given-names>Kristina</given-names> <surname>Toutanova</surname></string-name> (<year>2019</year>). <chapter-title>&#8220;BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding&#8221;</chapter-title>. In: <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)</source>. <publisher-name>Association for Computational Linguistics</publisher-name>, <fpage>4171</fpage>&#8211;<lpage>4186</lpage>. <pub-id pub-id-type="doi">10.18653/v1/N19-1423</pub-id>.</mixed-citation></ref>
<ref id="B11"><mixed-citation publication-type="webpage"><string-name><surname>Droste-H&#252;lshoff</surname>, <given-names>Annette von</given-names></string-name> (<year>1979</year>). <source>Die Judenbuche</source>. <publisher-name>Insel Verlag</publisher-name>. <uri>https://www.projekt-gutenberg.org/droste/judenbch/index.html</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B12"><mixed-citation publication-type="book"><string-name><surname>Elson</surname>, <given-names>David K.</given-names></string-name> and <string-name><given-names>Kathleen R.</given-names> <surname>McKeown</surname></string-name> (<year>2010</year>). <chapter-title>&#8220;Automatic Attribution of Quoted Speech in Literary Narrative&#8221;</chapter-title>. In: <source>Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence</source>. AAAI&#8217;10. <publisher-loc>Atlanta, Georgia</publisher-loc>: <publisher-name>AAAI Press</publisher-name>, <fpage>1013</fpage>&#8211;<lpage>1019</lpage>.</mixed-citation></ref>
<ref id="B13"><mixed-citation publication-type="webpage"><source>GROBID</source> (<year>2008&#8211;2022</year>). <uri>https://github.com/kermitt2/grobid</uri>. swh: 1:dir:dab86b296 e3c3216e2241968f0d63b68e8209d3c.</mixed-citation></ref>
<ref id="B14"><mixed-citation publication-type="journal"><string-name><surname>Hochreiter</surname>, <given-names>Sepp</given-names></string-name> and <string-name><given-names>J&#252;rgen</given-names> <surname>Schmidhuber</surname></string-name> (<year>1997</year>). <article-title>&#8220;Long Short-Term Memory&#8221;</article-title>. In: <source>Neural Computation</source> <volume>9</volume> (<issue>8</issue>), <fpage>1735</fpage>&#8211;<lpage>1780</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id>.</mixed-citation></ref>
<ref id="B15"><mixed-citation publication-type="journal"><string-name><surname>Hohl Trillini</surname>, <given-names>Regula</given-names></string-name> and <string-name><given-names>Sixta</given-names> <surname>Quassdorf</surname></string-name> (<year>2010</year>). <article-title>&#8220;A &#8216;key to all quotations&#8217;? A corpusbased parameter model of intertextuality&#8221;</article-title>. In: <source>Literary and Linguistic Computing</source> <volume>25</volume> (<issue>3</issue>), <fpage>269</fpage>&#8211;<lpage>286</lpage>. <pub-id pub-id-type="doi">10.1093/llc/fqq003</pub-id>.</mixed-citation></ref>
<ref id="B16"><mixed-citation publication-type="webpage"><string-name><surname>Kleist</surname>, <given-names>Heinrich von</given-names></string-name> (<year>1978</year>). <chapter-title>&#8220;Michael Kohlhaas&#8221;</chapter-title>. In: <source>Werke und Briefe in vier B&#228;nden</source>. Ed. by <string-name><given-names>Michael</given-names> <surname>Holzinger</surname></string-name>. <publisher-name>CreateSpace Independent Publishing Platform</publisher-name>, <fpage>7</fpage>&#8211;<lpage>113</lpage>. <uri>http://www.zeno.org/nid/2000516902X</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B17"><mixed-citation publication-type="book"><string-name><surname>Molz</surname>, <given-names>Johannes</given-names></string-name> (<year>2020</year>). <source>A Close and Distant Reading of Shakespearean Intertextuality: Towards a Mixed Method Approach for Literary Studies</source>. Open Publishing in the Humanities. <publisher-name>Universit&#228;tsbibliothek Ludwig-Maximilians-Universit&#228;t</publisher-name>. <pub-id pub-id-type="doi">10.5282/oph.4</pub-id>.</mixed-citation></ref>
<ref id="B18"><mixed-citation publication-type="book"><string-name><surname>Papay</surname>, <given-names>Sean</given-names></string-name> and <string-name><given-names>Sebastian</given-names> <surname>Pad&#243;</surname></string-name> (<year>2019</year>). <chapter-title>&#8220;Quotation Detection and Classification with a Corpus-Agnostic Model&#8221;</chapter-title>. In: <source>Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2019)</source>. <publisher-name>INCOMA Ltd.</publisher-name>, <fpage>888</fpage>&#8211;<lpage>894</lpage>. <pub-id pub-id-type="doi">10.26615/978-954-452-056-4_103</pub-id>.</mixed-citation></ref>
<ref id="B19"><mixed-citation publication-type="webpage"><string-name><surname>Pareti</surname>, <given-names>Silvia</given-names></string-name>, <string-name><given-names>Tim</given-names> <surname>O&#8217;Keefe</surname></string-name>, <string-name><given-names>Ioannis</given-names> <surname>Konstas</surname></string-name>, <string-name><given-names>James R.</given-names> <surname>Curran</surname></string-name>, and <string-name><given-names>Irena</given-names> <surname>Koprinska</surname></string-name> (<year>2013</year>). <chapter-title>&#8220;Automatically Detecting and Attributing Indirect Quotations&#8221;</chapter-title>. In: <source>Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing</source>. <publisher-name>Association for Computational Linguistics</publisher-name>, <fpage>989</fpage>&#8211;<lpage>999</lpage>. <uri>https://aclanthology.org/D13-1101</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B20"><mixed-citation publication-type="journal"><string-name><surname>Prasad</surname>, <given-names>Animesh</given-names></string-name>, <string-name><given-names>Manpreet</given-names> <surname>Kaur</surname></string-name>, and <string-name><given-names>Min-Yen</given-names> <surname>Kan</surname></string-name> (<year>2018</year>). <article-title>&#8220;Neural ParsCit: a deep learning-based reference string parser&#8221;</article-title>. In: <source>International Journal on Digital Libraries</source> <volume>19</volume> (<issue>4</issue>), <fpage>323</fpage>&#8211;<lpage>337</lpage>. <pub-id pub-id-type="doi">10.1007/s00799-018-0242-1</pub-id>.</mixed-citation></ref>
<ref id="B21"><mixed-citation publication-type="journal"><string-name><surname>Reeve</surname>, <given-names>Jonathan</given-names></string-name> (<year>2020</year>). <source>JonathanReeve/text-matcher: First Zenodo release</source>. Zenodo. version 0.1.6. <pub-id pub-id-type="doi">10.5281/zenodo.3937738</pub-id>.</mixed-citation></ref>
<ref id="B22"><mixed-citation publication-type="book"><string-name><surname>Schaum</surname>, <given-names>Konrad</given-names></string-name> (<year>2004</year>). <source>Ironie und Ethik in Annette von Droste-H&#252;lshoffs Judenbuche</source>. <publisher-name>Beitr&#228;ge zur neueren Literaturgeschichte</publisher-name>; [Folge 3], Bd. 204. <season>Winter</season>. Chap. Die Judenbuche als Sittengem&#228;lde, <fpage>99</fpage>&#8211;<lpage>194</lpage>.</mixed-citation></ref>
<ref id="B23"><mixed-citation publication-type="book"><string-name><surname>Scheible</surname>, <given-names>Christian</given-names></string-name>, <string-name><given-names>Roman</given-names> <surname>Klinger</surname></string-name>, and <string-name><given-names>Sebastian</given-names> <surname>Pad&#243;</surname></string-name> (<year>2016</year>). <chapter-title>&#8220;Model Architectures for Quotation Detection&#8221;</chapter-title>. In: <source>Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source>. <publisher-name>Association for Computational Linguistics</publisher-name>, <fpage>1736</fpage>&#8211;<lpage>1745</lpage>. <pub-id pub-id-type="doi">10.18653/v1/P16-1164</pub-id>.</mixed-citation></ref>
<ref id="B24"><mixed-citation publication-type="book"><string-name><surname>Smith</surname>, <given-names>David A.</given-names></string-name>, <string-name><given-names>Ryan</given-names> <surname>Cordell</surname></string-name>, <string-name><given-names>Elizabeth Maddock</given-names> <surname>Dillon</surname></string-name>, <string-name><given-names>Nick</given-names> <surname>Stramp</surname></string-name>, and <string-name><given-names>John</given-names> <surname>Wilkerson</surname></string-name> (<year>2014</year>). <chapter-title>&#8220;Detecting and Modeling Local Text Reuse&#8221;</chapter-title>. In: <source>Proceedings of the 14th ACM/IEEE-CS Joint Conference on Digital Libraries</source>. JCDL &#8217;14. <publisher-loc>London, United Kingdom</publisher-loc>: <publisher-name>IEEE Press</publisher-name>, <fpage>183</fpage>&#8211;<lpage>192</lpage>.</mixed-citation></ref>
<ref id="B25"><mixed-citation publication-type="webpage"><collab>TEI Consortium</collab>, eds. (<year>2022</year>). <source>TEI P5: Guidelines for Electronic Text Encoding and Interchange, Version 4.4.0</source>. <uri>https://www.tei-c.org/Guidelines/P5/</uri> (visited on 04/29/2022).</mixed-citation></ref>
<ref id="B26"><mixed-citation publication-type="book"><string-name><surname>Wallace</surname>, <given-names>Eric</given-names></string-name>, <string-name><given-names>Yizhong</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>Sujian</given-names> <surname>Li</surname></string-name>, <string-name><given-names>Sameer</given-names> <surname>Singh</surname></string-name>, and <string-name><given-names>Matt</given-names> <surname>Gardner</surname></string-name> (<year>2019</year>). <chapter-title>&#8220;Do NLP Models Know Numbers? Probing Numeracy in Embeddings&#8221;</chapter-title>. In: <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>. <publisher-name>Association for Computational Linguistics</publisher-name>, <fpage>5307</fpage>&#8211;<lpage>5315</lpage>. <pub-id pub-id-type="doi">10.18653/v1/D19-1534</pub-id>.</mixed-citation></ref>
<ref id="B27"><mixed-citation publication-type="webpage"><string-name><surname>Winko</surname>, <given-names>Simone</given-names></string-name> (<year>2017&#8211;2020</year>). <source>The making of plausibility in interpretive texts. Analyses of argumentative practices in literary studies</source>. DFG-funded research project (grant no. 372804438). <uri>https://gepris.dfg.de/gepris/projekt/372804438?language=en</uri> (visited on 11/02/2023).</mixed-citation></ref>
<ref id="B28"><mixed-citation publication-type="book"><string-name><surname>Winko</surname>, <given-names>Simone</given-names></string-name> and <string-name><given-names>Fotis</given-names> <surname>Jannidis</surname></string-name> (<year>2015</year>). <chapter-title>&#8220;Wissen und Inferenz &#8211; Zum Verstehen und Interpretieren literarischer Texte am Beispiel von Hans Magnus Enzensbergers Gedicht Fr&#252;hschriften&#8221;</chapter-title>. In: <source>Literatur interpretieren: Interdisziplin&#228;re Beitr&#228;ge zur Theorie und Praxis</source>. Ed. by <string-name><given-names>Jan</given-names> <surname>Borkowski</surname></string-name>, <string-name><given-names>Stefan</given-names> <surname>Descher</surname></string-name>, <string-name><given-names>Felicitas</given-names> <surname>Ferder</surname></string-name>, and <string-name><given-names>Philipp David</given-names> <surname>Heine</surname></string-name>. <publisher-name>Brill | mentis</publisher-name>, <fpage>221</fpage>&#8211;<lpage>250</lpage>. <pub-id pub-id-type="doi">10.30965/9783957438973</pub-id>.</mixed-citation></ref>
<ref id="B29"><mixed-citation publication-type="journal"><string-name><surname>Wu</surname>, <given-names>Yonghui</given-names></string-name>, <string-name><given-names>Mike</given-names> <surname>Schuster</surname></string-name>, <string-name><given-names>Zhifeng</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Quoc V.</given-names> <surname>Le</surname></string-name>, <string-name><given-names>Mohammad</given-names> <surname>Norouzi</surname></string-name>, <string-name><given-names>Wolfgang</given-names> <surname>Macherey</surname></string-name>, <string-name><given-names>Maxim</given-names> <surname>Krikun</surname></string-name>, <string-name><given-names>Yuan</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>Qin</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>Klaus</given-names> <surname>Macherey</surname></string-name>, <string-name><given-names>Jeff</given-names> <surname>Klingner</surname></string-name>, <string-name><given-names>Apurva</given-names> <surname>Shah</surname></string-name>, <string-name><given-names>Melvin</given-names> <surname>Johnson</surname></string-name>, <string-name><given-names>Xiaobing</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>&#321;ukasz</given-names> <surname>Kaiser</surname></string-name>, <string-name><given-names>Stephan</given-names> <surname>Gouws</surname></string-name>, <string-name><given-names>Yoshikiyo</given-names> <surname>Kato</surname></string-name>, <string-name><given-names>Taku</given-names> <surname>Kudo</surname></string-name>, <string-name><given-names>Hideto</given-names> <surname>Kazawa</surname></string-name>, <string-name><given-names>Keith</given-names> <surname>Stevens</surname></string-name>, <string-name><given-names>George</given-names> <surname>Kurian</surname></string-name>, <string-name><given-names>Nishant</given-names> <surname>Patil</surname></string-name>, <string-name><given-names>Wei</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>Cliff</given-names> <surname>Young</surname></string-name>, <string-name><given-names>Jason</given-names> <surname>Smith</surname></string-name>, <string-name><given-names>Jason</given-names> <surname>Riesa</surname></string-name>, <string-name><given-names>Alex</given-names> <surname>Rudnick</surname></string-name>, <string-name><given-names>Oriol</given-names> <surname>Vinyals</surname></string-name>, <string-name><given-names>Greg</given-names> <surname>Corrado</surname></string-name>, <string-name><given-names>Macduff</given-names> <surname>Hughes</surname></string-name>, and <string-name><given-names>Jeffrey</given-names> <surname>Dean</surname></string-name> (<year>2016</year>). <source>Google&#8217;s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation</source>. <pub-id pub-id-type="doi">10.48550/ARXIV.1609.08144</pub-id>.</mixed-citation></ref>
</ref-list>
</back>
</article>