www.researchgate.net Open in urlscan Pro
2606:4700::6811:2169  Public Scan

URL: https://www.researchgate.net/publication/355216002_A_Survey_on_Legal_Question_Answering_Systems
Submission: On February 07 via manual from AT — Scanned from DE

Form analysis 3 forms found in the DOM

GET search

<form method="GET" action="search" class="lite-page__header-search-input-wrapper"><input type="hidden" name="context" readonly="" value="publicSearchHeader"><input placeholder="Search for publications, researchers, or questions" name="q"
    autocomplete="off" class="lite-page__header-search-input"><button
    class="nova-legacy-c-button nova-legacy-c-button--align-center nova-legacy-c-button--radius-full nova-legacy-c-button--size-s nova-legacy-c-button--color-white nova-legacy-c-button--theme-ghost nova-legacy-c-button--width-square lite-page__header-search-button"
    type="submit" width="square"><span class="nova-legacy-c-button__label"><svg aria-hidden="true"
        class="nova-legacy-e-icon nova-legacy-e-icon--size-s nova-legacy-e-icon--theme-bare nova-legacy-e-icon--color-inherit nova-legacy-e-icon--luminosity-medium">
        <use xlink:href="/m/493903042280588/images/icons/nova/icon-stack-s.svg#magnifier-s"></use>
      </svg></span></button></form>

Name: loginFormPOST https://www.researchgate.net/login?_sg=NZSX3-VZCmFiWxDRuWYUvQYlkxGqKRZCl9gV6VaP7angHv4mPlXb3o_rVz8XToi46yu1IvXhQSYlSw

<form method="post" action="https://www.researchgate.net/login?_sg=NZSX3-VZCmFiWxDRuWYUvQYlkxGqKRZCl9gV6VaP7angHv4mPlXb3o_rVz8XToi46yu1IvXhQSYlSw" name="loginForm" id="headerLoginForm"><input type="hidden" name="request_token"
    value="aad-AKCozDfKJJQ8VEgHNaJRCHuLqsxFzKJV/03H0MEFd6XOzxNvIllJfqwNKN8dFRW0J4sq3HAMBHMDlLPgzFKKvyduCQtwDARVs1Bm4eflODzvaq8f6iG+48GK4pRQ+UJQBiRXy3rs++gDoNn9zRqgJ0CRbqPw7i3IWwK1pkwrk89729zWS0QEJPDOtev072FRlIEb/g18leSenyrcwWXrdI3RtYQjvB/34wEmpJ354mO87jmfyUiePkzgm7nS4k6rskIHoBd5lH6Fs4kp5OU="><input
    type="hidden" name="urlAfterLogin" value="publication/355216002_A_Survey_on_Legal_Question_Answering_Systems"><input type="hidden" name="invalidPasswordCount" value="0"><input type="hidden" name="headerLogin" value="yes">
  <div class="lite-page__header-login-item"><label class="nova-legacy-e-text nova-legacy-e-text--size-m nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-inherit lite-page__header-login-label"
      for="input-header-login">Email <div class="lite-page-tooltip "><svg aria-hidden="true" class="nova-legacy-e-icon nova-legacy-e-icon--size-s nova-legacy-e-icon--theme-bare nova-legacy-e-icon--color-inherit nova-legacy-e-icon--luminosity-medium">
          <use xlink:href="/m/493903042280588/images/icons/nova/icon-stack-s.svg#info-circle-s"></use>
        </svg>
        <div class="lite-page-tooltip__content lite-page-tooltip__content--above">
          <div class="nova-legacy-e-text nova-legacy-e-text--size-s nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-inherit"><b>Tip:</b> Most researchers use their institutional email address as their
            ResearchGate login</div>
          <div class="lite-page-tooltip__arrow lite-page-tooltip__arrow--above">
            <div class="lite-page-tooltip__arrow-tip"></div>
          </div>
        </div>
      </div></label></div><input type="email" required="" placeholder="" id="input-header-login" name="login" autocomplete="email" tabindex="1"
    class="nova-legacy-e-input__field nova-legacy-e-input__field--size-m lite-page__header-login-item nova-legacy-e-input__ambient nova-legacy-e-input__ambient--theme-default">
  <div class="lite-page__header-login-item"><label class="lite-page__header-login-label"
      for="input-header-password">Password</label><a class="nova-legacy-e-link nova-legacy-e-link--color-blue nova-legacy-e-link--theme-bare lite-page__header-login-forgot" href="application.LostPassword.html">Forgot password?</a></div><input
    type="password" required="" placeholder="" id="input-header-password" name="password" autocomplete="current-password" tabindex="2"
    class="nova-legacy-e-input__field nova-legacy-e-input__field--size-m lite-page__header-login-item nova-legacy-e-input__ambient nova-legacy-e-input__ambient--theme-default"><label
    class="nova-legacy-e-checkbox lite-page__header-login-checkbox"><input type="checkbox" class="nova-legacy-e-checkbox__input" aria-invalid="false" name="setLoginCookie" tabindex="3" value="yes" checked=""><span
      class="nova-legacy-e-checkbox__checkmark"></span><span class="nova-legacy-e-checkbox__label"> Keep me logged in</span></label>
  <div
    class="nova-legacy-l-flex__item nova-legacy-l-flex nova-legacy-l-flex--gutter-m nova-legacy-l-flex--direction-column@s-up nova-legacy-l-flex--align-items-stretch@s-up nova-legacy-l-flex--justify-content-center@s-up nova-legacy-l-flex--wrap-nowrap@s-up">
    <div class="nova-legacy-l-flex__item"><button
        class="nova-legacy-c-button nova-legacy-c-button--align-center nova-legacy-c-button--radius-m nova-legacy-c-button--size-m nova-legacy-c-button--color-blue nova-legacy-c-button--theme-solid nova-legacy-c-button--width-full" type="submit"
        width="full" tabindex="4"><span class="nova-legacy-c-button__label">Log in</span></button></div>
    <div class="nova-legacy-l-flex__item nova-legacy-l-flex__item--align-self-center@s-up">
      <div class="nova-legacy-e-text nova-legacy-e-text--size-s nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-inherit">or</div>
    </div>
    <div class="nova-legacy-l-flex__item">
      <div
        class="nova-legacy-l-flex__item nova-legacy-l-flex nova-legacy-l-flex--gutter-m nova-legacy-l-flex--direction-column@s-up nova-legacy-l-flex--align-items-center@s-up nova-legacy-l-flex--justify-content-flex-start@s-up nova-legacy-l-flex--wrap-nowrap@s-up">
        <div class="nova-legacy-l-flex__item">
          <a href="connector/google"><div style="display:inline-block;width:247px;height:40px;text-align:left;border-radius:2px;white-space:nowrap;color:#444;background:#4285F4"><span style="margin:1px 0 0 1px;display:inline-block;vertical-align:middle;width:38px;height:38px;background:url('images/socialNetworks/logos-official-2019-05/google-logo.svg') transparent 50% no-repeat"></span><span style="color:#FFF;display:inline-block;vertical-align:middle;padding-left:15px;padding-right:42px;font-size:16px;font-family:Roboto, sans-serif">Continue with Google</span></div></a>
        </div>
      </div>
    </div>
  </div>
</form>

Name: loginFormPOST https://www.researchgate.net/login?_sg=NZSX3-VZCmFiWxDRuWYUvQYlkxGqKRZCl9gV6VaP7angHv4mPlXb3o_rVz8XToi46yu1IvXhQSYlSw

<form method="post" action="https://www.researchgate.net/login?_sg=NZSX3-VZCmFiWxDRuWYUvQYlkxGqKRZCl9gV6VaP7angHv4mPlXb3o_rVz8XToi46yu1IvXhQSYlSw" name="loginForm" id="modalLoginForm"><input type="hidden" name="request_token"
    value="aad-AKCozDfKJJQ8VEgHNaJRCHuLqsxFzKJV/03H0MEFd6XOzxNvIllJfqwNKN8dFRW0J4sq3HAMBHMDlLPgzFKKvyduCQtwDARVs1Bm4eflODzvaq8f6iG+48GK4pRQ+UJQBiRXy3rs++gDoNn9zRqgJ0CRbqPw7i3IWwK1pkwrk89729zWS0QEJPDOtev072FRlIEb/g18leSenyrcwWXrdI3RtYQjvB/34wEmpJ354mO87jmfyUiePkzgm7nS4k6rskIHoBd5lH6Fs4kp5OU="><input
    type="hidden" name="urlAfterLogin" value="publication/355216002_A_Survey_on_Legal_Question_Answering_Systems"><input type="hidden" name="invalidPasswordCount" value="0"><input type="hidden" name="modalLogin" value="yes">
  <div class="nova-legacy-l-form-group nova-legacy-l-form-group--layout-stack nova-legacy-l-form-group--gutter-s">
    <div class="nova-legacy-l-form-group__item nova-legacy-l-form-group__item--width-auto@m-up"><label
        class="nova-legacy-e-text nova-legacy-e-text--size-m nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-xxs nova-legacy-e-text--color-inherit nova-legacy-e-label" for="input-modal-login-label"><span
          class="nova-legacy-e-label__text">Email <div class="lite-page-tooltip "><span class="nova-legacy-e-text nova-legacy-e-text--size-m nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-grey-500">·
              Hint</span>
            <div class="lite-page-tooltip__content lite-page-tooltip__content--above">
              <div class="nova-legacy-e-text nova-legacy-e-text--size-s nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-inherit"><b>Tip:</b> Most researchers use their institutional email address as
                their ResearchGate login</div>
              <div class="lite-page-tooltip__arrow lite-page-tooltip__arrow--above">
                <div class="lite-page-tooltip__arrow-tip"></div>
              </div>
            </div>
          </div></span></label><input type="email" required="" placeholder="Enter your email" id="input-modal-login" name="login" autocomplete="email" tabindex="1"
        class="nova-legacy-e-input__field nova-legacy-e-input__field--size-m nova-legacy-e-input__ambient nova-legacy-e-input__ambient--theme-default"></div>
    <div class="nova-legacy-l-form-group__item nova-legacy-l-form-group__item--width-auto@m-up">
      <div class="lite-page-modal__forgot"><label class="nova-legacy-e-text nova-legacy-e-text--size-m nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-xxs nova-legacy-e-text--color-inherit nova-legacy-e-label"
          for="input-modal-password-label"><span
            class="nova-legacy-e-label__text">Password</span></label><a class="nova-legacy-e-link nova-legacy-e-link--color-blue nova-legacy-e-link--theme-bare lite-page-modal__forgot-link" href="application.LostPassword.html">Forgot password?</a>
      </div><input type="password" required="" placeholder="" id="input-modal-password" name="password" autocomplete="current-password" tabindex="2"
        class="nova-legacy-e-input__field nova-legacy-e-input__field--size-m nova-legacy-e-input__ambient nova-legacy-e-input__ambient--theme-default">
    </div>
    <div class="nova-legacy-l-form-group__item nova-legacy-l-form-group__item--width-auto@m-up"><label class="nova-legacy-e-checkbox"><input type="checkbox" class="nova-legacy-e-checkbox__input" aria-invalid="false" checked="" value="yes"
          name="setLoginCookie" tabindex="3"><span class="nova-legacy-e-checkbox__checkmark"></span><span class="nova-legacy-e-checkbox__label"> Keep me logged in</span></label></div>
    <div class="nova-legacy-l-form-group__item nova-legacy-l-form-group__item--width-auto@m-up"><button
        class="nova-legacy-c-button nova-legacy-c-button--align-center nova-legacy-c-button--radius-m nova-legacy-c-button--size-m nova-legacy-c-button--color-blue nova-legacy-c-button--theme-solid nova-legacy-c-button--width-full" type="submit"
        width="full" tabindex="4"><span class="nova-legacy-c-button__label">Log in</span></button></div>
    <div class="nova-legacy-l-form-group__item nova-legacy-l-form-group__item--width-auto@m-up">
      <div
        class="nova-legacy-l-flex__item nova-legacy-l-flex nova-legacy-l-flex--gutter-m nova-legacy-l-flex--direction-column@s-up nova-legacy-l-flex--align-items-center@s-up nova-legacy-l-flex--justify-content-flex-start@s-up nova-legacy-l-flex--wrap-nowrap@s-up">
        <div class="nova-legacy-l-flex__item">
          <div class="nova-legacy-e-text nova-legacy-e-text--size-s nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-inherit">or</div>
        </div>
        <div class="nova-legacy-l-flex__item">
          <div
            class="nova-legacy-l-flex__item nova-legacy-l-flex nova-legacy-l-flex--gutter-m nova-legacy-l-flex--direction-column@s-up nova-legacy-l-flex--align-items-center@s-up nova-legacy-l-flex--justify-content-flex-start@s-up nova-legacy-l-flex--wrap-nowrap@s-up">
            <div class="nova-legacy-l-flex__item">
              <a href="connector/google"><div style="display:inline-block;width:247px;height:40px;text-align:left;border-radius:2px;white-space:nowrap;color:#444;background:#4285F4"><span style="margin:1px 0 0 1px;display:inline-block;vertical-align:middle;width:38px;height:38px;background:url('images/socialNetworks/logos-official-2019-05/google-logo.svg') transparent 50% no-repeat"></span><span style="color:#FFF;display:inline-block;vertical-align:middle;padding-left:15px;padding-right:42px;font-size:16px;font-family:Roboto, sans-serif">Continue with Google</span></div></a>
            </div>
          </div>
        </div>
        <div class="nova-legacy-l-flex__item">
          <div class="nova-legacy-e-text nova-legacy-e-text--size-s nova-legacy-e-text--family-sans-serif nova-legacy-e-text--spacing-none nova-legacy-e-text--color-grey-500" align="center">No account?
            <a class="nova-legacy-e-link nova-legacy-e-link--color-blue nova-legacy-e-link--theme-decorated" href="signup.SignUp.html?hdrsu=1&amp;_sg%5B0%5D=BiXArZ-UnjfZwdRLTxV6Jgmrrn202CokkpD_A6EmkDOcEQo6L3V_FtUyq92GeZShRczG1Qs8XvceGSzEXf5bYF2Hns0">Sign up</a>
          </div>
        </div>
      </div>
    </div>
  </div>
</form>

Text Content

WE VALUE YOUR PRIVACY

We and our partners store and/or access information on a device, such as cookies
and process personal data, such as unique identifiers and standard information
sent by a device for personalised ads and content, ad and content measurement,
and audience insights, as well as to develop and improve products. With your
permission we and our partners may use precise geolocation data and
identification through device scanning. You may click to consent to our and our
partners’ processing as described above. Alternatively you may click to refuse
to consent or access more detailed information and change your preferences
before consenting.
Please note that some processing of your personal data may not require your
consent, but you have a right to object to such processing. Your preferences
will apply to a group of websites. You can change your preferences at any time
by returning to this site or visit our privacy policy.
MORE OPTIONSDISAGREEAGREE
PreprintPDF Available


A SURVEY ON LEGAL QUESTION ANSWERING SYSTEMS

 * October 2021

Authors:
Jorge Martinez-Gil
 * Software Competence Center Hagenberg



Download file PDFRead file
Preprints and early-stage research may not have been peer reviewed yet.
Download file PDF
Read file
Download citation
Copy link Link copied

--------------------------------------------------------------------------------

Read file
Download citation
Copy link Link copied
References (95)





ABSTRACT

Many legal professionals think that the explosion of information about local,
regional, national, and international legislation makes their practice more
costly, time-consuming, and even error-prone. The two main reasons for this are
that most legislation is usually unstructured, and the tremendous amount and
pace with which laws are released causes information overload in their daily
tasks. In the case of the legal domain, the research community agrees that a
system allowing to generate automatic responses to legal questions could
substantially impact many practical implications in daily activities. The degree
of usefulness is such that even a semi-automatic solution could significantly
help to reduce the workload to be faced. This is mainly because a Question
Answering system could be able to automatically process a massive amount of
legal resources to answer a question or doubt in seconds, which means that it
could save resources in the form of effort, money, and time to many
professionals in the legal sector. In this work, we quantitatively and
qualitatively survey the solutions that currently exist to meet this challenge.

Discover the world's research

 * 20+ million members
 * 135+ million publications
 * 700k+ research projects

Join for free


Public Full-text 1



Content uploaded by Jorge Martinez-Gil
Author content

All content in this area was uploaded by Jorge Martinez-Gil on Oct 15, 2021
Content may be subject to copyright.
arXiv:2110.07333v1 [cs.IR] 12 Oct 2021
A Survey on Legal Question Answering Systems
Jorge Martinez-Gil
Software Competence Center Hagenberg
Softwarepark 32a, 4232
Hagenberg im Muhlkreis, Austria
jorge. martinez-gil@ scch. at
Abstract
Many legal professionals think that the explosion of information about local,
regional, national, and
international legislation makes their practice more costly, time-consuming, and
even error-prone. The
two main reasons for this are that most legislation is usually unstructured, and
the tremendous amount
and pace with which laws are released causes information overload in their daily
tasks. In the case
of the legal domain, the research community agrees that a system allowing to
generate automatic
responses to legal questions could substantially impact many practical
implications in daily activities.
The degree of usefulness is such that even a semi-automatic solution could
significantly help to reduce
the workload to be faced. This is mainly because a Question Answering system
could be able to auto-
matically process a massive amount of legal resources to answer a question or
doubt in seconds, which
means that it could save resources in the form of effort, money, and time to many
professionals in
the legal sector. In this work, we quantitatively and qualitatively survey the
solutions that currently
exist to meet this challenge.
Keywords: Knowledge Engineering, Data Engineering, Legal Intelligence, Legal
Information
Processing, Legal Question Answering
Preprint submitted to Computer Science Review October 15, 2021





1. Introduction
Legal Intelligence (LI) is a broad umbrella covering a group of technologies
aimed at automating
tasks that formerly required human intelligence, communication, and
decision-making in the legal
domain. Many applications use LI technology to support services, including legal
decision support,
adaptive interfaces, and mining litigation data for strategic legal advice, etc.
In the context of this
work, we focus on a subset of the Question Answering (QA) challenge; the design
of solutions for
automatic Legal Question Answering (LQA). The LQA challenge is about building
methods and tools
that involve accurately ingesting the query information from a person as an
input and automatically
offering a response tailored to the person’s needs as an output.
The LQA challenge is far from being trivial since one of the legal domain’s
significant problems
is that newly generated legislation is usually formatted unstructured. Moreover,
the vast volume
and speed at which legislation is made available usually lead to an information
overload problem.
Therefore, finding practical solutions is considered a significant research
challenge, and it is a matter
of intense research. In fact, in recent times, we have witnessed an explosion in
the number of solutions
for LQA.
The truth is that this domain is a perfect candidate for data and knowledge
scientists and prac-
titioners to validate their systems because it meets all the requirements that
make the problem at-
tractive: there is a massive amount of information, which also grows and grows
every day (e.g., the
vast amount of legislation at local, regional, national and international levels
that is generated every
day), and this information often lacks structure (almost all public
administrations publish their laws
in unstructured formats and without annotation, at most with some metadata). As
a result, people
in the sector face severe difficulties navigating this sea of information.
Numerous solutions claim to solve the problem to overcome the above-mentioned
problems (with
varying degrees of success). This is mainly because these solutions are
explicitly designed to automat-
ically process many legal sources to answer a question or doubt in a matter of
seconds, which means
2


it could save resources in the form of effort, money, and time for many legal
professionals. Therefore,
both academia and industry have a strong interest in presenting their solutions
as an excellent candi-
date to overcome some of the limitations in the legal field by trying to
automatically provide answers
to given questions, rather than presenting the professionals with a list of
related items (Kolomiyets &
Moens, 2011).
If we look at the historical evolution of this type of solution, we can realize
that in the past, most
techniques were based on the development or refinement of information retrieval
(IR) techniques that
would allow legal texts to be processed to answer questions. However, in recent
times, both academia
and industry are turning more to neural network-based solutions. However, both
have significant
advantages and disadvantages. For example, while classic IR techniques require
very long processing
times, which can be mitigated with high-performance hardware architectures,
neural network-based
techniques require a lot of training time, but once trained, they are much
faster, and they usually
deliver good results. However, the significant problems associated with these
neural architectures
are the low interpretability of the final models, the tremendous amount of
resources needed for their
training, and the great difficulty in transferring the learned models.
The tremendous impact that QA systems can have on a wide range of academic
disciplines and
business models has attracted much attention from both the research community
and the legal indus-
try. As a result, there already several surveys covering several aspects of this
challenge, e.g., (Wang,
2006; Kolomiyets & Moens, 2011; H¨offner et al., 2017; Diefenbach et al., 2018;
Franco et al., 2020;
Dimitrakis et al., 2020; da Silva et al., 2020). However, the present survey is
different from the previous
ones in many ways. First, it is, to the best of our knowledge, the first attempt
to compile the existing
literature to address the LQA problem. Moreover, we intend to do it informative
and transverse,
without focusing on a specific computational method or family of information
sources. Finally, it is
the most updated to date. We are collecting the latest advances related to the
advances in neural
computation and knowledge graphs.
3


















Therefore, to articulate this survey’s organization, we intend to answer the five
questions that we
consider key to meeting the challenge. This means that the contribution of the
present work is based
on an objective and depth answer to the following five main research questions:
•RQ1. What are the most significant LQA research works to date?
•RQ2. How do the primary data sources on which LQA relies look like?
•RQ3. How do the current LQA solutions address the problem of interpretability?
•RQ4. How is the quality of LQA systems currently assessed?
•RQ5. What are the research challenges that remain open in this context?
The rest of the work is structured as follows: Section 2 addresses RQ1 by
presenting the state-of-
the-art LQA methods categorized by the different conceptual approaches they are
based on. Section
3 addresses RQ2 by explaining the primary sources of data, information, or
knowledge from which
existing solutions are usually sourced. Section 4 addresses RQ3, introducing the
concept of inter-
pretability, explaining its paramount importance in the legal domain, and
analyzing which solutions
are the most advanced in this respect. Section 5 addresses RQ4, explaining which
are the most
common methods for the evaluation and quality assessment of the different LQA
systems. Section 6
addresses RQ5 explaining the challenges that industry and academia will face
soon if they want their
solutions adopted by professionals in the sector. Finally, we note the
conclusions and lessons learned
that could be drawn from this work.
2. Related Works
In the context of this work, we assume that a question usually comes in the form
of a natural
language sentence, which usually starts with an interrogative word and expresses
some information
need of the user, although sometimes a question could also take a form of an
imperative construct
4


and starts with a verb. The answer to a question, which may or may not be known
in advance,
would be a word, number, date, sentence, or paragraph that completely satisfies
the information need
expressed by the question. The standard process of automatically answering a
question involves three
fundamental phases: Query Understanding (to extract the core elements that make
up the question),
Sources Retrieval (to identify a few documents that may contain the answer from
a large source pool),
and Answer Extraction (to find a short phrase or sentence that answers the user’s
question directly).
However, not all methods implement the different phases similarly, as we will see
throughout the
section.
The problem of automatically match an answer to a question is widely assumed to
be an impor-
tant research challenge that tries to address one aspect of Artificial
Intelligence (AI) that will allow
computers to perform routine and tedious tasks. Nevertheless, automatically
answering questions is
genuinely complex and encompasses many challenges and nuances that make it
unfeasible to be cov-
ered by just an application. This means that most of the associated challenges
could have an impact
on both industry and academia. This is because having information systems that
can correctly answer
questions asked by a person opens a wide range of possibilities to impact, from
the most basic research
to the most advanced business models.
For this reason, there are numerous solutions to the problem of QA. While in the
past, most
techniques were based on the development or refinement of natural language
techniques that would
allow legal texts to be processed to answer questions, in recent times, both
academia and industry
are turning more to neural network-based solutions inspired in the seminal work
of Mikolov et al.
(Mikolov et al., 2013). For example, BERT (Devlin et al., 2019) and ELMo (Peters
et al., 2018)
or USE (Cer et al., 2018). These approaches have tremendous advantages and
disadvantages. For
example, while traditional techniques require properly tuning up different
workflow pipelines, neural
network-based techniques can be trained once and use many times later. The
problem is the lack of
interpretability, i.e., the limited capacity a person might have to understand
why the methods work
so well and the number of resources required for training and adaptation to new
cases.
5










In this survey, we focus strictly on existing LQA work, i.e., questions and
answers restricted to the
world of law. Table 1 shows a summary of the features that make them different
from other question
answering systems. First of all, in LQA, we work in a closed environment.
Moreover, the questions
can only be related to the legal domain. In addition, all types of questions are
allowed, whether they
are definition, comparison, confirmation, etc. The analysis of the texts, both
questions and sources to
extract the answers, is not limited. Nor are the types of sources that can be
used to find the answer
limited. Last but not least, in the context of the current survey, we do not
include dialogue systems
but only question answering solutions.
Several different classifications schemes can be potentially established for
categorizing the main
families of systems built in this domain. QA is a form of information retrieval,
and this means that
solutions are defined by how they represent the input information and the
retrievable entity as well
as by the ranking function. To properly articulate this section, we have focused
on the different
computational approaches for building LQA solutions. Techniques of this kind
have been widely
used in various forms of research on Yes/No Answers, Multiple Choice Question
Answering, Classical
Solutions from the Information Retrieval field, Ontology-based solutions, Neural
solutions, and Other
approaches.
•Yes/No Answers
•Legal Multiple Choice Question Answering
•Classical IR-based Solution
•Big Data solutions
•Ontology-based solutions
•Neural solutions
•Other approaches, e.g., Knowledge Graphs
6


Table 1: Kind of QA systems covered within this survey
Features of Legal Question Answering
Environment [Open] Closed
[Closed]
Domain
[General-purpose] Legal
[Mathematical]
[Legal]
[Other]
Kind of Questions [Factoid] All
[Definition]
[Confirmation]
[Causal]
[Comparison]
Kind of Analysis
[Morphological] All
[Syntactical]
[Semantic]
[Pragmatic]
Sources [Structured] All
[Semi-structured]
[Unstructured]
Kind of dialogue
[Question Answering] Question Answering
[Dialog]
[Spoken Dialog]
7


Each of the different branches of research has its peculiarities. Thus, for
example, the first types
based on Yes/No Answers and Multiple Choice Question Answering assume a
simplified version of the
problem where it is unnecessary to worry about generating candidate answers,
only about confirming
or not the veracity of a sentence or constructing a ranking of answers.
Classical Information Retrieval
techniques, in principle, should be one the most numerous families since they
are one of the natural
ways of tackling the problem. On the other hand, Big Data techniques, even if
they do not represent
an apparent scientific breakthrough, do represent a considerable advance in
technical characteristics,
where even unfeasible but straightforward techniques on a regular computer
(think, for example of
the calculation of regular expressions) can have a good performance. Finally,
the novel techniques
based on neural architectures with the best-expected results and other
approaches come to fill some
of the gaps related to neural networks, such as, for example, the lack of
interpretability. We then
explain what search strategy we followed to capture the primary sources of the
literature, we review
the different approaches, and we conclude the section with a summary.
2.1. Data Sources and Search Strategy
To choose the primary sources of this survey, we have built a search string so
that it includes
two major search terms: ’Method’ and ’Field’. The first major search term
represents the employed
methodology, namely, ’question answering’, whereas the second major search term
illustrates the fields
in which the method should have been utilized, i.e., ’legal’. This term includes
all sorts of technologies
and synonyms of legal in which that application should occur. Terms like ’legal
domain’, ’legal sector’,
’legal industry’, ’legal area’, ’legal documents’, ’legal material’,
’legislation’, and ’regulatory’. In
addition, through snowballing, we have come up with articles that contain terms
such as ’due diligence’.
Queries were run on Google Scholar1, DBLP2, Microsoft Academic3, and Semantic
Scholar4.
1https://scholar.google.com/
2https://dblp.org/
3https://academic.microsoft.com/home
4https://www.semanticscholar.org/
8






Table 2: Inclusion and exclusion criteria for work selection
Inclusion Exclusion
Written in English Written in other languages rather than English
Published in journal, conference or workshop Preprints when the paper is
available
Described an algorithm, technique or system Thesis, editorials, prefaces, or
summaries
Uses, at least, one LQA dataset QA method but no application in LQA
Published from January 1990 Published by a predatory publisher
Published until June 2021 Patents or patent applications
Moreover, we have had to use some criteria to decide whether a particular work
should be consid-
ered. Table 2 shows the criteria we have followed in deciding whether a paper
should be considered
for inclusion or exclusion in this survey. This table follows the recommendation
stated by (Soares &
Parreiras, 2020) for the study of QA systems.
As a result of the screening process, we have obtained 65 primary sources
analyzed and categorized
below within their corresponding family of solutions.
2.2. Yes/No Answers
This approach is the simplest of all. Given a question and a known answer, one
tries to determine
whether the associated answer is true or false. In this way, the systems do not
need to generate
candidate answers. Nevertheless, answering yes/no questions in the legal area is
very different from
other domains. One of its inherent features is that legal statements need a
thorough examination of
predicate-argument structures and semantic abstraction in these statements. The
methods from this
family usually check if there is a high degree of co-occurrence between
questions and answers in corpora
of a legal nature. Some variants might result in searching for some kinds of
regular expressions.
Works based on this approach, such as (Kim et al., 2013) have developed different
approaches
to answer yes/no questions relevant to civil laws in legal bar exams. A bar
examination is intended
9






Table 3: Summary of approaches based on Yes/No
Work Approach
(Kim et al., 2013) Antonym detection and Semantic Analysis
(Kim et al., 2014) TF-IDF, LDA and SVM
(Kim et al., 2016) Paraphrasing detection
(Taniguchi & Kano, 2016) Case-role analysis
(Kano et al., 2017) Using linguistic structures
(Taniguchi et al., 2018) Using Framenet
to determine whether a candidate is qualified to practice law in a given
jurisdiction. There is a
recurring concern in the literature on whether it is possible to pass this type
of test without human
supervision. The same author (Kim et al., 2014) has also implemented some
unsupervised baseline
models (TF-IDF and Latent Dirichlet Allocation (LDA)) and a supervised model,
Ranking SVM, for
facing the challenges. The model features are a set of words and scores of an
article based on the
corresponding baseline models. In addition, a final enhancement includes the use
of paraphrasing
(Kim et al., 2016). Other authors (Taniguchi & Kano, 2016; Taniguchi et al.,
2018) has addressed the
problem by case-role analysis and Framenet5, respectively. Finally, Kano et al.
has explored, for the
first time, linguistic structures (Kano et al., 2017). Table 3 summarizes all
these works.
This branch of research was one of the first to yield results. Its strength is
based on a superficial
understanding of the problem with solutions that tend to work well. The major
problem of this model
is that questions in human language express a well-defined information
requirement on the one hand,
but they also convey more information than an essential list of relevant
keywords since they represent
syntactic and semantic links between those keywords on the other. Therefore,
solutions tend to work
well only to a limited extent, i.e., when the questions are simple.
5https://framenet.icsi.berkeley.edu/fndrupal/
10

























Pros: High interpretability. Methods can be used in many other domains for
fact-checking
Cons: The operating model is trivial and leads to poor results at the present
time
2.3. Legal Multiple Choice Question Answering
Legal Multiple Choice Question Answering (LMQA) consists of correctly answer
questions in a
scenario where the possible answers are already given beforehand. In this model,
there are two clear
facilitators. It is unnecessary to generate candidate answers since this family
of methods assumes
that candidate answers are a different problem and should be studied apart.
Furthermore, it is known
that the correct answer is among the given ones (even in situations where None
of above answers are
allowed). Therefore, the goal is to learn a scoring function S(F, z ) with a
normalization parameter z
(whereby zor the normalization factor is usually used so that the values
associated with each answer
are in the same numerical range) such that the score of the correct choice is
higher than the score of
the other hypotheses and their corresponding probabilities.
Some works have used the classical Latent Semantic Analysis to obtain good
results in this con-
text (Deerwester et al., 1990). In general, all techniques based on
co-occurrence are very appropriate
because it is usually sufficient to check the number of times that the question
and each of the answers
appear together in some textual corpora of a legal nature (Li et al., 2003). Of
course, one improvement
is to use semantic similarity detection techniques to detect different
formulations of the same informa-
tion (Han et al., 2013). In addition, crowdsourcing techniques have also been
very successful (Aydin
et al., 2014). There are variations of the above that attempt to compute
co-occurrence in a more
sophisticated way so that text windows and regular expressions are taken into
account (Martinez-Gil
et al., 2019b). More recently, (Chitta & Hudek, 2019) has proposed an
improvement through multi-
class classifiers to identify the answer to the question. The key idea is to
handle imbalanced data
by generating synthetic instances of the minority answer categories. Table 4
chronologically lists the
works that we have mentioned above.
11














Table 4: Summary of approaches based on Multiple Choice
Work Approach
(Deerwester et al., 1990) Latent Semantic Analysis
(Li et al., 2003) Co-occurrence
(Han et al., 2013) Synonym detection
(Aydin et al., 2014) Crowdsourcing
(Martinez-Gil et al., 2019b) Reinforced Co-occurrence
(Chitta & Hudek, 2019) Synthetic minority oversampling
One of the fundamental keys for LMCQA to work well is the choice of the legal
corpus to be worked
on, both in terms of extension and quality since most methods require the
calculation of semantic
similarity, textual co-occurrences, etc.
Pros: High speed. High interpretability. Many existing methods for calculating
rankings
Cons: Knowing all answers beforehand is not a realistic situation in the real
world
2.4. Classical IR-based Solutions
In this context of LQA, classical IR-based solutions usually represent words in
the form of discrete
and atomic units. This family of methods assumes that it is vital that the
solution appropriately
identify either the exact answer or source and specific paragraph containing the
relevant information.
The lawyer then should read and interpret the excerpt for the client. For
example, the first approach
(and the simplest) could query the number of Google results for a specific
question and a given
answer together. However, this solution has brought several problems like the
lack of context for the
formulated question.
To overcome the problem of the lack of context, word processing models such as
LSA (Deerwester
et al., 1990) and term frequency-inverse document frequency (TF-IDF) partially
solve these ambigu-
ities by using terms that appear in a similar context based on their vector
representation. Then they
12
















group the semantic space into the same semantic cluster. Within this family of
methods, one of the
best-known QA systems is IBM Watson (Ferrucci et al., 2013), which is very
popular for its victory
in the televised show Jeopardy (Tesauro et al., 2013). Although in recent times,
IBM Watson has
become a generic umbrella that includes other business analytics capabilities.
If we restrict ourselves to LQA, one of the most popular initiatives is
ResPubliQA. This evaluation
task was proposed at the Cross-Language Evaluation Forum (CLEF) (Pe˜nas et al.,
2009, 2010) a
consists of given a pool of independent questions in natural language about
European legislation,
proposed systems should return a passage (but not an exact response) that
answers each question.
This initiative was a remarkable success. In fact, thanks to ResPubliQA, this
family of methods has
been the subject of intensive research during the last decade.
In relation to the works of this family, it is possible to identify,
Brueninghaus and Ashley that have
addressed the problem for the first time with a focus on information extraction
methods (Br¨uninghaus
& Ashley, 2001). Quaresma et al. focused on Portuguese legislation using a
classical IR pipeline
(Quaresma & Rodrigues, 2005). Some years after, (Maxwell & Schafer, 2008) went a
step further
beyond by using conceptual and contextual search. Another strategy was proposed
by (Monroy
et al., 2008) that returns several relevant passages extracted from different
legal sources. The relevant
passages allows generating answers to questions formulated in plain language
using graph algorithms.
A qualitative improvement was presented (Monroy et al., 2009) to lemmatizing and
using manual
and automatic thesauri for improving question-based document retrieval. For the
construction of
the graph, the authors followed the approach of representing all the articles as
a graph structure. In
parallel, (Tran et al., 2013) proposed mining reference information. Last but
not least, (Rodrigo et al.,
2013) proposed a new IR-based system with the lessons learned from ResPubliQA.
More recently, (Carvalho et al., 2016) have introduced a new approach for
lexical to discourse-level
corpus modeling, and (Bach et al., 2017) proposes a novel solution based on
Conditional Random
Fields (CRFs), which are methods that resorts on statistical modeling in order
to segment and label
13



























sequence data. Furthermore, (van Kuppevelt & van Dijck, 2017) decided to explore
how to perform
network analysis and visualization as a proper way to deal with Dutch case law.
Meanwhile, (Delfino
et al., 2018) has brought for the first time the Portuguese version of the
thesaurus OpenWordNet6to
the table.
The last works in this direction are those from (Hoshino et al., 2018) by using
predicate-argument
structure, (McElvain et al., 2019a,b) over plain text and focusing on
non-factoid kind of questions.
Non-factoid questions are those whose answer is not directly accessible in the
target document, which
demands some inference and perhaps extra processing to obtain an answer. Wehnert
et al. have
tried a new application of the popular BM25 approach (Wehnert et al., 2019), and
the latest to date,
(Verma et al., 2020) has had a focus on relevant subsection retrieval to answer
the questions of legal
nature. The story so far can be summarized in Table 5.
As can be seen, this family is extensive. Moreover, in its bosom, many and very
diverse proposals
have been born to face the problem. Before the latest advances in Deep Learning,
it was the dominant
family, and its solutions were considered good until those times. However, in
recent times very tough
competitors have emerged.
Pros: High performance, High interpretability
Cons: In recent times, accuracy has been overtaken by neural solutions
2.5. Big Data solutions
In recent times, the arrival of novel Big Data approaches has come up with many
advantages and
challenges for the legal sector. The possibilities of Big Data like the effective
and efficient processing
of massive amounts of data have introduced many opportunities and challenges.
Nevertheless, some
of the limitations that have traditionally burdened this area of knowledge
remain. Perhaps the most
illustrative of these limitations is the inability of the systems to understand
the question instead of
6https://github.com/own-pt/openWordnet-PT
14
















Table 5: Summary of approaches based on classical and novel approaches for
Information Retrieval
Work Approach
(Br¨uninghaus & Ashley, 2001) Information extraction methods
(Quaresma & Rodrigues, 2005) Classical IR pipeline
(Maxwell & Schafer, 2008) Conceptual and contextual search
(Monroy et al., 2008) Exploitation of Graph-based algorithms
(Monroy et al., 2009) Exploitation of Graph-based algorithms
(Tran et al., 2013) Mining reference information
(Rodrigo et al., 2013) Lessons learned from ResPubliQA
(Carvalho et al., 2016) Lexical to discourse level corpus modeling
(Kim & Goebel, 2017) Cascaded textual entailment
(Bach et al., 2017) Conditional Random Fields
(van Kuppevelt & van Dijck, 2017) Network Analysis
(Delfino et al., 2018) Exploitation of word senses and relations
(Hoshino et al., 2018) Predicate Argument Structure
(McElvain et al., 2019a) Non-Factoid questions
(McElvain et al., 2019b) Non-Factoid questions
(Wehnert et al., 2019) BM25 and Elasticsearch
(Verma et al., 2020) Relevant subsection retrieval
15




































Table 6: Summary of approaches based on Big Data
Work Approach
(Bennett et al., 2017) Scalable architecture by design
(Mimouni et al., 2017) Approximate Answers and Rich Semantics
(Martinez-Gil et al., 2019a) Mutual Information Exchange
merely using its huge computational potential to search for statistical patterns
in corpora of a legal
nature.
For this reason, the existing approaches are focused on developing good ideas
that allow to answer
the questions asked and obviate the performance details since these can be
solved with these new forms
of high-performance computing. Some outstanding works in this direction are
proposed by Bennet
et al. with a focus on the scalability of the solution by design (Bennett et
al., 2017), or Mimouni
et al. for handling the problem of complex queries by working with approximate
answers and richer
semantics (Mimouni et al., 2017). Last but not least, there is also a novel
approach based on the
concept of mutual information exchange (Martinez-Gil et al., 2019a), which, when
applied to large
volumes of data, have demonstrated better performance than classic co-occurrence
methods. Table 6
lists chronologically these works.
On the other hand, these applications entail processing and handling large
amounts of data quickly,
and they facilitate the automation of specific low-level computing operations. So
in this way, most
technical limitations disappear, and therefore, methods that were previously
considered too expensive
in terms of resource consumption in the form of time are no longer a problem.
For example, the
execution of queries based on the concept of regular expression that is very
expensive in terms of
computing time. However, the scientific difficulties to model appropriate
inter-dependencies between
the concepts from the questions remain, so the results are not optimal yet.
16














Pros: Very good performance, no limitations of a technical nature
Cons: Inability of the systems to properly understand the meaning of a complex
question
2.6. Ontology-based solutions
The capacity to construct a semantic representation of the data paired with the
accompanying
domain knowledge is one of the significant benefits of adopting domain ontologies.
Ontologies can also
be used to establish connections between various sorts of semantic knowledge. As
a result, ontologies
can be utilized to develop various data-search strategies. The legal domain has
not remained oblivious
to this line of research.
Examples of works that deal with the first approach are Lame et al. proposed for
the first time
the integration of NLP and ontologies to solve LQA problems (Lame, 2004).
Moreover, Xu et al.
proposed a system for relation extraction and textual evidence(Xu et al., 2016).
In addition, Fawei
et al. (Fawei et al., 2018), in the first instance, proposed a criminal law and
procedure ontology.
Meanwhile, in the second instance, proposed a way to semi-automatically
construct ontologies for the
legal section (Fawei et al., 2019). Another approach has consisted of Semantic
Role Labeling (Veena
et al., 2019), and last but not least, Kourtin et al. have opted for an
RDF-SPARQL tailored design
(Kourtin et al., 2021). Table 7 shows us a summary of the works mentioned above.
Solutions based on ontological foundations have always promised to provide the
correct answer to
a question. However, in practice, things are not rather complicated. There are
very few ontologies,
and those that exist are hardly reusable. So it is costly and error-prone to
develop legal ontological
models from scratch. Not to mention the learning curve that dealing with
description logics present,
and the difficulty of performing large-scale reasoning in reasonable time.
Pros: Logical reasoning provides accurate answers. High interpretability
Cons: Models very expensive to build, very difficult to reuse, slow reasoning
times
17














Table 7: Summary of approaches based on ontologies
Work Approach
(Lame, 2004) Combination of NLP and ontologies
(Xu et al., 2016) Relation extraction and textual evidence
(Fawei et al., 2018) Using a criminal law and procedure ontology
(Fawei et al., 2019) Using semi-automated ontology construction
(Veena et al., 2019) Semantic Role Labeling
(Kourtin et al., 2021) RDF-SPARQL tailored design
2.7. Neural solutions
Solutions based on architectures inspired by neural models have adequately and
consistently solved
most benchmarks in the QA field. These results are outstanding and have made most
of the community
decide to use their resources to explore this direction. In recent times,
several works have been
presented. There are two significant lines of research: A branch that tries to
develop solutions that
exploit structured resources that have led to a family that we will see later.
Furthermore, another
branch that tries to develop methods capable of exploiting unstructured
information.
Exemplary works that operate with unstructured text using the most modern
machine learning
techniques are, for example, works such as (Kim et al., 2015; Do et al., 2017)
used for the first time a
convolutional neural network to LQA. A few years after, the same authors
proposed an improvement
using deep siamese networks (Kim et al., 2017). Other authors decide to explore
new kinds of systems
using LSTM (John et al., 2016), the concept of neural attention (Morimoto et
al., 2017). The concept
of neural attention consists of freeing the encoder-decoders architecture from a
fixed-length internal
representation. Furthermore, the exploitation of a Multi-task Convolutional
Neural Network (Xiao
et al., 2017) has also been proposed.
Some recent works such as (Cer et al., 2018; Devlin et al., 2019; Peters et al.,
2018) have made
some of the most critical advances in question answering in recent times. These
approaches have con-
18
































sistently achieved top results in most academic competitions with a clear
general-purpose orientation.
Moreover, its performance in the legal sector is more than remarkable.
Futhermore, (Collarana et al., 2018) introduced for the first time a new
technique based on Match-
LSTM and a Pointer Layer, whereas (Nicula et al., 2018) proposed to use Deep
Learning over candidate
contexts. (Liu & Luo, 2018) built a system using Deep Neural Networks. In
parallel, Ravichander
et al. proposed a new strategy based on the combination of computational and
legal perspectives
(Ravichander et al., 2019). And last, but not least, Zhong et al. proposed two
novel alternatives using
Reinforcement Learning and Deep Learning respectively (Zhong et al., 2020b,c).
Among the latest published works in this family of methods, we can highlight the
following: The
Neural Attentive Text Representation from (Kien et al., 2020), the introduction
of Few-shot Learning
in the legal domain (Wu et al., 2020). Novel techniques for language modeling
(Huang et al., 2020b).
Deep learning restricted to the world of building (Zhong et al., 2020a). And
diverse applications of
the successful BERT to implement LQA solutions (Holzenberger et al., 2020; Zhang
& Xing, 2021).
Table 8 shows a chronological summary of the aforementioned solutions.
The advances in the numeric representation of words has opened a complete family
of methods.
Many different training methods are able to generate word embeddings from
unstructured data, mak-
ing the novel semantic analysis models achieve state-of-art performance. As for
the legal point of
view, it is critical to find out an efficient way to represent the semantic meaning
of the longer pieces
of text, such as phrases, sentences, or documents, to achieve a more reasonable
interpretation.
On the other hand, one major limiting factor in this context is that there are
usually no legal
datasets to train neural models at a scale properly. Hence, the performance
expected for an LQA
system is usually worse than in a generic one.
Pros: These methods are capable of achieving the best results to date
Cons: Very poor interpretability. Huge amount of data needed for training
As a final note, it is worth mentioning that although the lack of
interpretability of neural solutions is
19

























Table 8: Summary of LQA approaches based on neural networks
Work Approach
(Kim et al., 2015) Convolutional Neural Network
(John et al., 2016) Long-Short Term Memory
(Do et al., 2017) Convolutional Neural Network
(Kim et al., 2017) Deep Siamese Networks
(Morimoto et al., 2017) Neural Attention
(Xiao et al., 2017) Multi-task Convolutional Neural Network
(Cer et al., 2018) Universal Sentence Encoder
(Devlin et al., 2019) Bidirectional Encoder Representations from Transformers
(Peters et al., 2018) Embeddings from Language Model
(Collarana et al., 2018) Match-LSTM + Pointer Layer
(Nicula et al., 2018) Deep Neural Networks over candidate contexts
(Liu & Luo, 2018) Deep Neural Networks
(Ravichander et al., 2019) Combination of computational and legal perspectives
(Zhong et al., 2020b) Reinforcement Learning
(Zhong et al., 2020c) Deep Neural Networks
(Kien et al., 2020) Neural Attentive Text Representation
(Wu et al., 2020) Few-shot Learning
(Holzenberger et al., 2020) Based on BERT
(Huang et al., 2020b) Language Modeling
(Zhong et al., 2020a) Deep Neural Networks over building regulation
(Zhang & Xing, 2021) Based on BERT
20












































always mentioned, an effort is being made in recent times to develop frameworks
that help the human
operator (Ribeiro et al., 2016; Lundberg & Lee, 2017). These frameworks are
intended to enable
machine learning engineers and relevant domain specialists to analyze the
end-to-end solutions and
discover differences that could lead to a sub-optimal performance concerning the
desired objectives.
2.8. Towards a new family of solutions
Without detracting from the merit of neural solutions, many legal professionals
are not satisfied
with just an answer to their question and demand much more. They demand to
understand why
the solution has opted for such an answer and not another alternative. This
brings up one of the
ghosts that has always been associated with neural computing, that is, that its
behavior is similar
to that of a black-box since it is possible to give it an input and obtain an
output. However, it is
humanly impossible to understand how the connections of thousands of neurons
have worked to give
rise to such an answer, or even the real meaning behind the feature vectors
obtained after subjecting
the model to the training of a neural nature. For example, the base configuration
of BERT (Devlin
et al., 2019) requires a configuration consisting of 12 layers of neurons and 12
windows of attention.
Therefore, there are usually issues related to the interpretability of the
resulting model. This is where
this family of methods comes in.
While Linked Data is now well-established and has a strong community behind it
that has been
studying it for several years, there is a lack of work on its application to the
legal domain. It is true,
however, that some proposals have begun to be put on the table. Nevertheless,
what really promises
to impact this area is the combination of Knowledge Graphs and Machine Learning
to create a new
generation of LQA systems.
To date, some authors thought that exploring the Linked Data route was the way
to go (He et al.,
2013; Filtz, 2017). However, the solutions are still unrealistic, as most of the
legislation is still not
published in a structured way. Furthermore, automatically structuring it is
currently unfeasible. In
addition, the learning curve for formulating questions in graph-oriented
languages has led to the search
21












Table 9: Summary of LQA approaches based on other approaches
Work Approach
(He et al., 2013) Linked Data
(Filtz, 2017) Linked Data
(Tong et al., 2019) Legal Knowledge Graph
(Sovrano et al., 2020) Legal Knowledge Graph
(Huang et al., 2020a) Legal Knowledge Graph
(Filtz et al., 2021) Legal Knowledge Graph
for alternative approaches. In recent times, several works have also emerged
that use for the first time
the potential of knowledge graphs to overcome some of the gaps of neural
solutions, such as the lack of
interpretability and the need for large amounts of data for training (Tong et
al., 2019; Sovrano et al.,
2020; Huang et al., 2020a; Filtz et al., 2021). In addition, technologies such
as GPT-37have been
developed to convert questions formulated in natural language to their
equivalent in graph-oriented
language. Table 9 lists the works above.
Legal Knowledge Graphs has attracted much attention in recent times due to that
an increasing
number of researchers point out that they can bring a high degree of accuracy
together with the
highest interpretability that can be achieved to date. However, the other side
of the coin says that
this approach can offer good results at a high cost (in terms of money, time, and
effort needed).
This is mainly because this technology faces some obstacles in its development
related to the amount
of engineering work to build graphs properly and the necessary improvement of
natural language
translators to graph query language.
Pros: High accuracy. High interpretability. High performance
Cons: No good or cheap methods for automatizing the whole process
7https://openai.com/
22























2.9. Summary
It is clear from the literature that working with information concerning
legislation and case law
has always been attractive to computer scientists and practitioners applying for
the latest advances in
language and information technologies. These technologies have proven to be very
useful for solving
several problems that have traditionally affected legal information processing.
In practice, the daily
activities of these legal professionals requires checking vast amounts of legal
material necessary to assess
the relevant information pieces and identify the correct fragment they need.
With this objective in
mind, the LQA discipline was born and has been developed.
We provide a general summary of the different methods followed to build LQAs in
the last decades.
Figure 1 shows how, at present, of the 65 primary sources analyzed, the vast
majority belong to the
families of neural solutions and IR-based methods, although some families, such
as other approaches,
e.g., Legal Knowledge Graphs, are getting traction in recent times.
Neural Solutions
32.3%
Information Retrieval
26.1%
Yes-No Answers
9.2%
LMCQA
9.2%
Ontologies
9.2% Other approaches
9.2%
Big Data
4.8%
Figure 1: Summary of existing families to meet the challenge of LQA
Next, we will look at the sources of data, information, and knowledge that are
commonly used
when building LQA solutions.
23


3. Data, Information and Knowledge Sources
LQA solutions are a kind of system that analyzes one or various data sources to
answer a given
question automatically. These data sources are usually structured or
unstructured. The first ones
are commonly referred to as legal corpora, while the latter are usually
considered Legal Knowledge
Bases (LKBs). Please note that in this domain, data is considered the atomic
unit with which the
systems work. If this data is structured to facilitate its interpretation, it is
considered information.
Finally, when relationships are created to facilitate understanding and
communication, it is considered
knowledge.
Depending on the sources to be exploited, two techniques for tackling the
problem are working with
unstructured legal corpora or working LKBs. Sometimes a third option is also
considered concerning
semi-structured sources (although this scenario is usually considered structured
as well). Each one
has its own set of benefits and drawbacks. Working with structured LKBs, for
example, let system
designers take advantage of the knowledge represented by using so-called
inference engines to infer
new information and answer questions. The fact is that not easy to implement
these systems, so they
have been progressively replaced by more efficient systems based on lighter
knowledge models such
as knowledge graphs and other enhanced lexical semantic models (Yih et al.,
2013). However, at the
beginning of this survey, we mentioned that legislation usually comes in an
unstructured form, so it is
more realistic to design QA systems being prepared to ingest vast amounts of
data from unstructured
sources. We will now focus on the two fundamental kinds of data sources that LQA
solutions typically
use to find the appropriate answers to the questions asked. We divide these
sources into unstructured
and structured.
3.1. Unstructured sources
As we have seen repeatedly in this survey, most legislation is generated in an
unstructured manner
and at high speed. This means that designing methods capable of working with
unstructured text is
usually the norm when dealing with legal information. LQAs solutions exploiting
unstructured sources
24




have several practical benefits as most of them have been specifically designed to
efficiently process
vast amounts of textual data (usually represented in legal language). These
enormous amounts of data
come from existing documents, legal databases, and so on. Most of the methods of
the Information
Retrieval and Neural Solutions families are based on this strategy.
The impossibility of having structured legal information on a large scale, of
high quality, and
covering all the current legislation, makes researchers and practitioners
consider using methods to
exploit unstructured resources of a legal nature. For this reason, the most
frequent type of LQA
solution needs to identify the relevant legal documents or information pieces
and split them into
candidate passages, and then, face the main key to the problem that consists of
extracting the answer
from the candidate passages.
The current generation of LQA solutions has evolved to extract answers from a
wide range of dif-
ferent plain machine-readable resources. These LQA solutions exploit the massive
set of unstructured
information available on some sources to retrieve information about any
particular question. It is
important to note that these QA systems are possible mainly due to the
extraction functions that
are usually either heuristic or learned from a dataset such the Stanford
Question Answering Dataset
(SQuAD) (Rajpurkar et al., 2016) or JEC-QA (Zhong et al., 2020c) in this
particular case. Moreover,
since these extraction methods can process questions about different domains and
topics, they are
highly reusable.
Concrete examples in the legal domain are ResPubliQA 2010 collection, i.e., a
subset of the JRC-
ACQUIS Multilingual Parallel Corpus and the EUROPARL collection which are
multilingual parallel
collections. The JRC-ACQUIS is a legal corpus representing the complete body of
the European Union
documents, and it is commonly used to build systems. Other interesting examples
are the House of
Lords judgments corpus8, etc. Moreover, some recent works such as (Chalkidis &
Kampas, 2019)
show us one of the clearest examples of how LQA solutions can benefit from
unstructured corpora to
8https://publications.parliament.uk/pa/ld/ldjudgmt.htm
25









build successful systems.
Pros: Information can be easily added or updated
Cons: Reliability of answers is low, paraphrasing is frequent, problematic
representation of
unstructured data
3.2. Structured sources
The most popular form of the structured source is Legal Knowledge Bases (LKBs)
(Bench-Capon &
Coenen, 1992). The LKBs are knowledge management systems being able to describe
legal entities and
the relationships between them. Legal entities are usually represented as nodes,
while the relationships
between them are represented as edges. Each legal entity can be classified using
one or more classes.
At the same time, classes can be organized hierarchically. The same is valid for
relationships. Nodes
also stand for literals. Moreover, most LKBs can perform reasoning enabling the
derivation of inferred
facts. Most of ontology-based, Linked Data, and Legal Knowledge Graphs use these
kinds of structured
sources.
For example, an LQA solution consuming knowledge from a LKB might be focused on
translating
a question formulated by a user to a specific query of an LKB, such as in
(Kourtin et al., 2021). This
makes these LQA solutions perfect candidates for answering factoid formulations
with high accuracy.
The reason is that factoid formulations are simple questions that could be
answered with a precise
fact, so the corresponding answer can be obtained merely by knowing an entity
and a property in the
LKB. Nevertheless, many problems may arise when many facts need to be
discovered, connected, and
combined to elaborate.
Furthermore, there may be several reasons why the generation of structured
information is not so
popular in this domain, but in general, it is widely assumed that generating
such information might
be too expensive in terms of resource consumption, it might be subject to many
errors what it would
make it difficult and expensive to be maintained (Martinez-Gil, 2015).
26








Despite these problems, general-purpose datasets to test solutions have been
developed in recent
times. For example, the series of Question Answering over Linked Data (QALD)
(Cimiano et al.,
2013). However, and unfortunately, the development of structured sources of
legal content is much
more limited. Although the process on how the extraction function is usually
learned is well described
in (Zou et al., 2014). The good news is that the use of Knowledge Graphs of a
legal nature is being
investigated to obtain better results (Filtz, 2017; Huang et al., 2020a; Filtz
et al., 2021).
Pros: Reliability of answers is high, not need of complex natural language
processing
Cons: Limited information stored, sourcing is error-prone and expensive to
build, difficult
interoperability in data sources
4. The problem of the interpretability
Historically, the legal sector is as interested in knowing the answer to a
question as in knowing
its associated explanation. It is something inherent to this domain that
differentiates it from many
disciplines where a black-box would have no significant problems operating. In
the legal industry,
apart from the scenario whereby the system is assumed to have the required
information and the user
is willing to accept the answers, another critical factor comes into play: the
need for explanation.
Therefore, the problem of interpretability is highly relevant for legal
professionals.
Methods and techniques for answering specific questions are in high demand, and
as a result,
several solutions for LQA have been developed to respond to this need. System
designers claim that
the capability to answer questions through computers automatically could help
alleviate a problem
involving tedious tasks such as an extensive information search that is, in
general, time-consuming.
Moreover, therefore, by automatically providing hints concerning a vast number
of legal topics, lots
of resources in the form of effort, money, and time could be saved.
The problem is that designing a LQA solution with a high degree of accuracy,
interpretability,
and performance at the same time is far from being trivial. So one often has to
choose at most
27












two characteristics out of the three possible ones. Interpretability being, in
the legal sector, almost
a mandatory requirement. In addition, if it is taken into account that, as we
have seen, black-box
solutions currently dominate this application domain, it is possible to deduce
that one of the root
causes that is limiting the LQA growth is the fact that the automatic generation
of explanations is
still not very satisfactory for the legal professionals.
There are many proposals to work with legal texts; some of these proposals are
based on variations
of the concept of the distributional assumption terms that appear in a similar
context base, the
calculation of synonyms, others are based on the co-occurrence of words in the
same textual corpus,
different neural architectures, different legal sources, etc. In principle, it is
challenging to discern
which approach could perform better than the others. This always depends on the
use case and the
context in which they are applied. But there is at least one factor that can be
known in advance: the
solution’s interpretability. That is, a human operator can fully understand the
model used to answer
questions. A system that is worthy of people’s trust must not only be effective
and efficient; it must
also be able to communicate the reason for the outputs it produces (Atkinson et
al., 2020).
If we did not have a fundamental requirement, the interpretability of the
solution, it is clear that
the family of neural solutions might be a perfect candidate to build the best
LQA solutions to date.
However, understanding the model, which is an essential aspect in the legal
domain, is not the only
obstacle. For example, finding large data sets representing solved cases is very
difficult or expensive,
or the neural model is usually difficult to be reused in problems of a similar
nature. In this respect,
some of the latest IR systems are superior because they do not require training,
and they can always
be reused over any corpora.
For all these reasons we have detailed, explaining the answers offered is very
important. Moreover,
explanations should be contrastive. This means that in addition to explaining
why a particular
outcome has been offered, a good explanation should communicate why other
outcomes have been
discarded. According to recent work (Doshi-Velez & Kim, 2017), the
interpretability of a model is
28






measured according to three levels:
•Application level interpretability is the level at which a domain expert could
understand how the
model works. Therefore, a good knowledge of the legal area and how the quality
of the result
can be determined is needed. This level should be sufficient for specialized
professionals in the
sector.
•Human level interpretability is a more superficial level than the application
level because it is
not necessary to rely on an expert to examine the model but an average person.
This makes the
experiments cheaper and more straightforward because it is possible to find many
more people
for the validation phase of the model. In addition, it could help the widespread
adoption of such
solutions.
•Function level interpretability does not need people. It is assumed that anyone
could interpret
the resulting model. For example, everyone can understand how decision trees
work. However,
the limitation of functional level interpretability when working with text is
limited (Martinez-Gil
& Chaves-Gonzalez, 2020).
To start working, a system with good interpretability at the application level
should be sufficient.
At least for a professional in the sector. However, this is often not enough
because not all professionals
are specialized in the same legal field. So the aim here is to reach high levels
of interpretability
regarding the application level, human level, and function level at the same
time, as well as the need
for many fewer resources for training and the possibility of a simple transfer
learning stage. If not,
the user will not necessarily assume that the system has a good understanding of
the domain, which
could be used to explain the predictions. As a result, it will choose not to use
it.
This create big problems since the simplest regression-based models are easy for
users to interpret.
However, they do not lead to the best results. The opposite is true for models
of a neural nature.
Therefore, accuracy and interpretability are often considered orthogonal goals.
Therefore, building a
solution that reconciles the two features is far from being trivial.
29




To date, no fully interpretable LQA solution delivery highly accurate results
have been achieved.
However, the use of Legal Knowledge Graphs seems promising because they combine
the compu-
tational power of machine learning with the structured nature of knowledge bases
(in the form of
graphs). Time will tell if this family of methods can deliver what it promises.
In the meantime, users
of highly accurate neural LQA solutions can count on some frameworks that allow
them to get great
clues about how systems make the decisions they do (Ribeiro et al., 2016;
Lundberg & Lee, 2017).
5. Evaluation
Several datasets Sovrano et al. (2021) and performance metrics have been
proposed to evaluate
the results of a given QA system. According to Rodrigo et al. (Rodrigo & Pe˜nas,
2017) the method
to obtain the final score depends on the evaluation measure selected. Each
measure evaluates a
different set of features. Hence, researchers must select the measure depending
on the objectives of
the evaluation.
The most relevant metrics to evaluate the quality of LQA solutions are the
following: Without
ranking, ranking by question, ranking for all questions, cardinal confidence
self-score, and other met-
rics, for example, to evaluate the quality of the artificially generated text. We
are going to see them
in detail below.
5.1. Without Ranking
Although accuracy is a simple and intuitive measure, it only acknowledges
correct answers and
does not consider the amount of incorrect answers. However, its application is
very popular because
it is easy to understand.
accuracy =correct answers
total answers (1)
30










5.2. Ranking by question
When the system offers several answers in order of priority and what is of
interest is to average
the correct answer in the ordered list, the following Mean Average Precision
(MAP) is usually used.
The MAP compares the ground truth of correct answers to the answers offered by
the system and
returns a score. The higher the score, the more accurate the LQA solution is.
MAP =PQ
q=1 AveP (q)
Q(2)
whereby Qis the number of questions and AveP is defined as
AveP =Pn
k=1 P(k)×rel(k)
number of relevant documents (3)
whereby rel(k) is a function returning 1 if the answer is correct
It is also usual to use the Mean Reciprocal Rank (MRR). MRR is a statistic
measure for evaluating
any process that produces a list of possible answers to a set of queries,
ordered by the probability of
correctness.
MRR =1
|Q|
|Q|
X
i=1
1
ranki
(4)
When there is a list of questions, precision and recall are evaluated based on
the complete list
of known distinct instances of the answers. Precision is the fraction of answers
that are correct in
relation to the found ones:
precision ={f ound answers} ∩ {correct answers}
{found answers}(5)
The recall is the fraction of the found answers in relation to the correct ones.
31


recall ={f ound answers} ∩ {correct answers}
{correct answers}(6)
F-measure combines precision and recall into a harmonic mean to summarize both
metrics.
f−measure =2·precision ·recall
precision +recall (7)
Precision can be optimized at the expense of recall and vice versa. For this
reason, it is convenient
to report the two measures together, or at least, the combination of them
through an F-measure.
5.3. Ranking for all questions
The most popular metric for this scenario is the so-called Confidence Weighted
Score (CWS),
which is based on the notion of average precision. CWS requires LQA solution to
return only one
answer per question and rank all the answers according to the background truth.
Then, the CWS
metrics reward solutions returning correct answers at top positions in the
ranking.
CW S =1
|Q|
|Q|
X
i=1
C(i)
i(8)
whereby Qis the number of questions and Cis defined as
C=
i
X
j=1
I(j) (9)
being I(j) the function that returns 1 if answer j is correct and 0 if it is
not.
5.4. Cardinal Confidence Self-Score
The most popular metrics in this context are K and K1 (Peters et al., 2005). K
and K1 are based
on a utility function that returns -1 if the answer is incorrect and 1 if it is
correct. Both measures
weigh this value with the confidence score given by the LQA solution. A positive
value does not
32




indicate more correct answers than incorrect ones, but that the sum of scores
from correct answers is
higher than the sum of scores from incorrect answers (Rodrigo & Pe˜nas, 2017).
5.5. Unanswered different from incorrect
In the context of LQA solutions, most of the time, it is preferable not to
respond incorrectly
(Pe˜nas & Rodrigo, 2011). This is because it is usually unsuitable for confusing
the user with incorrect
answers that undermine the system’s trustworthiness. This idea is not new, but
despite several
previous attempts, there is no commonly accepted measure to assess non-response.
To deal with this situation, the metric Correctness at one c@1 has been
proposed.
c@1 = 1
total questions ·(correct answers +correct answers
total questions ·unanswered questions) (10)
5.6. Other metrics
In recent times, new alternative metrics have emerged to evaluate the quality of
responses (Chen
et al., 2019). These metrics are of particular relevance for some families of
LQA construction methods,
especially those that aim to generate a textual sentence and not just answer a
factual question. A
summary of these alternative metrics is shown below.
5.6.1. BERT Score and its variants
BERTScore is a family of alternative metrics that first tries to get the BERT
embeddings (Devlin
et al., 2019) of each term and calculate the current answer and the ground truth
through a BERT
model separately. Then, a mapping is computed between candidate and reference
terms by means of
pairwise cosine similarity. This mapping is then aggregated into precision and
recall scores, and then
into a harmonic mean.
33










5.6.2. Sentence Mover’s Similarity
SMS is another alternative metric based on mover’s distance to evaluate textual
information such as
machine-generated text. Its application in LQA assessment is done by firstly
computing an embedding
for each sentence in an answer as an average of the ELMo embeddings (Peters et
al., 2018). Then, a
mapping function is used to obtain the distance of moving a candidate answer’s
sentences to match
the reference answer.
5.6.3. N-gram based metrics
These are metrics developed for evaluating machine translation, e.g. BLEU,
ROUGE, etc. The
key idea is to evaluate a candidate sentence by determining the n-grams in the
candidate sentence
that also appear in the reference sentence (Chen et al., 2019). In the context
of LQA solutions, these
metrics can be adapted to measure the quality of the generated answer in
relation to the answer
specified in the ground truth.
5.7. Summary
In the following, we show a summary of the different families of methods to build
LQAs and the
most appropriate ways to automatically determine their quality. Table 10 shows a
tabular summary
of these correspondences.
For example, metrics such as BERTScore, SMS, and N-Gram do not make sense in the
Yes/No
Answer or LMCQA family of methods since the answers are known in advance and are
unnecessary
to generate them. Alternatively, when working with ontologies, the measure of
recall is much more
important than precision. Because if the answer was wrong, it is due to the
access to a wrong ontology
model rather than to the reasoning algorithm used to obtain the answer.
Moreover, ranking measures
do not make much sense either.
34






Table 10: Most useful evaluation methods to determine the accuracy of LQA
solutions
Family of Methods Evaluation
Yes/No Answers Acc, KK1, c@1
LMCQA Acc, PR, MAP, MRR, CWS, KK1, c@1
Classical IR-based solutions Acc, PR, MAP, MRR, CWS, KK1, c@1, BScore, SMS,
NGram
Big Data solutions Acc, PR, MAP, MRR, CWS, KK1, c@1
Ontology-based solutions Acc, Recall, KK1, c@1
Neural solutions Acc, PR, MAP, MRR, CWS, KK1, c@1, BScore, SMS, NGram
Other approaches Acc, PR, MAP, MRR, CWS, KK1, c@1, BScore, SMS, NGram
6. Open Research Challenges
Given the state-of-the-art and the open problems that need to be investigated,
it is possible to
identify some significant gaps that should be filled to facilitate the adoption of
LQA solutions at
scale. Among the Open Research Challenges (ORC) that can be identified are the
lack of proposals
for transfer learning in the legal sector, the lack of solutions to deal with
the complicated nuances of
legal language, the improvement of the integration of distributed and
heterogeneous data sources, as
well as the importance of the multilingual capabilities of the different
solutions. In the following, we
review each of them.
6.1. ORC1. Transfer Learning
One important concept related to knowledge representation is Transfer Learning.
These methods
are based on the possibility of using a given model to export it to other
scenarios of analogous
nature. In other words, the knowledge representation that has been learned to
complete one task
can be generalized to help complete other tasks. A suitable knowledge
representation method must
determine which factors and features will be exploited and thus reuse them in
another task.
Most traditional machine learning approaches can be used to build models capable
of addressing
35


various problems given enough data and time. However, in practice, the amount of
data and time
available are frequently limited. For this reason, transfer learning has
received considerable attention
from such communities as Deep Learning and Big Data communities to date. These
communities’
problems are very resource-intensive in the form of data and time. This fact
explains why strategies
of this kind are often seen as a way to alleviate such issues.
To date, transfer learning methods in the field of LQA solutions have been only
superficially
explored. Only the work of Yan et al. has attempted to bring solutions to the
table (Yan et al., 2020).
It will be necessary for the community to intensify its research in this area in
the near future. It is a
great way to boost the accuracy of solutions while reusing training data and
computational resources.
6.2. ORC2. Nuances of legal language
While in many application domains of QA systems, there is an apparent problem of
ambiguity,
i.e., most of the time, the questions are formulated so that a single statement
may assume different
meanings. In legal language, the problem is just the opposite. The language is
so precise that it
cannot give rise to multiple interpretations. Although this may seem a
facilitator a priori, this fact
brings a series of disadvantages that should be faced.
Classical stemming techniques cannot be freely used to obtain the root of the
terms. Thus, for
example, the sentence the contract is void is very different from the sentence
the contract is voidable.
Furthermore, even though the terms void and voidable contain the same root, a
similar meaning could
not be assumed.
This forces us to have optimal performance methods to really understand the
meaning of the
information being processed. In the near future, we will have to continue
working in this direction.
This would be facilitated by exploring legal texts’ inherent characteristics to
utilize these features for
building LQA solutions. For example, properties such as references between
documents or structured
relationships in legal statements should be investigated since they are often a
great help in processing
those different nuances that legal language brings.
36




6.3. ORC3. Distributed Knowledge
We have seen that although approaches based on structured corpora often yield
good results, it is
often difficult to use them in practice mainly due to the cost when building such
structures (i.e., it
is expensive in terms of effort, money, and time needed) and it is often
complicated to find experts
with enough knowledge for curating them. Therefore, it is usually desirable to
access all possible
sources to maximize the chances of success in answering a question while
minimizing the expenditure
of resources.
The problem is that the different data representations in the sources make
automatic aggregation,
interlinking, and integration very difficult. Therefore, the development of
efficient methods to prop-
erly access differently structured sources of legal nature, which can serve as a
knowledge source for
computer systems in search of the ideal answer to a given question, is a
critical challenge that should
be faced in the near future.
For these reasons, further research is necessary to make the task easier, by
different interlinking
corpora based on diverse matching methodologies or by transforming and properly
integrating other
data source types into novel knowledge models, e.g., linked data, knowledge,
graphs, etc. using some
semantic labeling techniques, or even creating a unique global representation
such as the universal
schema such as the one proposed in (Riedel et al., 2013).
6.4. ORC4. Multilingualism
Working with several languages simultaneously is a recurring problem, not only
in the design of
LQA solutions. Nevertheless, the truth is that in this domain, it is of
fundamental importance since
there is important legislation and regulations that are not always written in
the same language spoken
in the territories where they apply.
The good news in this regard is the latest advances in word vectorization, which
brings with it the
possibility of working with abstract representations that are
language-independent (Grave et al., 2018).
37






In this sense, representing words by vectors can be considered one of the main
recent breakthroughs
in deep learning for natural language processing (Goldberg, 2017). The benefits
of adopting vectors
are multiple since, in addition to working with abstract representations of
words, they have other
associated advantages, such as low computational complexity, because these
vectors are efficient to
compute using simple operations.
The advances in word representation have a substantial impact on semantic
analysis. The future
semantic analysis models of legal text will improve LQA performance thanks to
various deep learning
methods to build multilingual word embeddings from unstructured text. However,
to achieve a better
performance, it is necessary to explore practical approaches describing the
meaning of textual pieces,
such as sentences, paragraphs, or even documents, from a legal perspective.
7. Conclusions
QA technology is becoming an essential solution in many areas overloaded by the
constant gener-
ation of large amounts of information. Automatically answering specific questions
can contribute to
alleviating the problem of dealing with those vast amounts of data. In the legal
domain, good LQA
solutions are in high demand mainly due to the problems of information overload
that legal profession-
als have to face in their daily activities. As a result, several solutions for
LQA have been developed as
a response to that situation. The primary reason for that is that the capability
to answer questions
through computers automatically could help alleviate a problem involving tedious
tasks such as an
extensive information search that is, in general, time-consuming and
error-prone.
We have surveyed the different LQA solutions that have been proposed in the
literature to date.
To do that, we have proceeded with the analysis of the essential methods in LQA
over the last
decades. We have identified the data sources on which existing LQA solutions have
been implemented
to automatically answer questions of legal nature. We have seen that there is no
such solution that
can provide high accuracy, interpretability, and performance at the same time.
Nevertheless, there are
38




usually strategies capable of optimizing two of them at the expense of the
other, being interpretability
of significant importance in the legal domain since the community has always
found more useful
systems whereby a few hundredths more precision does not compensate for the
other disadvantages of
interpretability and performance. Furthermore, we have made an overview of the
most popular ways
to determine the quality of LQA solutions from a strictly quantitative point of
view. Last but not
least, we have made a review of the research challenges that are still open
today and that will have
to be faced by the community soon.
This concludes this survey of more than two decades of research efforts in the
field of LQA.
It is expected that soon, more and more proposals for highly accurate,
interpretable, and efficient
QA systems specialized in legal matters will see the light of day. It seems also
reasonable to think
that LQA solutions will not replace legal experts and their unique, specialized
knowledge. However,
all indications are that these solutions will undoubtedly and significantly
transform the traditional
delivery of legal services in the near future.
Competing interest
The author has no competing interest to declare.
Acknowledgments
This research work has been partially supported by the Austrian Ministry for
Transport, Innovation
and Technology, the Federal Ministry of Science, Research and Economy, and the
Province of Upper
Austria in the frame of the COMET center SCCH.
39


References
Atkinson, K., Bench-Capon, T. J. M., & Bollegala, D. (2020). Explanation in AI
and law: Past, present
and future. Artif. Intell.,289 , 103387. URL:
https://doi.org/10.1016/j.artint.2020.103387.
doi:10.1016/j.artint.2020.103387.
Aydin, B. I., Yilmaz, Y. S., Li, Y., Li, Q., Gao, J., & Demirbas, M. (2014).
Crowd-
sourcing for multiple-choice question answering. In C. E. Brodley, & P. Stone
(Eds.),
Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, July
27 -
31, 2014, Qu´ebec City, Qu´ebec, Canada (pp. 2946–2953). AAAI Press. URL:
http://www.aaai.org/ocs/index.php/IAAI/IAAI14/paper/view/8634.
Bach, N. X., Cham, L. T. N., Thien, T. H. N., & Phuong, T. M. (2017). Question
analysis for
vietnamese legal question answering. In T. T. Nguyen, A. Le, S. Tojo, L. Nguyen,
& X. Phan (Eds.),
9th International Conference on Knowledge and Systems Engineering, KSE 2017,
Hue, Vietnam,
October 19-21, 2017 (pp. 154–159). IEEE. URL:
https://doi.org/10.1109/KSE.2017.8119451.
doi:10.1109/KSE.2017.8119451.
Bench-Capon, T. J. M., & Coenen, F. (1992). The maintenance of legal knowledge
based
systems. Artif. Intell. Rev.,6, 129–143. URL:
https://doi.org/10.1007/BF00150230.
doi:10.1007/BF00150230.
Bennett, Z., Russell-Rose, T., & Farmer, K. (2017). A scalable approach to legal
question an-
swering. In J. Keppens, & G. Governatori (Eds.), Proceedings of the 16th edition
of the Inter-
national Conference on Artificial Intelligence and Law, ICAIL 2017, London,
United Kingdom,
June 12-16, 2017 (pp. 269–270). ACM. URL:
https://doi.org/10.1145/3086512.3086547.
doi:10.1145/3086512.3086547.
Br¨uninghaus, S., & Ashley, K. D. (2001). Improving the representation of legal
case texts with infor-
mation extraction methods. In R. P. Loui (Ed.), Proceedings of the Eigths
International Conference
40











on Artificial Intelligence and Law, ICAIL 2001, St. Louis, Missouri, USA, May
21-25, 2001 (pp.
42–51). ACM. URL: https://doi.org/10.1145/383535.383540.
doi:10.1145/383535.383540.
Carvalho, D. S., Tran, V. D., Van Tran, K., Lai, V. D., & Nguyen, M.-L. (2016).
Lexical to discourse-
level corpus modeling for legal question answering. In Tenth International
Workshop on Juris-
Informatics (JURISIN).
Cer, D., Yang, Y., Kong, S., Hua, N., Limtiaco, N., John, R. S., Constant, N.,
Gua jardo-Cespedes,
M., Yuan, S., Tar, C., Strope, B., & Kurzweil, R. (2018). Universal sentence
encoder for en-
glish. In E. Blanco, & W. Lu (Eds.), Proceedings of the 2018 Conference on
Empirical Methods
in Natural Language Processing, EMNLP 2018: System Demonstrations, Brussels,
Belgium, Oc-
tober 31 - November 4, 2018 (pp. 169–174). Association for Computational
Linguistics. URL:
https://doi.org/10.18653/v1/d18-2029. doi:10.18653/v1/d18-2029.
Chalkidis, I., & Kampas, D. (2019). Deep learning in law: early adaptation and
le-
gal word embeddings trained on large corpora. Artif. Intell. Law,27 , 171–198.
URL:
https://doi.org/10.1007/s10506-018-9238-9. doi:10.1007/s10506-018-9238-9.
Chen, A., Stanovsky, G., Singh, S., & Gardner, M. (2019). Evaluating question
answering eval-
uation. In A. Fisch, A. Talmor, R. Jia, M. Seo, E. Choi, & D. Chen (Eds.),
Proceedings of
the 2nd Workshop on Machine Reading for Question Answering, MRQA@EMNLP 2019,
Hong
Kong, China, November 4, 2019 (pp. 119–124). Association for Computational
Linguistics. URL:
https://doi.org/10.18653/v1/D19-5817. doi:10.18653/v1/D19-5817.
Chitta, R., & Hudek, A. K. (2019). A reliable and accurate multiple choice
question answering
system for due diligence. In Proceedings of the Seventeenth International
Conference on Artificial
Intelligence and Law, ICAIL 2019, Montreal, QC, Canada, June 17-21, 2019 (pp.
184–188). ACM.
URL: https://doi.org/10.1145/3322640.3326711. doi:10.1145/3322640.3326711.
Cimiano, P., L´opez, V., Unger, C., Cabrio, E., Ngomo, A. N., & Walter, S.
(2013). Multilin-
41












gual question answering over linked data (QALD-3): lab overview. In P. Forner,
H. M¨uller,
R. Paredes, P. Rosso, & B. Stein (Eds.), Information Access Evaluation.
Multilinguality, Multi-
modality, and Visualization - 4th International Conference of the CLEF
Initiative, CLEF 2013,
Valencia, Spain, September 23-26, 2013. Proceedings (pp. 321–332). Springer
volume 8138 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-642-40802-1_30.
doi:10.1007/978-3-642-40802-1\_30.
Collarana, D., Heuss, T., Lehmann, J., Lytra, I., Maheshwari, G., Nedelchev, R.,
Schmidt, T., &
Trivedi, P. (2018). A question answering system on regulatory documents. In
JURIX (pp. 41–50).
Deerwester, S. C., Dumais, S. T., Landauer, T. K., Furnas, G. W., & Harshman, R.
A. (1990).
Indexing by latent semantic analysis. J. Am. Soc. Inf. Sci.,41 , 391–407.
Delfino, P., Cuconato, B., Passos, G. P., Zaverucha, G., & Rademaker, A. (2018).
Using openwordnet-
pt for question answering on legal domain. In F. Bond, P. Vossen, & C. Fellbaum
(Eds.),
Proceedings of the 9th Global Wordnet Conference, GWC 2018, Nanyang
Technological Univer-
sity (NTU), Singapore, January 8-12, 2018 (pp. 105–112). Global Wordnet
Association. URL:
https://www.aclweb.org/anthology/2018.gwc-1.13/.
Devlin, J., Chang, M., Lee, K., & Toutanova, K. (2019). BERT: pre-training of
deep bidirectional transformers for language understanding, . (pp. 4171–4186).
URL:
https://doi.org/10.18653/v1/n19-1423. doi:10.18653/v1/n19-1423.
Diefenbach, D., L´opez, V., Singh, K. D., & Maret, P. (2018). Core techniques of
question
answering systems over knowledge bases: a survey. Knowl. Inf. Syst.,55 ,
529–569. URL:
https://doi.org/10.1007/s10115-017-1100-y. doi:10.1007/s10115-017-1100-y.
Dimitrakis, E., Sgontzos, K., & Tzitzikas, Y. (2020). A survey on question
answer-
ing systems over linked data and documents. J. Intell. Inf. Syst.,55 , 233–259.
URL:
https://doi.org/10.1007/s10844-019-00584-7. doi:10.1007/s10844-019-00584-7.
42











Do, P., Nguyen, H., Tran, C., Nguyen, M., & Nguyen, M. (2017). Legal question
answer-
ing using ranking SVM and deep convolutional neural network. CoRR,abs/1703.05320
. URL:
http://arxiv.org/abs/1703.05320.arXiv:1703.05320.
Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable
machine learning.
arXiv preprint arXiv:1702.08608 , .
Fawei, B., Pan, J. Z., Kollingbaum, M. J., & Wyner, A. Z. (2018). A methodology
for a criminal law
and procedure ontology for legal question answering. In R. Ichise, F. L´ecu´e,
T. Kawamura, D. Zhao,
S. Muggleton, & K. Kozaki (Eds.), Semantic Technology - 8th Joint International
Conference, JIST
2018, Awaji, Japan, November 26-28, 2018, Proceedings (pp. 198–214). Springer
volume 11341 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-030-04284-4_14.
doi:10.1007/978-3-030-04284-4\_14.
Fawei, B., Pan, J. Z., Kollingbaum, M. J., & Wyner, A. Z. (2019). A
semi-automated on-
tology construction for legal question answering. New Gener. Comput.,37 ,
453–478. URL:
https://doi.org/10.1007/s00354-019-00070-2. doi:10.1007/s00354-019-00070-2.
Ferrucci, D. A., Levas, A., Bagchi, S., Gondek, D., & Mueller, E. T. (2013).
Watson: Beyond jeop-
ardy! Artif. Intell.,199-200 , 93–105. URL:
https://doi.org/10.1016/j.artint.2012.06.009.
doi:10.1016/j.artint.2012.06.009.
Filtz, E. (2017). Building and processing a knowledge-graph for legal data. In
E. Blomqvist,
D. Maynard, A. Gangemi, R. Hoekstra, P. Hitzler, & O. Hartig (Eds.), The
Semantic Web -
14th International Conference, ESWC 2017, Portoroˇz, Slovenia, May 28 - June 1,
2017, Pro-
ceedings, Part II (pp. 184–194). volume 10250 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-319-58451- 5_13.
doi:10.1007/978-3-319-58451-5\_13.
Filtz, E., Kirrane, S., & Polleres, A. (2021). The linked legal data landscape:
linking legal data across
different countries. Artificial Intelligence and Law, (pp. 1–55).
43












Franco, W., Avila, C. V. S., Oliveira, A., Maia, G., Brayner, A., Vidal, V. M.
P., Carvalho,
F., & Pequeno, V. M. (2020). Ontology-based question answering systems over
knowledge
bases: A survey. In J. Filipe, M. Smialek, A. Brodsky, & S. Hammoudi (Eds.),
Proceed-
ings of the 22nd International Conference on Enterprise Information Systems,
ICEIS 2020,
Prague, Czech Republic, May 5-7, 2020, Volume 1 (pp. 532–539). SCITEPRESS. URL:
https://doi.org/10.5220/0009392205320539. doi:10.5220/0009392205320539.
Goldberg, Y. (2017). Neural Network Methods for Natural Language Process-
ing. Synthesis Lectures on Human Language Technologies. Morgan & Clay-
pool Publishers. URL: https://doi.org/10.2200/S00762ED1V01Y201703HLT037.
doi:10.2200/S00762ED1V01Y201703HLT037.
Grave, E., Bojanowski, P., Gupta, P., Joulin, A., & Mikolov, T. (2018). Learning
word vectors for
157 languages. arXiv preprint arXiv:1802.06893 , .
Han, L., Kashyap, A. L., Finin, T., Mayfield, J., & Weese, J. (2013). Umbc
ebiquity-core: Se-
mantic textual similarity systems. In M. T. Diab, T. Baldwin, & M. Baroni
(Eds.), Proceedings
of the Second Joint Conference on Lexical and Computational Semantics, *SEM
2013, June 13-
14, 2013, Atlanta, Georgia, USA (pp. 44–52). Association for Computational
Linguistics. URL:
https://www.aclweb.org/anthology/S13-1005/.
He, S., Liu, S., Chen, Y., Zhou, G., Liu, K., & Zhao, J. (2013). Casia@qald-3: A
question answering system over linked data. In P. Forner, R. Navigli, D. Tufis, &
N. Ferro (Eds.), Working Notes for CLEF 2013 Conference , Valencia, Spain,
Septem-
ber 23-26, 2013 . CEUR-WS.org volume 1179 of CEUR Workshop Proceedings. URL:
http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-HeEt2013.pdf.
H¨offner, K., Walter, S., Marx, E., Usbeck, R., Lehmann, J., & Ngomo, A. N.
(2017). Survey
on challenges of question answering in the semantic web. Semantic Web,8,
895–920. URL:
https://doi.org/10.3233/SW-160247. doi:10.3233/SW-160247.
44










Holzenberger, N., Blair-Stanek, A., & Van Durme, B. (2020). A dataset for
statutory reasoning in
tax law entailment and question answering. arXiv preprint arXiv:2005.05257, .
Hoshino, R., Taniguchi, R., Kiyota, N., & Kano, Y. (2018). Question answering
system for le-
gal bar examination using predicate argument structure. In K. Kojima, M.
Sakamoto, K. Mi-
neshima, & K. Satoh (Eds.), New Frontiers in Artificial Intelligence - JSAI-isAI
2018 Workshops,
JURISIN, AI-Biz, SKL, LENLS, IDAA, Yokohama, Japan, November 12-14, 2018,
Revised Se-
lected Papers (pp. 207–220). Springer volume 11717 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-030-31605- 1_16.
doi:10.1007/978-3-030-31605-1\_16.
Huang, W., Jiang, J., Qu, Q., & Yang, M. (2020a). AILA: A question answering
system
in the legal domain. In C. Bessiere (Ed.), Proceedings of the Twenty-Ninth
International
Joint Conference on Artificial Intelligence, IJCAI 2020 (pp. 5258–5260).
ijcai.org. URL:
https://doi.org/10.24963/ijcai.2020/762. doi:10.24963/ijcai.2020/762.
Huang, W., Liao, X., Xie, Z., Qian, J., Zhuang, B., Wang, S., & Xiao, J.
(2020b). Generating reason-
able legal text through the combination of language modeling and question
answering. In C. Bessiere
(Ed.), Proceedings of the Twenty-Ninth International Joint Conference on
Artificial Intelligence,
IJCAI 2020 (pp. 3687–3693). ijcai.org. URL:
https://doi.org/10.24963/ijcai.2020/510.
doi:10.24963/ijcai.2020/510.
John, A., Caro, L. D., Boella, G., & Bartolini, C. (2016). An approach to in-
formation retrieval and question answering in the legal domain. In Proceedings
of the 10th International Workshop on Juris-informatics (JURISIN 2016). URL:
http://orbilu.uni.lu/bitstream/10993/29385/1/paper%2018.pdf.
Kano, Y., Hoshino, R., & Taniguchi, R. (2017). Analyzable legal yes/no question
answering system
using linguistic structures. In K. Satoh, M. Kim, Y. Kano, R. Goebel, & T.
Oliveira (Eds.), COL-
IEE 2017. 4th Competition on Legal Information Extraction and Entailment, held
in conjunction
with the 16th International Conference on Artificial Intelligence and Law (ICAIL
2017) in King’s
45









College London, UK (pp. 57–67). EasyChair volume 47 of EPiC Series in Computing
. URL:
https://easychair.org/publications/paper/zdP.
Kien, P. M., Nguyen, H., Bach, N. X., Tran, V., Nguyen, M. L., & Phuong, T. M.
(2020). Answering
legal questions by learning neural attentive text representation. In D. Scott,
N. Bel, & C. Zong
(Eds.), Proceedings of the 28th International Conference on Computational
Linguistics, COLING
2020, Barcelona, Spain (Online), December 8-13, 2020 (pp. 988–998).
International Commit-
tee on Computational Linguistics. URL:
https://doi.org/10.18653/v1/2020.coling-main.86.
doi:10.18653/v1/2020.coling-main.86.
Kim, M., & Goebel, R. (2017). Two-step cascaded textual entailment for legal bar
exam question
answering. In J. Keppens, & G. Governatori (Eds.), Proceedings of the 16th
edition of the In-
ternational Conference on Artificial Intelligence and Law, ICAIL 2017, London,
United Kingdom,
June 12-16, 2017 (pp. 283–290). ACM. URL:
https://doi.org/10.1145/3086512.3086550.
doi:10.1145/3086512.3086550.
Kim, M., Lu, Y., & Goebel, R. (2017). Textual entailment in legal bar exam
question answering using
deep siamese networks. In S. Arai, K. Kojima, K. Mineshima, D. Bekki, K. Satoh,
& Y. Ohta
(Eds.), New Frontiers in Artificial Intelligence - JSAI-isAI Workshops, JURISIN,
SKL, AI-Biz,
LENLS, AAA, SCIDOCA, kNeXI, Tsukuba, Tokyo, Japan, November 13-15, 2017, Revised
Se-
lected Papers (pp. 35–48). Springer volume 10838 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-319-93794- 6_3.
Kim, M., Xu, Y., & Goebel, R. (2014). Legal question answering using ranking SVM
and syn-
tactic/semantic similarity. In T. Murata, K. Mineshima, & D. Bekki (Eds.), New
Frontiers in
Artificial Intelligence - JSAI-isAI 2014 Workshops, LENLS, JURISIN, and GABA,
Kanagawa,
Japan, October 27-28, 2014, Revised Selected Papers (pp. 244–258). Springer
volume 9067 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-662-48119-6_18.
doi:10.1007/978-3-662-48119-6\_18.
46










Kim, M., Xu, Y., & Goebel, R. (2015). Applying a convolutional neural network to
legal ques-
tion answering. In M. Otake, S. Kurahashi, Y. Ota, K. Satoh, & D. Bekki (Eds.),
New Fron-
tiers in Artificial Intelligence - JSAI-isAI 2015 Workshops, LENLS, JURISIN, AAA,
HAT-
MASH, TSDAA, ASD-HR, and SKL, Kanagawa, Japan, November 16-18, 2015, Revised Se-
lected Papers (pp. 282–294). volume 10091 of Lecture Notes in Computer Science.
URL:
https://doi.org/10.1007/978-3-319-50953- 2_20.
doi:10.1007/978-3-319-50953-2\_20.
Kim, M., Xu, Y., Goebel, R., & Satoh, K. (2013). Answering yes/no questions in
legal bar ex-
ams. In Y. I. Nakano, K. Satoh, & D. Bekki (Eds.), New Frontiers in Artificial
Intelligence
- JSAI-isAI 2013 Workshops, LENLS, JURISIN, MiMI, AAA, and DDS, Kanagawa, Japan,
October 27-28, 2013, Revised Selected Papers (pp. 199–213). Springer volume 8417
of Lec-
ture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-319-10061-6_14.
doi:10.1007/978-3-319-10061-6\_14.
Kim, M., Xu, Y., Lu, Y., & Goebel, R. (2016). Question answering of bar exams by
para-
phrasing and legal text analysis. In S. Kurahashi, Y. Ohta, S. Arai, K. Satoh, &
D. Bekki
(Eds.), New Frontiers in Artificial Intelligence - JSAI-isAI 2016 Workshops,
LENLS, HAT-
MASH, AI-Biz, JURISIN and SKL, Kanagawa, Japan, November 14-16, 2016, Revised
Selected
Papers (pp. 299–313). Springer volume 10247 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-319-61572- 1_20.
doi:10.1007/978-3-319-61572-1\_20.
Kolomiyets, O., & Moens, M. (2011). A survey on question answering technol-
ogy from an information retrieval perspective. Inf. Sci.,181 , 5412–5434. URL:
https://doi.org/10.1016/j.ins.2011.07.047. doi:10.1016/j.ins.2011.07.047.
Kourtin, I., Mbarki, S., & Mouloudi, A. (2021). A legal question answering
ontology-based system. In
Formalising Natural Languages: Applications to Natural Language Processing and
Digital Humani-
ties: 14th International Conference, NooJ 2020, Zagreb, Croatia, June 5–7, 2020,
Revised Selected
Papers 14 (pp. 218–229). Springer International Publishing.
47










van Kuppevelt, D. E., & van Dijck, G. (2017). Answering legal research questions
about dutch
case law with network analysis and visualization. In A. Z. Wyner, & G. Casini
(Eds.), Legal
Knowledge and Information Systems - JURIX 2017: The Thirtieth Annual Conference,
Lux-
embourg, 13-15 December 2017 (pp. 95–100). IOS Press volume 302 of Frontiers in
Artifi-
cial Intelligence and Applications. URL:
https://doi.org/10.3233/978-1-61499-838-9-95.
doi:10.3233/978-1-61499-838-9-95.
Lame, G. (2004). Using NLP techniques to identify legal ontology components:
Concepts and re-
lations. Artif. Intell. Law,12 , 379–396. URL:
https://doi.org/10.1007/s10506-005-4160-3.
doi:10.1007/s10506- 005-4160-3.
Li, Y., Bandar, Z., & McLean, D. (2003). An approach for measuring semantic
similarity between
words using multiple information sources. IEEE Trans. Knowl. Data Eng.,15 ,
871–882. URL:
https://doi.org/10.1109/TKDE.2003.1209005. doi:10.1109/TKDE.2003.1209005.
Liu, L., & Luo, J. (2018). A question answering system based on deep learning.
In D. Huang,
M. M. Gromiha, K. Han, & A. Hussain (Eds.), Intelligent Computing Methodologies
- 14th
International Conference, ICIC 2018, Wuhan, China, August 15-18, 2018,
Proceedings, Part
III (pp. 173–181). Springer volume 10956 of Lecture Notes in Computer Science.
URL:
https://doi.org/10.1007/978-3-319-95957- 3_19.
doi:10.1007/978-3-319-95957-3\_19.
Lundberg, S. M., & Lee, S. (2017). A unified approach to interpreting model
predictions. In I. Guyon,
U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, & R.
Garnett (Eds.),
Advances in Neural Information Processing Systems 30: Annual Conference on
Neural Information
Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA (pp.
4765–4774). URL:
https://proceedings.neurips.cc/paper/2017/.
Martinez-Gil, J. (2015). Automated knowledge base management: A survey. Com-
put. Sci. Rev.,18, 1–9. URL: https://doi.org/10.1016/j.cosrev.2015.09.001.
doi:10.1016/j.cosrev.2015.09.001.
48













Martinez-Gil, J., & Chaves-Gonzalez, J. M. (2020). A novel method based on
symbolic regres-
sion for interpretable semantic similarity measurement. Expert Syst. Appl.,160,
113663. URL:
https://doi.org/10.1016/j.eswa.2020.113663. doi:10.1016/j.eswa.2020.113663.
Martinez-Gil, J., Freudenthaler, B., & Tjoa, A. M. (2019a). A general frame-
work for multiple choice question answering based on mutual information and
rein-
forced co-occurrence. Trans. Large Scale Data Knowl. Centered Syst.,42, 91–110.
URL:
https://doi.org/10.1007/978-3-662-60531- 8_4. doi:10.1007/978-3-662-60531-8\_4.
Martinez-Gil, J., Freudenthaler, B., & Tjoa, A. M. (2019b). Multiple choice
question answering
in the legal domain using reinforced co-occurrence. In S. Hartmann, J. K¨ung, S.
Chakravarthy,
G. Anderst-Kotsis, A. M. Tjoa, & I. Khalil (Eds.), Database and Expert Systems
Applications
- 30th International Conference, DEXA 2019, Linz, Austria, August 26-29, 2019,
Proceedings,
Part I (pp. 138–148). Springer volume 11706 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-030-27615- 7_10.
doi:10.1007/978-3-030-27615-7\_10.
Maxwell, K. T., & Schafer, B. (2008). Concept and context in legal information
retrieval. In
E. Francesconi, G. Sartor, & D. Tiscornia (Eds.), Legal Knowledge and
Information Systems -
JURIX 2008: The Twenty-First Annual Conference on Legal Knowledge and
Information Systems,
Florence, Italy, 10-13 December 2008 (pp. 63–72). IOS Press volume 189 of
Frontiers in Arti-
ficial Intelligence and Applications. URL:
https://doi.org/10.3233/978-1-58603-952-3-63.
doi:10.3233/978-1-58603-952-3-63.
McElvain, G., Sanchez, G., Matthews, S., Teo, D., Pompili, F., & Custis, T.
(2019a). West-
search plus: A non-factoid question-answering system for the legal domain. In B.
Pi-
wowarski, M. Chevalier, ´
E. Gaussier, Y. Maarek, J. Nie, & F. Scholer (Eds.), Proceedings
of the 42nd International ACM SIGIR Conference on Research and Development in
Informa-
tion Retrieval, SIGIR 2019, Paris, France, July 21-25, 2019 (pp. 1361–1364).
ACM. URL:
https://doi.org/10.1145/3331184.3331397. doi:10.1145/3331184.3331397.
49












McElvain, G., Sanchez, G., Teo, D., & Custis, T. (2019b). Non-factoid question
answering in the
legal domain. In B. Piwowarski, M. Chevalier, ´
E. Gaussier, Y. Maarek, J. Nie, & F. Scholer (Eds.),
Proceedings of the 42nd International ACM SIGIR Conference on Research and
Development in
Information Retrieval, SIGIR 2019, Paris, France, July 21-25, 2019 (pp.
1395–1396). ACM. URL:
https://doi.org/10.1145/3331184.3331431. doi:10.1145/3331184.3331431.
Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013).
Distributed representations
of words and phrases and their compositionality. In Advances in Neural
Information Processing
Systems 26: 27th Annual Conference on Neural Information Processing Systems
2013. Proceedings
of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States. (pp.
3111–3119).
Mimouni, N., Nazarenko, A., & Salotti, S. (2017). Answering complex queries on
legal networks:
A direct and a structured IR approaches. In U. Pagallo, M. Palmirani, P.
Casanovas, G. Sar-
tor, & S. Villata (Eds.), AI Approaches to the Complexity of Legal Systems -
AICOL Inter-
national Workshops 2015-2017: AICOL-VI@JURIX 2015, AICOL-VII@EKAW 2016, AICOL-
VIII@JURIX 2016, AICOL-IX@ICAIL 2017, and AICOL-X@JURIX 2017, Revised Selected
Pa-
pers (pp. 451–464). Springer volume 10791 of Lecture Notes in Computer Science.
URL:
https://doi.org/10.1007/978-3-030-00178- 0_31.
doi:10.1007/978-3-030-00178-0\_31.
Monroy, A., Calvo, H., & Gelbukh, A. F. (2008). Using graphs for shallow
question answering
on legal documents. In A. F. Gelbukh, & E. F. Morales (Eds.), MICAI 2008:
Advances in
Artificial Intelligence, 7th Mexican International Conference on Artificial
Intelligence, Atizap´an
de Zaragoza, Mexico, October 27-31, 2008, Proceedings (pp. 165–173). Springer
volume 5317 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-540-88636-5_15.
doi:10.1007/978-3-540-88636-5\_15.
Monroy, A., Calvo, H., & Gelbukh, A. F. (2009). NLP for shallow question
answering of legal
documents using graphs. In A. F. Gelbukh (Ed.), Computational Linguistics and
Intelligent Text
Processing, 10th International Conference, CICLing 2009, Mexico City, Mexico,
March 1-7, 2009.
50








Proceedings (pp. 498–508). Springer volume 5449 of Lecture Notes in Computer
Science. URL:
https://doi.org/10.1007/978-3-642-00382- 0_40.
doi:10.1007/978-3-642-00382-0\_40.
Morimoto, A., Kubo, D., Sato, M., Shindo, H., & Matsumoto, Y. (2017). Legal
question answering
system using neural attention. In K. Satoh, M. Kim, Y. Kano, R. Goebel, & T.
Oliveira (Eds.),
COLIEE 2017. 4th Competition on Legal Information Extraction and Entailment,
held in conjunc-
tion with the 16th International Conference on Artificial Intelligence and Law
(ICAIL 2017) in
King’s College London, UK (pp. 79–89). EasyChair volume 47 of EPiC Series in
Computing.
URL: https://easychair.org/publications/paper/pqm.
Nicula, B., Ruseti, S., & Rebedea, T. (2018). Improving deep learning for
multiple choice ques-
tion answering with candidate contexts. In G. Pasi, B. Piwowarski, L. Azzopardi,
& A. Han-
bury (Eds.), Advances in Information Retrieval - 40th European Conference on IR
Research, ECIR
2018, Grenoble, France, March 26-29, 2018, Proceedings (pp. 678–683). Springer
volume 10772 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-319-76941-7_62.
doi:10.1007/978-3-319-76941-7\_62.
Pe˜nas, A., Forner, P., Rodrigo, ´
A., Sutcliffe, R. F. E., Forascu, C., & Mota, C. (2010). Overview of
respubliqa 2010: Question answering evaluation over european legislation. In M.
Braschler, D. Har-
man, & E. Pianta (Eds.), CLEF 2010 LABs and Workshops, Notebook Papers, 22-23
Septem-
ber 2010, Padua, Italy. CEUR-WS.org volume 1176 of CEUR Workshop Proceedings.
URL:
http://ceur-ws.org/Vol-1176/CLEF2010wn-MLQA10-PenasEt2010.pdf.
Pe˜nas, A., Forner, P., Sutcliffe, R. F. E., Rodrigo, ´
A., Forascu, C., Alegria, I., Giampiccolo, D.,
Moreau, N., & Osenova, P. (2009). Overview of respubliqa 2009: Question
answering evaluation
over european legislation. In C. Peters, G. M. D. Nunzio, M. Kurimo, T. Mandl,
D. Mostefa,
A. Pe˜nas, & G. Roda (Eds.), Multilingual Information Access Evaluation I. Text
Retrieval Ex-
periments, 10th Workshop of the Cross-Language Evaluation Forum, CLEF 2009,
Corfu, Greece,
September 30 - October 2, 2009, Revised Selected Papers (pp. 174–196). Springer
volume 6241 of
51








Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-642-15754-7_21.
doi:10.1007/978-3-642-15754-7\_21.
Pe˜nas, A., & Rodrigo, ´
A. (2011). A simple measure to assess non-response. In D. Lin, Y. Mat-
sumoto, & R. Mihalcea (Eds.), The 49th Annual Meeting of the Association for
Computational
Linguistics: Human Language Technologies, Proceedings of the Conference, 19-24
June, 2011,
Portland, Oregon, USA (pp. 1415–1424). The Association for Computer Linguistics.
URL:
https://www.aclweb.org/anthology/P11-1142/.
Peters, C., Clough, P. D., Gonzalo, J., Jones, G. J. F., Kluck, M., & Magnini,
B. (Eds.) (2005).
Multilingual Information Access for Text, Speech and Images, 5th Workshop of the
Cross-Language
Evaluation Forum, CLEF 2004, Bath, UK, September 15-17, 2004, Revised Selected
Papers volume
3491 of Lecture Notes in Computer Science. Springer. URL:
https://doi.org/10.1007/b138934.
doi:10.1007/b138934.
Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., &
Zettlemoyer, L. (2018).
Deep contextualized word representations. In Proceedings of NAACL-HLT (pp.
2227–2237).
Quaresma, P., & Rodrigues, I. (2005). A question-answering system for portuguese
juridical doc-
uments. In G. Sartor (Ed.), The Tenth International Conference on Artificial
Intelligence and
Law, Proceedings of the Conference, June 6-11, 2005, Bologna, Italy (pp.
256–257). ACM. URL:
https://doi.org/10.1145/1165485.1165536. doi:10.1145/1165485.1165536.
Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). Squad: 100,000+
questions for machine
comprehension of text. In Proceedings of the 2016 Conference on Empirical
Methods in Natural
Language Processing (pp. 2383–2392).
Ravichander, A., Black, A. W., Wilson, S., Norton, T. B., & Sadeh, N. M. (2019).
Question answering
for privacy policies: Combining computational and legal perspectives. In K.
Inui, J. Jiang, V. Ng,
& X. Wan (Eds.), Proceedings of the 2019 Conference on Empirical Methods in
Natural Language
52









Processing and the 9th International Joint Conference on Natural Language
Processing, EMNLP-
IJCNLP 2019, Hong Kong, China, November 3-7, 2019 (pp. 4946–4957). Association
for Computa-
tional Linguistics. URL: https://doi.org/10.18653/v1/D19-1500.
doi:10.18653/v1/D19-1500.
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). ”why should I trust you?”:
Explaining the predictions
of any classifier. In B. Krishnapuram, M. Shah, A. J. Smola, C. C. Aggarwal, D.
Shen, & R. Rastogi
(Eds.), Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge
Discovery
and Data Mining, San Francisco, CA, USA, August 13-17, 2016 (pp. 1135–1144).
ACM. URL:
https://doi.org/10.1145/2939672.2939778. doi:10.1145/2939672.2939778.
Riedel, S., Yao, L., McCallum, A., & Marlin, B. M. (2013). Relation extraction
with matrix factoriza-
tion and universal schemas. In Proceedings of the 2013 Conference of the North
American Chapter
of the Association for Computational Linguistics: Human Language Technologies
(pp. 74–84).
Rodrigo, ´
A., & Pe˜nas, A. (2017). A study about the future evaluation of
question-answering sys-
tems. Knowl. Based Syst.,137 , 83–93. URL:
https://doi.org/10.1016/j.knosys.2017.09.015.
doi:10.1016/j.knosys.2017.09.015.
Rodrigo, ´
A., P´erez-Iglesias, J., Pe˜nas, A., Garrido, G., & Araujo, L. (2013). An-
swering questions about european legislation. Expert Syst. Appl.,40 , 5811–5816.
URL:
https://doi.org/10.1016/j.eswa.2013.05.008. doi:10.1016/j.eswa.2013.05.008.
da Silva, J. W. F., Venceslau, A. D. P., Sales, J. E., Maia, J. G. R., Pinheiro,
V. C. M.,
& Vidal, V. M. P. (2020). A short survey on end-to-end simple question answering
sys-
tems. Artif. Intell. Rev.,53 , 5429–5453. URL:
https://doi.org/10.1007/s10462-020-09826-5.
doi:10.1007/s10462- 020-09826-5.
Soares, M. A. C., & Parreiras, F. S. (2020). A literature review on question
answering tech-
niques, paradigms and systems. J. King Saud Univ. Comput. Inf. Sci.,32 ,
635–646. URL:
https://doi.org/10.1016/j.jksuci.2018.08.005. doi:10.1016/j.jksuci.2018.08.005.
53














Sovrano, F., Palmirani, M., Distefano, B., Sapienza, S., & Vitali, F. (2021). A
dataset for evaluating
legal question answering on private international law. In J. Maranh˜ao, & A. Z.
Wyner (Eds.), ICAIL
’21: Eighteenth International Conference for Artificial Intelligence and Law,
S˜ao Paulo Brazil,
June 21 - 25, 2021 (pp. 230–234). ACM. URL:
https://doi.org/10.1145/3462757.3466094.
doi:10.1145/3462757.3466094.
Sovrano, F., Palmirani, M., & Vitali, F. (2020). Legal knowledge extraction for
knowledge graph
based question-answering. In V. Serena, J. Harasta, & P. Kremen (Eds.), Legal
Knowledge and
Information Systems - JURIX 2020: The Thirty-third Annual Conference, Brno,
Czech Republic,
December 9-11, 2020 (pp. 143–153). IOS Press volume 334 of Frontiers in
Artificial Intelligence
and Applications. URL: https://doi.org/10.3233/FAIA200858.
doi:10.3233/FAIA200858.
Taniguchi, R., Hoshino, R., & Kano, Y. (2018). Legal question answering system
using framenet.
In K. Kojima, M. Sakamoto, K. Mineshima, & K. Satoh (Eds.), New Frontiers in
Artificial
Intelligence - JSAI-isAI 2018 Workshops, JURISIN, AI-Biz, SKL, LENLS, IDAA,
Yokohama,
Japan, November 12-14, 2018, Revised Selected Papers (pp. 193–206). Springer
volume 11717 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-030-31605-1_15.
doi:10.1007/978-3-030-31605-1\_15.
Taniguchi, R., & Kano, Y. (2016). Legal yes/no question answering system using
case-role analysis.
In S. Kurahashi, Y. Ohta, S. Arai, K. Satoh, & D. Bekki (Eds.), New Frontiers in
Artificial Intelli-
gence - JSAI-isAI 2016 Workshops, LENLS, HAT-MASH, AI-Biz, JURISIN and SKL,
Kanagawa,
Japan, November 14-16, 2016, Revised Selected Papers (pp. 284–298). Springer
volume 10247 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-319-61572-1_19.
doi:10.1007/978-3-319-61572-1\_19.
Tesauro, G., Gondek, D., Lenchner, J., Fan, J., & Prager, J. M. (2013). Analysis
of
watson’s strategies for playing jeopardy! J. Artif. Intell. Res.,47, 205–251.
URL:
https://doi.org/10.1613/jair.3834. doi:10.1613/jair.3834.
54












Tong, P., Zhang, Q., & Yao, J. (2019). Leveraging domain context for question
answering over
knowledge graph. Data Science and Engineering,4, 323–335.
Tran, O. T., Bach, N. X., Nguyen, M. L., & Shimazu, A. (2013). Answering legal
questions by mining
reference information. In Y. I. Nakano, K. Satoh, & D. Bekki (Eds.), New
Frontiers in Artificial
Intelligence - JSAI-isAI 2013 Workshops, LENLS, JURISIN, MiMI, AAA, and DDS,
Kanagawa,
Japan, October 27-28, 2013, Revised Selected Papers (pp. 214–229). Springer
volume 8417 of
Lecture Notes in Computer Science. URL:
https://doi.org/10.1007/978-3-319-10061-6_15.
doi:10.1007/978-3-319-10061-6\_15.
Veena, G., Gupta, D., Anil, A., & Akhil, S. (2019). An ontology driven question
answering system for
legal documents. In 2019 2nd International Conference on Intelligent Computing,
Instrumentation
and Control Technologies (ICICICT) (pp. 947–951). IEEE volume 1.
Verma, A., Morato, J., Jain, A., & Arora, A. (2020). Relevant subsection
retrieval for law domain
question answer system. In Data Visualization and Knowledge Engineering (pp.
299–319). Springer.
Wang, M. (2006). A survey of answer extraction techniques in factoid question
answering. Computa-
tional Linguistics,1, 1–14.
Wehnert, S., Hoque, S. A., Fenske, W., & Saake, G. (2019). Threshold-based
retrieval and textual
entailment detection on legal bar exam questions. arXiv preprint
arXiv:1905.13350 , .
Wu, J., Liu, J., & Luo, X. (2020). Few-shot legal knowledge question answering
system for covid-19
epidemic. In 2020 3rd International Conference on Algorithms, Computing and
Artificial Intelli-
gence (pp. 1–6).
Xiao, G., Mo, J., Chow, E., Chen, H., Guo, J., & Gong, Z. (2017). Multi-task CNN
for classification
of chinese legal questions. In O. Hussain, L. Jiang, X. Fei, C. Lan, & K. Chao
(Eds.), 14th IEEE
International Conference on e-Business Engineering, ICEBE 2017, Shanghai, China,
November 4-
55




6, 2017 (pp. 84–90). IEEE Computer Society. URL:
https://doi.org/10.1109/ICEBE.2017.22.
doi:10.1109/ICEBE.2017.22.
Xu, K., Reddy, S., Feng, Y., Huang, S., & Zhao, D. (2016). Question answering on
freebase via relation
extraction and textual evidence. In Proceedings of the 54th Annual Meeting of
the Association for
Computational Linguistics, ACL 2016, August 7-12, 2016, Berlin, Germany, Volume
1: Long Pa-
pers. The Association for Computer Linguistics. URL:
https://doi.org/10.18653/v1/p16-1220.
doi:10.18653/v1/p16-1220.
Yan, M., Zhang, H., Jin, D., & Zhou, J. T. (2020). Multi-source meta transfer
for low resource
multiple-choice question answering. In D. Jurafsky, J. Chai, N. Schluter, & J.
R. Tetreault (Eds.),
Proceedings of the 58th Annual Meeting of the Association for Computational
Linguistics, ACL
2020, Online, July 5-10, 2020 (pp. 7331–7341). Association for Computational
Linguistics. URL:
https://doi.org/10.18653/v1/2020.acl-main.654.
doi:10.18653/v1/2020.acl-main.654.
Yih, W., Chang, M., Meek, C., & Pastusiak, A. (2013). Question answering using
enhanced lexical se-
mantic models. In Proceedings of the 51st Annual Meeting of the Association for
Computational Lin-
guistics, ACL 2013, 4-9 August 2013, Sofia, Bulgaria, Volume 1: Long Papers (pp.
1744–1753). The
Association for Computer Linguistics. URL:
https://www.aclweb.org/anthology/P13-1171/.
Zhang, N.-N., & Xing, Y. (2021). Questions and answers on legal texts based on
bert-bigru. In Journal
of Physics: Conference Series (p. 012035). IOP Publishing volume 1828.
Zhong, B., He, W., Huang, Z., Love, P. E. D., Tang, J., & Luo, H. (2020a). A
building regulation
question answering system: A deep learning methodology. Adv. Eng. Informatics,46
, 101195. URL:
https://doi.org/10.1016/j.aei.2020.101195. doi:10.1016/j.aei.2020.101195.
Zhong, H., Wang, Y., Tu, C., Zhang, T., Liu, Z., & Sun, M. (2020b). Iteratively
questioning and
answering for interpretable legal judgment prediction. In The Thirty-Fourth AAAI
Conference on
Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of
Artificial Intelli-
56











gence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in
Artificial
Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020 (pp. 1250–1257).
AAAI Press.
URL: https://aaai.org/ojs/index.php/AAAI/article/view/5479.
Zhong, H., Xiao, C., Tu, C., Zhang, T., Liu, Z., & Sun, M. (2020c). JEC-QA: A
legal-domain
question answering dataset. In The Thirty-Fourth AAAI Conference on Artificial
Intelligence,
AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence
Conference,
IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial
Intelligence,
EAAI 2020, New York, NY, USA, February 7-12, 2020 (pp. 9701–9708). AAAI Press.
URL:
https://aaai.org/ojs/index.php/AAAI/article/view/6519.
Zou, L., Huang, R., Wang, H., Yu, J. X., He, W., & Zhao, D. (2014). Natural
language question
answering over RDF: a graph data driven approach. In C. E. Dyreson, F. Li, & M.
T. ¨
Ozsu
(Eds.), International Conference on Management of Data, SIGMOD 2014, Snowbird,
UT, USA,
June 22-27, 2014 (pp. 313–324). ACM. URL:
https://doi.org/10.1145/2588555.2610525.
doi:10.1145/2588555.2610525.
57








CITATIONS (0)


REFERENCES (95)




ResearchGate has not been able to resolve any citations for this publication.
The linked legal data landscape: linking legal data across different countries
Article
Full-text available
 * Dec 2021
 * Artif Intell Law

 * Erwin Filtz
 * Sabrina Kirrane
 * Axel Polleres

The European Union is working towards harmonizing legislation across Europe, in
order to improve cross-border interchange of legal information. This goal is
supported for instance via standards such as the European Law Identifier (ELI)
and the European Case Law Identifier (ECLI), which provide technical
specifications for Web identifiers and suggestions for vocabularies to be used
to describe metadata pertaining to legal documents in a machine readable format.
Notably, these ECLI and ELI metadata standards adhere to the RDF data format
which forms the basis of Linked Data, and therefore have the potential to form a
basis for a pan-European legal Knowledge Graph. Unfortunately, to date said
specifications have only been partially adopted by EU member states. In this
paper we describe a methodology to transform the existing legal information
system used in Austria to such a legal knowledge graph covering different steps
from modeling national specific aspects, to population, and finally the
integration of legal data from other countries through linked data. We
demonstrate the usefulness of this approach by exemplifying practical use cases
from legal information search, which are not possible in an automated fashion so
far.
View
Show abstract
Answering Legal Questions by Learning Neural Attentive Text Representation
Conference Paper
Full-text available
 * Jan 2020

 * Phi Manh Kien
 * Ha-Thanh Nguyen
 * Ngo Xuan Bach
 * Tu Minh Phuong

View
Legal Knowledge Extraction for Knowledge Graph Based Question-Answering
Chapter
Full-text available
 * Dec 2020

 * Francesco Sovrano
 * Monica Palmirani
 * Fabio Vitali

This paper presents the Open Knowledge Extraction (OKE) tools combined with
natural language analysis of the sentence in order to enrich the semantic of the
legal knowledge extracted from legal text. In particular the use case is on
international private law with specific regard to the Rome I Regulation EC
593/2008, Rome II Regulation EC 864/2007, and Brussels I bis Regulation EU
1215/2012. A Knowledge Graph (KG) is built using OKE and Natural Language
Processing (NLP) methods jointly with the main ontology design patterns defined
for the legal domain (e.g., event, time, role, agent, right, obligations,
jurisdiction). Using critical questions, underlined by legal experts in the
domain, we have built a question answering tool capable to support the
information retrieval and to answer to these queries. The system should help the
legal expert to retrieve the relevant legal information connected with topics,
concepts, entities, normative references in order to integrate his/her searching
activities.
View
Show abstract
Multi-source Meta Transfer for Low Resource Multiple-Choice Question Answering
Conference Paper
Full-text available
 * Jan 2020

 * Ming Yan
 * Hao Zhang
 * Di Jin
 * Joey Tianyi Zhou

View
A dataset for evaluating legal question answering on private international law
Conference Paper
 * Jun 2021

 * Francesco Sovrano
 * Monica Palmirani
 * Biagio Distefano
 * Fabio Vitali

View
Few-Shot Legal Knowledge Question Answering System for COVID-19 Epidemic
Conference Paper
 * Dec 2020

 * Jiaye Wu
 * Jie Liu
 * Xudong Luo

View
A building regulation question answering system: A deep learning methodology
Article
 * Nov 2020
 * ADV ENG INFORM

 * Botao Zhong
 * Wanlei He
 * Ziwei Huang
 * Hanbin Luo

Regulations play an important role in assuring the quality of a building’s
construction and minimizing its adverse environmental impacts. Engineers and the
like need to retrieve regulatory information to ensure a building conforms to
specified standards. Despite the availability of search engines and digital
databases that can be used to store regulations, engineers, for example, are
unable to retrieve information for domain-specific needs in a timely manner. As
a consequence, users often have to deal with the burden of browsing and
filtering information, which can be a time-consuming process. This research
develops a robust end-to-end methodology to improve the efficiency and
effectiveness of retrieving queries pertaining to building regulations. The
developed methodology integrates information retrieval with a deep learning
model of Natural Language Processing (NLP) to provide precise and rapid answers
to user’s questions from a collection of building regulations. The methodology
is evaluated and a prototype system to retrieve queries is developed. The
paper’s contribution is therefore twofold as it develops a: (1) methodology that
combines NLP and deep learning to be able to address queries raised about the
building regulations; and (2) chatbot of question answering system, which we
refer to as QAS4CQAR. Our proposed methodology has powerful feature
representation and learning capability and therefore can potentially be adopted
to building regulations in other jurisdictions.
View
Show abstract
Explanation in AI and law: Past, present and future
Article
 * Dec 2020
 * ARTIF INTELL

 * Katie Atkinson
 * Trevor Bench-Capon
 * Danushka Bollegala

Explanation has been a central feature of AI systems for legal reasoning since
their inception. Recently, the topic of explanation of decisions has taken on a
new urgency, throughout AI in general, with the increasing deployment of AI
tools and the need for lay users to be able to place trust in the decisions that
the support tools are recommending. This paper provides a comprehensive review
of the variety of techniques for explanation that have been developed in AI and
Law. We summarise the early contributions and how these have since developed. We
describe a number of notable current methods for automated explanation of legal
reasoning and we also highlight gaps that must be addressed by future systems to
ensure that accurate, trustworthy, unbiased decision support can be provided to
legal professionals. We believe that insights from AI and Law, where explanation
has long been a concern, may provide useful pointers for future development of
explainable AI.
View
Show abstract
AILA: A Question Answering System in the Legal Domain
Conference Paper
 * Jul 2020

 * Weiyi Huang
 * Jiahao Jiang
 * Qiang Qu
 * Min Yang

Question answering (QA) in the legal domain has gained increasing popularity for
people to seek legal advice. However, existing QA systems struggle to comprehend
the legal context and provide jurisdictionally relevant answers due to the lack
of domain expertise. In this paper, we develop an Artificial Intelligence Law
Assistant (AILA) for question answering in the domain of Chinese laws. AILA
system automatically comprehends users' natural language queries with the help
of the legal knowledge graph (KG) and provides the best matching answers for
given queries. In addition, AILA provides visual cues to interpret the input
queries and candidate answers based on the legal KG. Experimental results on a
large-scale legal QA corpus show the effectiveness of AILA. To the best of our
knowledge, AILA is the first Chinese legal QA system which integrates the domain
knowledge from legal KG to comprehend the questions and answers for ranking QA
pairs. AILA is available at http://bmilab.ticp.io:48478/.
View
Show abstract
A novel method based on symbolic regression for interpretable semantic
similarity measurement
Article
 * Jun 2020
 * EXPERT SYST APPL

 * Jorge Martinez-Gil
 * Jose M. Chaves-González

The problem of automatically measuring the degree of semantic similarity between
textual expressions is a challenge that consists of calculating the degree of
likeness between two text fragments that have none or few features in common
according to human judgment. In recent times, several machine learning methods
have been able to establish a new state-of-the-art regarding the accuracy, but
none or little attention has been paid to their interpretability, i.e. the
extent to which an end-user could be able to understand the cause of the output
from these approaches. Although such solutions based on symbolic regression
already exist in the field of clustering (Lensen et al., 2019), we propose here
a new approach which is being able to reach high levels of interpretability
without sacrificing accuracy in the context of semantic textual similarity.
After a complete empirical evaluation using several benchmark datasets, it is
shown that our approach yields promising results in a wide range of scenarios.
View
Show abstract
Show more




RECOMMENDATIONS

Discover more
Project


MPCENERGY

 * Georgios C. Chasparis
 * Thomas Natschläger
 * Thomas Grubinger
 * [...]
 * Christa Illibauer

View project
Project


ACCURATE AND EFFICIENT PROFILE MATCHING IN KNOWLEDGE BASES

 * Jorge Martinez-Gil

Accurate matching of job offers and applicant profiles
View project
Project


SEMANTIC SIMILARITY MEASUREMENT

 * Jorge Martinez-Gil

View project
Project


ONTOLOGY MATCHING AND META-MATCHING

 * Jorge Martinez-Gil
 * Jose F Aldana Montes
 * Ismael Navas Delgado

View project
Article


AUGMENTED AND CHALLENGING DATASETS WITH MULTI-STEP REASONING AND MULTI-SPAN
QUESTIONS FOR CHINESE JU...

December 2022
 * Qingye Meng
 * Ziyue Wang
 * Hang Chen
 * [...]
 * Shijin Wang

The existing judicial reading comprehension datasets are relatively simple, and
the answers to the questions can be obtained through single-step reasoning.
However, the content of legal documents in actual scenarios is complex, making
it problematic to infer correct results merely by single-step reasoning. To
solve this type of issue, we promote the difficulties of questions included in
Chinese ... [Show full abstract] Judicial Reading Comprehension (CJRC) dataset
and propose two augmented versions, CJRC2.0 and CJRC3.0. These datasets are
derived from Chinese judicial judgment documents in different fields and
annotated by judicial professionals. Compared to CJRC, there are more types of
judgment documents in the two datasets, and the questions become are more
challenging to answer. For CJRC2.0, we only preserve complex questions that
require to be solved by multi-step reasoning. Besides, we provide additional
supporting facts to the answers. For CJRC3.0, we introduce a new question type,
the multi-span question, which should be answered by extracting and combining
multiple spans in the documents. We implement two powerful baselines to evaluate
the difficulty of our proposed datasets. Our proposed datasets fill gaps in the
field of explainable legal machine reading comprehension.
Read more
Conference Paper
Full-text available


QUESTION ANSWERING OF BAR EXAMS BY PARAPHRASING AND LEGAL TEXT ANALYSIS

July 2017 · Lecture Notes in Computer Science
 * mi-young Kim
 * Ying Xu
 * Yao Lu
 * Randy Goebel

Our legal question answering system combines legal information retrieval and
textual entailment, and exploits paraphrasing and sentence-level analysis of
queries and legal statutes. We have evaluated our system using the training data
from the competition on legal information extraction/entailment (COLIEE)-2016.
The competition focuses on the legal information processing required to answer
yes/no ... [Show full abstract] questions from Japanese legal bar exams, and it
consists of three phases: legal ad-hoc information retrieval (Phase 1), textual
entailment (Phase 2), and a combination of information retrieval and textual
entailment (Phase 3). Phase 1 requires the identification of Japan civil law
articles relevant to a legal bar exam query. For this phase, we have used an
information retrieval approach using TF-IDF and a Ranking SVM. Phase 2 requires
decision on yes/no answer for previously unseen queries, which we approach by
comparing the approximate meanings of queries with relevant articles. Our
meaning extraction process uses a selection of features based on a kind of
paraphrase, coupled with a condition/conclusion/exception analysis of articles
and queries. We also identify synonym relations using word embedding, and detect
negation patterns from the articles. Our heuristic selection of attributes is
used to build an SVM model, which provides the basis for ranking a decision on
the yes/no questions. Experimental evaluation show that our method outperforms
previous methods. Our result ranked highest in the Phase 3 in the COLIEE-2016
competition.
View full-text
Chapter


COLIEE-2018: EVALUATION OF THE COMPETITION ON LEGAL INFORMATION EXTRACTION AND
ENTAILMENT

October 2019
 * Yoshinobu Kano
 * mi-young Kim
 * Masaharu Yoshioka
 * [...]
 * Ken Satoh

We summarize the evaluation of the 5th Competition on Legal Information
Extraction/Entailment 2018 (COLIEE-2018). The COLIEE-2018 tasks include two
tasks in each of statute law and case law. The case law component includes an
information retrieval (Task 1), and the confirmation of an entailment relation
between an existing case and an unseen case (Task 2). The statute law component
includes ... [Show full abstract] information retrieval (Task 3) and
entailment/question answering (Task 4). Participation was open to any group
based on any approach. 13 teams participated in the case law competition, and we
received results from 7 teams where 6 submissions to Task 1 (12 runs), and 4
submissions to Task 2 (8 runs). Regarding the statute law, there were
submissions of 17 runs from 8 teams (including 2 organizers’ runs) for Task 3
and 7 runs from 3 teams for Task 4. We describe each team’s approaches, our
official evaluation, and analysis on our data and submission results. We also
discuss possibilities for future competition tasks.
Read more
Chapter


MULTIPLE CHOICE QUESTION ANSWERING IN THE LEGAL DOMAIN USING REINFORCED
CO-OCCURRENCE

August 2019
 * Jorge Martinez-Gil
 * Bernhard Freudenthaler
 * A Min Tjoa

Nowadays, the volume of legal information available is continuously growing. As
a result, browsing and querying this huge legal corpus in search of specific
information is currently a tedious task exacerbated by the fact that data
presentation does not usually meet the needs of professionals in the sector. To
satisfy these ever-increasing needs, we have designed an appropriate solution to
provide ... [Show full abstract] an adaptive and intelligent solution for the
automatic answer of questions of legal content based on the computation of
reinforced co-occurrence, i.e. a very demanding type of co-occurrence that
requires large volumes of information but guarantees good results. This solution
is based on the pattern-based methods that have been already successfully
applied in information extraction research. An empirical evaluation over a
dataset of legal questions seems to indicate that this solution is promising.
Read more


Discover the world's research
Join ResearchGate to find the people and research you need to help your work.
Join for free

ResearchGate iOS App
Get it from the App Store now.
Install
Keep up with your stats and more
Access scientific knowledge from anywhere

or
Discover by subject area
 * Recruit researchers
 * Join for free
 * Login
   Email
   Tip: Most researchers use their institutional email address as their
   ResearchGate login
   
   PasswordForgot password?
   Keep me logged in
   Log in
   or
   Continue with Google
   
   Welcome back! Please log in.
   Email
   · Hint
   Tip: Most researchers use their institutional email address as their
   ResearchGate login
   
   PasswordForgot password?
   Keep me logged in
   Log in
   or
   Continue with Google
   No account? Sign up
   

Company
About us
News
Careers
Support
Help Center
Business solutions
Advertising
Recruiting

© 2008-2023 ResearchGate GmbH. All rights reserved.
 * Terms
 * Privacy
 * Copyright
 * Imprint