Papers Accepted to the JURIX 2011 Conference

My colleagues and I have had two papers (one long and one short) accepted for presentation at The 24th International Conference on Legal Knowledge and Information Systems (JURIX 2011). The papers are available on the links.
On Rule Extraction from Regulations
Adam Wyner and Wim Peters
Abstract
Rules in regulations such as found in the US Federal Code of Regulations can be expressed using conditional and deontic rules. Identifying and extracting such rules from the language of the source material would be useful for automating rulebook management and translating into an executable logic. The paper presents a linguistically-oriented, rule-based approach, which is in contrast to a machine learning approach. It outlines use cases, discusses the source materials, reviews the methodology, then provides initial results and future steps.
Populating an Online Consultation Tool
Sarah Pulfrey-Taylor, Emily Henthorn, Katie Atkinson, Adam Wyner, and Trevor Bench-Capon
Abstract
The paper addresses the extraction, formalisation, and presentation of public policy arguments. Arguments are extracted from documents that comment on public policy proposals. Formalising the information from the arguments enables the construction of models and systematic analysis of the arguments. In addition, the arguments are represented in a form suitable for presentation in an online consultation tool. Thus, the forms in the consultation correlate with the formalisation and can be evaluated accordingly. The stages of the process are outlined with reference to a working example.
Shortlink to this page.
By Adam Wyner

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License.

Workshop on Modelling Policy-making (MPM 2011)

In conjunction with
The 24th International Conference on Legal Knowledge and Information Systems (JURIX 2011)
Wednesday December 14, 2011
University of Vienna
Vienna, Austria
Context:
As the European Union develops, issues about governance, legitimacy, and transparency become more pressing. National governments and the EU Commission realise the need to promote widespread, deliberative democracy in the policy-making cycle, which has several phases: 1) agenda setting, 2) policy analysis, 3) lawmaking, 4) administration and implementation, and 5) monitoring. As governments must become more efficient and effective with the resources available, modern information and communications technology (ICT) are being drawn on to address problems of information processing in the phases. One of the key problems is policy content analysis and modelling, particularly the gap between on the one hand policy proposals and formulations that are expressed in quantitative and narrative forms and on the other hand formal models that can be used to systematically represent and reason with the information contained in the proposals and formulations.
Submission Focus:
The workshop invites submissions of original research about the application of ICT to the early phases of the policy cycle, namely those before the legislators fix the legislation: agenda setting, policy analysis, and lawmaking. The research should seek to address the gap noted above. The workshop focuses particularly on using and integrating a range of subcomponents – information extraction, text processing, representation, modelling, simulation, reasoning, and argument – to provide policy making tools to the public and public administrators.
Intended Audience:
Legal professionals, government administrators, political scientists, and computer scientists.
Areas of Interest:

  • information extraction from natural language text
  • policy ontologies
  • formal logical representations of policies
  • transformations from policy language to executable policy rules
  • argumentation about policy proposals
  • web-based tools that support participatory policy-making
  • tools for increasing public understanding of arguments behind policy decisions
  • visualising policies and arguments about policies
  • computational models of policies and arguments about policies
  • integration tools
  • multi-agent policy simulations

Preliminary Workshop Schedule:
09:45-10:00 Workshop Opening comments
10:00-11:00 Paper Session 1

  • Using PolicyCommons to support the policy-consultation process: investigating a new workflow and policy-deliberation data model
    Neil Benn and Ann Macintosh
  • A Problem Solving Model for Regulatory Policy Making
    Alexander Boer, Tom Van Engers and Giovanni Sileno

11:00-11:15 Break (coffee, tea, air etc.)
11:15-12:15 Paper Session 2

  • Linking Semantic Enrichment to Legal Documents
    Akos Szoke, Andras Forhecz, Krisztian Macsar and Gyorgy Strausz
  • Semantic Models and Ontologies in Modelling Policy-making
    Adam Wyner, Katie Atkinson and Trevor Bench-Capon

12:15-13:15 Lunch break
13:15-14:45 Paper Session 3

  • Consistent Conceptual Descriptions to Support Formal Policy Model Development: Metamodel and Approach
    Sabrina Scherer and Maria Wimmer
  • The Policy Modeling Tool of the IMPACT Argumentation Toolbox
    Thomas Gordon
  • Ontologies for Governance, Risk Management and Policy Compliance
    Jorge Gonzalez-Conejero, Albert Merono-Penuela and David Fernandez Gamez

14:45-15:00 Break (coffee, tea, air etc.)
15:00-16:00 Paper Session 4 and Closing discussion

  • Policy making: How rational is it?
    Tom Van Engers, Ignace Snellen and Wouter Van Haaften
  • Closing discussion

Workshop Registration and Location:
Please see the JURIX 2011 website for all information about registration and location.
Webpage URL:
http://wyner.info/LanguageLogicLawSoftware/?p=1157
Important Dates:

  • Submission: Monday, October 24
  • Review Notification: Monday, November 7
  • Final Version: Thursday, December 1
  • Workshop date: Wednesday, December 14

Author Guidelines:
Submit position papers of between 2-5 pages in length in PDF format and using the IOS Press style files and authors’ guidelines at:
IOS Press Author Instructions
Submit papers to:
MPM 2011 on EasyChair
Publication:
The position papers are available only in an electronic version from the following link:
Proceedings of the Workshop on Modelling Policy-making
A call for selected extended versions of the papers will be issued for a special issue of AI and Law on Modelling Policy-making.
Contact Information:
Adam Wyner, adam@wyner.info
Neil Benn, n.j.l.benn@leeds.ac.uk
Program Committee Co-Chairs:
Adam Wyner (University of Liverpool, UK)
Neil Benn (University of Leeds, UK)
Program Committee (Preliminary):
Katie Atkinson
Trevor Bench-Capon
Bruce Edmonds
Tom van Engers
Euripidis Loukis
Tom Gordon
Ann Macintosh
Gunther Schefbeck
Maria Wimmer
Radboud Winkels
By Adam Wyner

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License.

Draft Materials for LEX 2011

Draft post
At the links below, you can find the slides and hands on materials on GATE for the LEX summer school on Managing Legal Resources in the Semantic Web.
GATE Legislative Rulebook
By Adam Wyner

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License.

On ICAIL 2011 Discussion on Legal Corpus Development and Text Analytics

In this note, I point to various parts of a discussion on developing and analysing legal textual data raised at ICAIL 2011. Please feel free to add comments to this document (or to me in person, by email, on your blog and linked to this, etc), which I can then add to the post (I’m very happy to attribute contributions). The intention is to stimulate discussion on these matters to help the community of researchers move ahead on common interests.
Corpus Development
Unlike the situation from several years ago, we have accessible sources of large corpora of legal textual information. The World Legal Information Institutes provide free, independent and non-profit access to worldwide law. For example, one can go to the US site and download cases: United States v Grant [1961] USCA9 19; 286 F.2d 157 (19 January 1961); one can request zipped files or screen scrap cases. The LIIs have introduced standardised references and formats for cases. There are boolean and regex searches.
From the contacts that I have had (e.g. in the US and UK), the LIIs would be very happy to collaborate with academic researchers in the analysis of their data and in keeping with their primary mission. In particular, developing tools that can be integrated and deployed with their platforms might be a way to go, thereby addressing significant platform and dissemination issues.
Another source of corpora is public.resource.org, which distributes a range of corpora covering legislation, codes, and cases.
Analysis and Annotation
There are a range of issues about information retrieval and extraction. Others can speak about IR, statistical, machine learning approaches. What I know better is annotation, whether fully or semi automatic and manual. Here we have issues about what to annotate and how. Some low level information is unproblematic (e.g. entities of a range of sorts, sections, and sentiment); higher level information (e.g. factors) might be more complex. I have some suggestions for annotations for low level information; a good starting point for factors are the CATO factors, though there is a general issue about how to extend factor identification to other domains (CATO factors are specific for intellectual property).
One general problem with analysis is that different researchers might use different tools in their work and just report the results. This means results are not interchangeable, which is particularly problematic with annotation work. If a common ‘framework’ tool is used and some consensus is developed about (at least) low level annotation types, then work can proceed more collaboratively, transparently, and reproducibly. One can develop a more forceful argument for researchers (public service bodies and information providers) to promote such an open development methodology (among them are justification and traceability, see Wyner and Peters 2010 and David Lewis’s ICAIL 2011 keynote address on related points). General Architecture for Text Engineering is an open framework for text processing modules.
There are ‘open’ systems for text annotation — Open Calais and Open Up platform’s data enrichment service from The Stationery Office. However, there are intellectual property issues that need to be considered.
Another general issue is how to carry out manual annotation, for example to build gold standards, which are required for machine learning systems. There has been significant progress, for example, with TeamWare, which provides for curated, web-based annotation tools along with annotation analysis (e.g. inter-annotator agreement). For a short tutorial (for an experiment) on using TeamWare for annotation of some legal case factors, see Web-based Annotation Support for the Law. Wim Peters and I proposed to law school faculty to use this tool to support their student exercises for first and second year students since these exercises often require identifying and extracting information from cases. Wim and I think integrating annotation exercises into legal e-learning could both help to develop large annotated sets of data and to serve an important educational purpose. See our paper about some of these points and proposals.
Research Questions
Large corpora can be formed, tools can be applied to them, but for fund raising, the community needs to develop a range of motivating research questions and use cases. Asides from questions pursued in the AI and Law community, we might consult further with public bodies (National Center for State Courts and similar), legal information service providers (Lexis-Nexis, ThomsonReuters, Practical Law Company, law societies, political scientists, etc. The kinds of answers we look for partially guide how we structure not only the corpora, but moreso the annotations.
Funding Opportunities
Digging into Data and the Request for Proposals, but the due date is June 16 (I had been working on a proposal, but needed better research questions to hold local interest). Though the deadline is too soon to submit a proposal, it does demonstrate a widespread interest in funding bodies in the development and analysis of large corpora in the humanities and social sciences. The other obvious funding sources are national (US, UK, French, etc) and international (EU and Digging into Data).
By Adam Wyner

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License.

Talk at BILETA 2011

I’m giving a talk tomorrow, April 11 2011, at BILETA, the annual conference of the British & Irish Law, Education and Technology Association at Manchester Metropolitan University School of Law. My collaborators are Wim Peters (University of Sheffield) and Fiona Beveridge (University of Liverpool).
The abstract and slides are below:
Web-based Software Tools to Support Students’ Empirical Study of the Law
Adam Wyner (University of Liverpool, Computer Science), Wim Peters (University of Sheffield, Computer Science), and Fiona Beveridge (University of Liverpool, Law School)
The paper investigates and proposes tools to support students in empirically investigating legal cases using text analytic software. Web-based tools can be used to engage and leverage the collective skills and ambitions of law students to crowd-source the development of legal resource materials. Law school students must develop skills in close textual analysis of legal source material such as legal cases. To use source material such as case decisions to reason about how precedents apply in case-based reasoning, law students must learn to identify a range of elements in legal cases, for example, parties, jurisdiction, material facts, legislative and case citations, cause of action, ratio decideni, and others. Moreover, students should be able to address complex queries to a case or a case base (a corpus of cases) in order to answer questions of particular legal interest; for example, about relationships between a judge, parties, cause of action, and ratio. Currently students either simply rely on their own analytic abilities to read a case or find answers to questions; legal search tools (e.g. Lexis-Nexis) provide search support, but are restricted to a limited number of coarse-grained parameters and cannot search for deep, particular semantic relationships in the text. To enable automated support of queries of the corpus, and so enable deep empirical research on cases, it is essential to have a corpus of legal cases which are annotated with machine readable (XML) tags that signal the semantic properties of passages of text. To create such a corpus requires a tool to annotate the text. Such a tool would reinforce students’ examination of the source document. The paper describes recent developments of tools using Semantic Web technologies, text analysis, and web-based annotation support. With the text analysis software, General Architecture for Text Engineering (GATE), which is customised for legal applications, law students can annotate legal cases for a fine-grained range of legally relevant concepts and linguistic relations; they can also use GATE to write grammars and automatically annotate the text. Using GATE TeamWare, an online text annotation tool that automatically evaluates interannotator agreement, students can collaboratively analyse and agree on a gold standard corpus of legal cases. The corpus can be automatically indexed using Lucene, thereby allowing fast results to complex queries over any string or annotation used.
The slides of the talk are here

Proceedings and Program for Workshop on Modelling Legal Cases and Legal Rules

in conjunction with JURIX 2010
December 15, 2010
Department of Computer Science, Ashton Building, Room 310
University of Liverpool, Liverpool, United Kingdom
Workshop Proceedings
Workshop Program
Session I

    14:30-14:35
    Welcome and Introductory remarks
    14:35-15:00
    Steven van Driel (Utrecht University) and Henry Prakken (Utrecht University and University of Groningen)
    Visualising the argumentation structure of an expert witness report with Rationale (extended abstract)
    15:00-15:25
    Thomas F. Gordon (Fraunhofer FOKUS)
    Analyzing open source license compatibility issues with Carneades
    15:25-15:40
    Martyn Lloyd-Kelly, Adam Wyner, and Katie Atkinson (University of Liverpool)
    Emotional argumentation schemes in legal cases (short position paper)
    15:40-16:00
    Short informal remarks

16:00-16:30 Tea
Session II

    16:30-16:55
    Anna Ronkainen (University of Helsinki)
    MOSONG, a fuzzy logic model of trade mark similarity
    16:55-17:20
    Adam Wyner and Trevor Bench-Capon (University of Liverpool)
    Visualising legal case-based reasoning argumentation schemes
    17:20-17:45
    Burkhard Schafer (University of Edinburgh)
    Say “cheese”: natural kinds, deontic logic and European Court of Justice decision C-210\/89
    17:45-18:00
    Short informal remarks

For general information, see JURIX 2010
By Adam Wyner
Distributed under the Creative Commons
Attribution-Non-Commercial-Share Alike 2.0

Legal Know-How Workshop Presentations

December 10, 2010, I gave a presentation at the International Society for Knowledge Organisation’s meeting on Legal Know-How. It was an interesting meeting, where I got the opportunity to present my work to members of the legal profession, hear what law firms are doing about knowledge management, and make some good new contacts.
The slides of all the talks, including mine, are available:
ISKO-UK Legal Know-How meeting
In a couple of weeks, ISKO will also add mp3s of the talks, so one can see the slides and hear the talks. Nice way to do things, as remarks and narration are almost more crucial than the slides themselves.
By Adam Wyner
Distributed under the Creative Commons
Attribution-Non-Commercial-Share Alike 2.0

Call for Papers: JURIX 2010 Workshop on Modelling Legal Cases and Legal Rules

I am organising a workshop at JURIX 2010
Modelling Legal Cases and Legal Rules
As part of the Jurix 2010 conference in Liverpool UK, we will hold a Workshop on Modelling Legal Cases and Legal Rules. This workshop is a follow on from successful workshops at Jurix 2007 and ICAIL 2009.
Legal cases and legal rules in common law contexts have been modelled in a variety of ways over the course of research in AI and Law to support different styles of reasoning for a variety of problem-solving contexts, such as decision-making, information retrieval, teaching, etc. Particular legal topic areas and cases have received wide coverage in the AI and Law literature including wild animals (e.g. Pierson v. Post, Young v. Hitchens, and Keeble v. Hickeringill), intellectual property (e.g. Mason v. Jack Daniel Distillery), and evidence (e.g. the Rijkbloem case). As well, some legal rules have been widely discussed, such as legal argument schemes (e.g. Expert Testimony) or rules of evidence (see Walton 2002). However, other areas have been less well covered. For example, there appears to be less research on modelling legal cases in civil law contexts; investigation of taxonomies and ontologies of legal rules would support abstraction and formalisation (see Sherwin 2009); additional legal rules could be brought under the scope of investigation, such as those bearing on criminal assault or causes of action.
The aim of this workshop is to provide a forum in which researchers can present their research on modelling legal cases and legal rules.
Papers are solicited that model a particular legal case or a small set of legal rules. Authors are free to choose the case or set of legal rules and analyse them according to the authors’ preferred model of representation; any theoretical discussion should be grounded in or exemplified by the case or rules at hand. Papers should make clear what are the particular distinctive features of their approach and why these features are useful in modelling the chosen case or rules. The workshop is an opportunity for authors to demonstrate the benefits of their approach and for group discussions to identify useful overlapping features as well as aspects to be further explored and developed.
Format of papers and submission guidelines
Full papers should not be more than 10 pages long and should be submitted in PDF format. It is suggested that the conference style files are used for formatting (see IOS Press site). All papers should provide:

  • A summary of the case or legal rules.
  • An overview of the representation technique, or reference to a full description of it.
  • The representation itself.
  • Discussion of any significant features.

Short position papers are also welcome from those interested in the topic but who do not wish to present a fully represented case or elaborate discussion of a set of legal rules; the short position papers can outline ideas, sketch directions of research, summarise or reflect on previously published work that has addressed the topic. A short position paper should be not more than five pages, giving a clear impression of what would be presented.
All submissions should be emailed as a PDF attachment to the workshop organiser, Adam Wyner, at: adam@wyner.info.
Programme Committee (Preliminary)

  • Kevin Ashley, University of Pittsburgh, USA
  • Katie Atkinson, University of Liverpool, UK
  • Floris Bex, University of Dundee, UK
  • Trevor Bench-Capon, University of Liverpool, UK
  • Tom Gordon, Fraunhofer, FOKUS, Germany
  • Robert Richards, Seattle, Washington, USA
  • Giovanni Sartor, European University Institute, Italy
  • Burkhard Schafer, Edinburgh Law School, Scotland
  • Douglas Walton, University of Windsor, Canada

Organisation
Organiser of this workshop is Adam Wyner, University of Liverpool, UK. You can contact the workshop organiser by sending an email to adam@wyner.info
Dates
Paper submission: Friday, November 5, 2010
Accepted Notification: Friday, November 12, 2010
Workshop Registration: Friday, November 19, 2010
December 15th, 2010 Jurix Workshops/Tutorials
December 16th-17th, 2010 Jurix 2010 Main Conference
By Adam Wyner
Distributed under the Creative Commons
Attribution-Non-Commercial-Share Alike 2.0

Presentation at Legal Know-how Workshop, Nov. 10, 2010

I have been invited to make a presentation on Textual information extraction and ontologies for legal case-based reasoning at a Legal Know-how Workshop, which is an industry oriented event organised by the International Society for Knowledge Management UK.
Date: 10 November 2010
Time: 13:30-19:00
Venue: University College London
Medical Sciences Building
A. V. Hill Lecture Theatre
Gower Street
London, WC1E 6BT
See the workshop website for registration fee (either free or under £25) and booking.
This will be a very interesting opportunity to hear from and talk with industry consultants and experts about the latest developments in legal knowledge management. My thanks to Stella Dextre Clarke of ISKO-UK for organising the event and inviting me to take part.

Programme

13:30 Registration
14:00 Welcome from ISKO-UK by Stella Dextre Clarke
14:05 Legal knowledge – the practitioner’s viewpoint
Melanie Farquharson, 3Kites Consulting

This session will focus on the practical situations in which lawyers look for knowledge in order to deliver legal services to their clients. It will identify some typical ‘use cases’ and consider ways in which knowledge can be delivered to the practitioner – even without them having to look for it.

14:35 Why lawyers need taxonomies – adventures in organising legal knowledge
Kathy Jacob & Lynley Barker, Pinsent Masons LLP;
Graham Barbour & Mark Fea, LexisNexis

This presentation will cover the practical issues encountered by a law firm in its quest to improve findability of one of its key resources – knowledge and information. We will discuss our approach to building taxonomies, the tools and processes deployed and how we anticipate our taxonomy will be applied and consumed by lawyers and publishers.
The LexisNexis part of the presentation will focus on the challenges of building and applying legal taxonomies to suit the breadth and depth of content they provide online. It will also examine ways in which taxonomies can be surfaced in the user interface and help to drive compelling functionality that improves the user’s search experience.

15:20 Taxonomy management at Clifford Chance
Mats Bergman, Clifford Chance

This talk will describe how taxonomy management works in practice at Clifford Chance. As an increasing number of core knowledge resources are making use of the same set of firm-wide taxonomies, the increased interdependencies necessitate the implementation of a controlled process for updating the taxonomies. A simple governance model will be presented. Some thoughts will follow on the evolution of taxonomy development within a larger organisation and the current challenge of using social tagging in conjunction with controlled vocabularies.

15:50 Refreshments (Lower Refectory)
16:20 Textual information extraction and ontologies for legal case-based reasoning
Adam Wyner, University of Liverpool

This talk gives a brief overview of current developments and prospects in two related areas of the legal semantic web for legal cases – textual information extraction and ontologies. Textual information extraction is a process of automatically annotating and extracting textual information from the legal case base (precedents), thereby identifying elements such as participants, the roles the participants play, the factors which were considered in arriving at a decision, and so on. The information is valuable not only for search (to find applicable precedents), but also to populate an ontology for legal case-based reasoning. An ontology is a formal representation of key aspects of the knowledge of legal professionals with which we can reason (e.g. given an assertion that something is a legal case, we can infer other properties) and with respect to which we can write rules (e.g. reasoning using case factors to arrive at a legal decision). Since it is expensive to manually populate an ontology (meaning to read cases and input the data into the ontology), we use textual information extraction to automatically populate the ontology. We conclude with an appeal for open source, collaborative development of legal knowledge systems among partners in academia, industry, and government.

17:00 Collaboration across boundaries
Gwenda Sippings & Gerard Bredenoord, Linklaters LLP

In this presentation, we will look at approaches to managing legal know-how in a major global law firm. We will describe several boundaries that we have to consider when organising our know-how, including boundaries between professionals, countries, internal and external resources and the well debated boundary between information and knowledge. We will also share some of the ways in which we are making our know-how available to the fee earners and other professionals in the firm, using social and technological solutions.

17:35 Reconciling the taxonomy needs of different users
Derek Sturdy, Tikit Knowledge Services

The last decade has seen the development of a substantial number of legal know-how and knowledge databases. It has also shown up a serious question on whether the metadata, and especially the taxonomies, that are applied to the various knowledge items, should be tailored to the particular needs of end-users, or whether, so to speak, "one size can fit all". In particular, this talk will discuss the overlapping, but discrete, needs of those using knowledge resources primarily for legal drafting and document production, and of those conducting legal research, and will address the relative value today, (as opposed to in 2000), of the effort put into internal metadata creation for those two sorts of end-users.

By Adam Wyner
Distributed under the Creative Commons
Attribution-Non-Commercial-Share Alike 2.0

Semantic Processing of Legal Texts Workshop

In this post you will find information on the Semantic Processing of Legal Texts workshop, held in conjunction with the Language Resources and Evaluation Conference. Below please find a link to the conference, information on the workshop, and a program for the conference.
LREC
Language Resources and Evaluation Conference, May 17-23, Malta.
LREC 2010 Workshop on
SEMANTIC PROCESSING OF LEGAL TEXTS (SPLeT-2010)

23 May 2010, Malta
Workshop Description
The legal domain represents a primary candidate for web-based information distribution, exchange and management, as testified by the numerous e-government, e-justice and e-democracy initiatives worldwide. The last few years have seen a growing body of research and practice in the field of Artificial Intelligence and Law which addresses a range of topics: automated legal reasoning and argumentation, semantic and cross-language legal information retrieval, document classification, legal drafting, legal knowledge discovery and extraction, as well as the construction of legal ontologies and their application to the law domain. In this context, it is of paramount importance to use Natural Language Processing techniques and tools that automate and facilitate the process of knowledge extraction from legal texts.
Over the last two years, a number of dedicated workshops and tutorials specifically focusing on different aspects of semantic processing of legal texts has demonstrated the current interest in research on Artificial Intelligence and Law in combination with Language Resources (LR) and Human Language Technologies (HLT). The LREC 2008 Workshop on “Semantic processing of legal texts” was held in Marrakech, Morocco, on the 27th of May 2008. The JURIX 2008 Workshop on “the Natural Language Engineering of Legal Argumentation: Language, Logic, and Computation (NaLEA)”, which focused on recent advances in natural language engineering and legal argumentation. The ICAIL 2009 Workshops “LOAIT ’09 – the 3rd Workshop on Legal Ontologies and Artificial Intelligence Techniques joint with the 2nd Workshop on Semantic Processing of Legal Texts” and “NALEA’09 – Workshop on the Natural Language Engineering of Legal Argumentation: Language, Logic, and Computation”, the former focusing on Legal Knowledge Representation with particular emphasis on the issue of ontology acquisition from legal texts, the latter tackling issues related to legal argumentation and linguistic technologies.
To continue this momentum, a 3rd Workshop on “Semantic Processing of Legal Texts” is being organised at the LREC conference to bring to the attention of the broader LR/HLT community the specific technical challenges posed by the semantic processing of legal texts and also share with the community the motivations and objectives which make it of interest to researchers in legal informatics. The outcome of these interactions are expected to advance research and applications and foster interdisciplinary collaboration within the legal domain.
The main goals of the workshop are to provide an overview of the state-of-the-art in legal knowledge extraction and management, to explore new research and development directions and emerging trends, and to exchange information regarding legal LRs and HLTs and their applications.
Areas of Interest
The workshop will focus on the topics of the automatic extraction of information from legal texts and the structural organisation of the extracted knowledge. Particular emphasis will be given to the crucial role of language resources and human language technologies. Papers are on, but not limited to, the following topics:

  • Building legal resources: terminologies, ontologies, corpora
  • Ontologies of legal texts, including subareas such as ontology acquisition, ontology customisation, ontology merging, ontology extension, ontology evolution, lexical information, etc.
  • Information retrieval and extraction from legal texts
  • Semantic annotation of legal texts
  • Legal text processing
  • Multilingual aspects of legal text semantic processing
  • Legal thesauri mapping
  • Automatic Classification of legal documents
  • Logical analysis of legal language
  • Automated parsing and translation of natural language arguments into a logical formalism
  • Linguistically-oriented XML mark up of legal arguments
  • Dialogue protocols for argumentation
  • Legal argument ontology
  • Computational theories of argumentation that are suitable to natural language
  • Controlled language systems for law
  • Workshop Chairs

  • Enrico Francesconi (Istituto di Teoria e Tecniche dell’Informazione Giuridica of CNR, Florence, Italy)
  • Simonetta Montemagni (Istituto di Linguistica Computazionale of CNR, Pisa, Italy)
  • Wim Peters (Natural Language Processing Research Group, University of Sheffield, UK)
  • Adam Wyner (Department of Computer Science, University College London, UK)
  • Program Committee

  • Johan Bos (University of Rome, Italy)
  • Danièle Bourcier (Humboldt Universität, Berlin, Germany)
  • Thomas R. Bruce (Cornell Law School, Ithaca, NY, USA)
  • Pompeu Casanovas (Institut de Dret i Tecnologia, UAB, Barcelona, Spain)
  • Alessandro Lenci (Dipartimento di Linguistica, Università di Pisa, Pisa, Italy)
  • Leonardo Lesmo (Dipartimento di Informatica, Università di Torino, Torino, Italy)
  • Raquel Mochales Palau (Catholic University of Leuven, Belgium)
  • Paulo Quaresma (Universidade de Évora, Portugal)
  • Erich Schweighofer (Universität Wien, Rechtswissenschaftliche Fakultät, Wien, Austria)
  • Manfred Stede (University of Potsdam, Germany)
  • Daniela Tiscornia (Istituto di Teoria e Tecniche dell’Informazione Giuridica of CNR, Florence, Italy)
  • Tom van Engers (Leibniz Center for Law, University of Amsterdam, Netherlands)
  • Stephan Walter (Euroscript, Luxembourg S.a.r.l.)
  • Radboud Winkels (Leibniz Center for Law, University of Amsterdam, Netherlands)
  • Program

  • 14:30-14:45 Welcome and introduction
  • 14:45-15:10
    A Description Language for Content Zones of German Court Decisions
    Florian Kuhn
  • 15:10-15:35
    Controlling the language of statutes and regulations for semantic processing
    Stefan Hoefler and Alexandra Bünzli
  • 15:35-16:00
    Named entity recognition in the legal domain for ontology population
    Mírian Bruckschen, Caio Northfleet, Douglas da Silva, Paulo Bridi, Roger Granada, Renata Vieira, Prasad Rao and Tomas Sander
  • 16:00-16:30
    Coffee break

  • 16:30-16:55
    Legal Claim Identification: Information Extraction with Hierarchically Labeled Data
    Mihai Surdeanu, Ramesh Nallapati and Christopher Manning
  • 16:55-17:20
    On the Extraction of Decisions and Contributions from Summaries of French Legal IT Contract Cases
    Manuel Maarek
  • 17:20-17:45
    Towards Annotating and Extracting Textual Legal Case Factors
    Adam Wyner and Wim Peters
  • 17:45-18:10
    Legal Rules Learning based on a Semantic Model for Legislation
    Enrico Francesconi