April 2nd, 2026 - Delft, The Netherlands

Text2Story 2026

Ninth International Workshop on Narrative Extraction from Texts
held in conjunction with the 48th European Conference on Information Retrieval

Call for papers

Overview

For eight years, Text2Story has gathered researchers on narrative understanding, detecting, representing, and reasoning over stories in text. Recent progress and discussions have been driven by Transformers and Large Language Models (LLMs), which are increasingly central to the field and have reshaped approaches to narrative extraction and interpretation. Despite rapid progress, challenges persist in fine-grained narrative structure, dialogue and multimodal narratives, robustness to domain shift, and evaluation beyond surface accuracy. The ninth edition of the workshop will spotlight the growing convergence of Information Retrieval (IR) with Transformers/LLMs and agentic AI. By centering these IR-aligned problems, Text2Story aims to advance both fundamental insights and practical applications in narrative search, understanding, and generation

Call for papers

Research papers submitted to the workshop should advance the understanding, modeling, and application of narratives across text and multimodal contexts. We encourage work on narrative extraction, representation, analysis, and generation, particularly in connection with Transformers, LLMs, and Agentic AI, as well as research that bridges narrative understanding and information retrieval. We encourage the submission of high-quality and original submissions covering the following topics and contributions focused on low and medium-resource languages.

    Narrative Information Extraction and Representation

  • Identification of participants, events, and temporal expressions
  • Temporal and causal relation extraction, ordering, and reasoning
  • Annotation protocols and narrative representation models
  • Lexical, Syntactic, and Semantic Ambiguity in Narrative Representation
  • LLMs for narrative extraction, understanding, and reasoning
  • Narrative Analysis and Generation

  • Argumentation and discourse analysis
  • Language models and transfer learning for narrative analysis
  • Comprehension of generated narratives
  • Narrative generation in low-resource and multilingual settings
  • Narrative text simplification, anonymization, and personalization
  • Narratives, LLMs, and Agentic AI

  • Understanding narratives in human–agent communication
  • Multi-turn and multi-agent narrative understanding
  • Consistency, coherence, and trust in agent-generated narratives
  • Narratives for agent planning, reasoning, and decision-making
  • Evaluating and benchmarking LLMs for narrative tasks
  • Narrative analysis of LLM-generated text
  • Data creation and augmentation with LLMs
  • Evaluation of narrative extraction with LLMs
  • Datasets, Resources, and Evaluation

  • Evaluation methodologies for narrative extraction and understanding
  • Creation and use of annotated datasets
  • Narrative-focused resources and benchmarks
  • LLMs as annotators and as curators
  • Ethics, Bias, and Misinformation

  • Bias detection and mitigation in generated stories
  • Ethical and fair narrative generation
  • Narratives in misinformation, disinformation, and fact-checking
  • Narrative Applications

  • Narrative-focused search and retrieval
  • Narrative summarization
  • Narrative-based question answering (Q&A)
  • Narrative-based user profiling
  • Sentiment and opinion detection in narratives
  • Social media narratives and public discourse
  • Automatic timeline generation and story evolution tracking
  • Storyline visualization and interactive exploration
  • Multimodal narratives (integration of text, image, audio, and video)

Objectives

Overall, the workshop aims to: (1) raise awareness within the Information Retrieval (IR) and NLP communities about the challenges and opportunities in narrative extraction, comprehension, and generation; (2) bridge the gap between academic research, industrial practice, and applied systems involving LLMs and Agentic AI for narrative tasks; (3) foster discussion on new methods, recent advances, and emerging challenges in narrative understanding, reasoning, and evaluation; (4) share experiences from projects, case studies, and applications structured around key research questions in narrative representation and generation; (5) explore the impact of automation, LLMs, and agentic systems on the creation and interpretation of narratives; (6) encourage reflection on both successful and unexpected outcomes in narrative research, contributing to a deeper understanding of limitations and open problems.

Important Dates

  • January 23rd, 2026
    Submission Deadline
  • February 21st, 2026
    Acceptance Notification
  • March 15th, 2026
    Camera-ready copies
  • April 2nd, 2026
    Workshop

Submissions

We expect contributions from researchers addressing all aspects of narrative extraction, representation, analysis, and generation. This includes the detection and formal modeling of events, participants, and their temporal and causal relations, as well as methods for reasoning and orderinge. Submissions that explore narrative comprehension and interpretation, such as the analysis or evaluation of LLM-generated narratives, are particularly encouraged.

We also invite work on innovative ways to present and interact with narrative information, including automatic timeline construction, multimodal summarization, and narrative visualization. Research tackling misinformation, bias, and verification of extracted facts, along with the development of datasets, annotation schemas, and evaluation methodologies, is highly valued. Contributions addressing low and medium-resource languages, multilingual and cross-lingual settings are especially welcome.

Building on these themes, the field now faces pressing questions that can guide authors in shaping their submissions. How can LLMs and Agentic AI be leveraged for more coherent, grounded, and explainable narrative understanding? What strategies enable the integration of multimodal content—text, image, audio, video—into unified and trustworthy narratives? How can models dynamically adapt to evolving narratives across domains, genres, and languages? What methods best support the evaluation and benchmarking of narrative systems, from extraction to generation and reasoning? How can we ensure transparency, fairness, and robustness in narrative systems that may influence real-world perceptions and decisions? To what extent can collaboration between human experts and automated systems enhance interpretability, cultural awareness, and inclusivity in narrative understanding?

Full Papers

up to 8 pages + references

Original and high-quality unpublished contributions to the theory and practical aspects of the narrative extraction task. Full papers should introduce existing approaches, describe the methodology and the experiments conducted in detail. Negative result papers to highlight tested hypotheses that did not get the expected outcome are also welcomed.

Short Papers

up to 5 pages + references

Unpublished short papers describing work in progress; position papers introducing a new point of view, a research vision or a reasoned opinion on the workshop topics; and dissemination papers describing project ideas, ongoing research lines, case studies or summarized versions of previously published papers in high-quality conferences/journals that is worthwhile sharing with the Text2Story community, but where novelty is not a fundamental issue.

Demos | Resource Papers

up to 5 pages + references

Unpublished papers presenting research/industrial demos; papers describing important resources (datasets or software packages) to the Text2Story community;


Papers must be submitted electronically in PDF format through Easy Chair . All submissions must be in English and formatted according to the one-column CEUR-ART style with no page numbers. Templates, available in either LaTeX or ODT (LibreOffice) format, can be found in the following zip folder. Do not use Word for the ODT template. CEUR requires the use of the Libertinus font family. Instructions on installing these fonts are found in the ODT template. There is also an Overleaf page for LaTeX users. Please also note that your work must include a Declaration on Generative AI to comply with the CEUR-WS GenAI policy. Both the LaTeX and LibreOffice templates already include this required section.

IMPORTANT: Please include between brackets the type of submission (full; negative results; work in progress; demo and resource; position; dissemination) in the paper title.

Papers submitted to Text2Story 2026 should be original work and different from papers that have been previously published, accepted for publication, or that are under review at other venues. Exceptions to this rule are "dissemination papers". Pre-prints submitted to ArXiv are eligible.

All papers will be refereed through a double-blind peer-review process by at least two members of the programme committee. We plan to publish the workshop proceedings as a CEUR volume, which will ensure they are indexed by DBLP and made available online via open access. Authors are therefore responsible for ensuring that their submissions do not infringe on any existing publication rights.

Organization

Organizing Committee

Program Committee

  • Alvaro Figueira (CRACS / INESC TEC and University of Porto)
  • Andreas Spitz (University of Konstanz)
  • Antoine Doucet (University of La Rochelle)
  • António Branco (Faculdade de Ciências da Universidade de Lisboa)
  • Arian Pasquali (University of Porto)
  • Bart Gajderowicz (University of Toronto)
  • Behrooz Mansouri (University of Southern Maine)
  • Brenda Salenave Santana (Universidade Federal de Pelotas)
  • Brucce Santos (Laboratório de Inteligência Artificial)
  • Bruno Martins (IST and INESC-ID - Instituto Superior Técnico, University of Lisbon)
  • Dhruv Gupta (Norwegian University of Science and Technology)
  • Evelin Amorim (INESC TEC)
  • Federico Pianzola (University of Groningen)
  • Florian Boudin (Université de Nantes)
  • Henrique Lopes Cardoso (University of Porto)
  • Hugo Sousa (University of Porto)
  • Irina Rabaev (Shamoon College of Engineering, SCE, Beer Sheva, Israel)
  • Ismail Sengor Altingovde (Middle East Technical University)
  • Jakub Piskorski (Joint Research Centre of the European Commission)
  • Joao Cordeiro (Universidade Beira Interior)
  • Joaquim Silva (Nova School of Science and Technology - NOVA LINCS)
  • Junbo Huang (University of Hamburg)
  • Lise Stork (University of Amsterdam)
  • Luca Cagliero (Politecnico di Torino)
  • Ludovic Moncla (LIRIS (UMR 5205 CNRS), INSA Lyon)
  • Luís Filipe Cunha (University of Minho)
  • Marc Spaniol (Université de Caen Normandie)
  • Marcelo Finger (University of São Paulo)
  • Mark Finlayson (Florida International University)
  • Nuno Guimarães (LIAAD - INESC TEC)
  • Pablo Gervás (Universidad Complutense de Madrid)
  • Paul Rayson (Lancaster University)
  • Paulo Quaresma (Universidade de Evora)
  • Purificação Silvano (Faculdade de Letras da Universidade do Porto)
  • Ricardo Marcacini (ICMC/USP)
  • Ross Purves (University of Zurich)
  • Sara Tonelli (Fondazione Bruno Kessler)
  • Sérgio Nunes (INESC TEC and Faculty of Engineering, University of Porto)
  • Sérgio Matos (IEETA, Universidade de Aveiro)
  • Sriharsh Bhyravajjula (UNIVERSITY OF WASHINGTON)
  • Thiago Pardo (University of São Paulo)
  • Udo Kruschwitz (University of Regensburg)
  • Valentina Bartalesi (ISTI-CNR)
  • Veysel Kocaman (John Snow Labs)
  • Yangyang Chen (Brandeis University)

Web and Dissemination Chair

  • Luís Filipe Cunha (INESC TEC & University of Porto)
  • Behrooz Mansouri (University of Southern Maine)

Invited Speakers

Knowledge Extraction from Fiction Using Generative Large Language Models and Multilingual Semantic Parsing

Speaker: Federico Pianzola, University of Groningen, The Netherlands

Keynote Card Federico Pianzola

Abstract: Extracting event knowledge from unstructured text is a well-known challenge in natural language processing (NLP) and is particularly difficult when dealing with fiction. Subtext, instead of explicit information, and figurative style in fictional narratives complicate event extraction. Recent advances in Large Language Models (LLMs) have improved performance in various NLP tasks. However, their effective ness in extracting events from fiction remains underexplored. In this talk, I present an evaluation of different techniques to extract narrative information from fiction, including the prompting of open-weights LLMs and the semantic parsing via Abstract Meaning Representation (AMR).
This work has been done in the context of the GOLEM project (“Graphs and Ontologies for Literary Evolution Models”).

Bio: Federico Pianzola is Assistant Professor of Computational Humanities at the University of Groningen (Netherlands), where he coordinates the Master’s in Digital Humanities. He is the Principal Investigator of the ERC Starting Grant project GOLEM and previously completed a Marie Sklodowska-Curie Global Fellowship at the University of Milan-Bicocca (Italy) and Sogang University (South Korea). He is member of the scientific advisory board of OPERAS (the Research Infrastructure supporting open scholarly communication in the social sciences and humanities in the European Research Area) and president of the International Society of the Empirical Study of Literature (IGEL). He is also managing editor of the journal Scientific Study of Literature and he is on the editorial board of three other journals: the Journal of Computational Literary Studies, the Journal of Cultural Analytics, and the Korean Journal of Digital Humanities. His most recent book Digital Social Reading: Sharing Fiction in the 21st Century has been published by MIT Press in January 2025.


Evaluating Generative AI and Agentic Systems: Methodologies for Systematic Assessment and Continuous Improvement

Speaker: Arian Pasquali, Faktion.com Antwerp, Belgium

Keynote Card Arian Pasquali

Abstract: As large language models and AI agents define the current era of enterprise AI adoption across business, government, and public sectors, organizations face a critical challenge: how to systematically evaluate and monitor these systems at scale while maintaining meaningful human oversight. This keynote addresses the realities encountered in real projects and the urgent need to comprehensively review evaluation methodologies for generative AI and agentic systems, from narrative extraction to generation and reasoning, with particular emphasis on the critical role of human-in-the-loop approaches in this context.
We begin by examining why traditional evaluation approaches and academic benchmarks fall short for generative AI in enterprise settings. While LLM benchmarks may show impressive scores on standardized tests, they often fail to predict real-world performance where outputs are diverse, context-dependent, domain-specific, and subject to business constraints. The talk presents a framework for designing domain-specific evaluation workflows that combine quantitative metrics, human judgment, emerging techniques such as LLM-as-a-Judge approaches and its challenges. We explore practical methods for collecting and analyzing evaluation signals from real-world usage while addressing the most common implementation difficulties.
A central focus is the transition from evaluation-driven research in laboratory settings to evaluation-driven development in industry practice. We discuss how to approach systematic continuous improvement of generative AI systems, including synthetic data generation, domain-specific metrics, data segmentation, and related tasks to enable development lifecycles with iterative improvements. By the end of this keynote, attendees will gain an industry perspective on how to design, implement, and operationalize evaluation methodologies that enable the development and deployment of trustworthy generative AI and agentic systems.

Bio: Arian Pasquali is an Applied AI Researcher at Orq.ai, an AI Engineering & Evaluation Platform focused on Generative AI and AI agents for enterprises, based in Amsterdam. He specializes in NLP and LLMs, with focus on evaluation and annotation workflows applied to AI agents. During the last years he has been bridging the gap between research and practical applications across government, education, and enterprise domains. His current research interests focus on designing processes for the systematic evaluation of large language models and agentic architectures. He has published research in top-tier conferences on keyword extraction, temporal narrative generation, and interactive systems for information retrieval. Arian is also an active contributor to the academic community, where he has served as a reviewer and chair for major international conferences, including ECIR and ECML-PKDD. His contributions have been recognized with several awards, including the World Summit Award 2019 for Government and Citizen Engagement, as well as paper awards at ECIR.

Attending


Text2Story 2026 will be held at the 48th European Conference on Information Retrieval (ECIR'26) in Delft, The Netherlands.

Registration at ECIR 2026 is required to attend the workshop (don't forget to select the Text2Story workshop).

Acknowledgements

This work is financed by National Funds through the FCT - Fundação para a Ciência e a Tecnologia, I.P. (Portuguese Foundation for Science and Technology) within the project StorySense, with reference 2022.09312.PTDC (DOI 10.54499/2022.09312.PTDC).