The 2nd Workshop on Practical LLM-assisted Data-to-Text Generation

News

  • 09/09/2024: The workshop programme and accepted papers list are now available!
  • 06/09/2024: Craig Thomson (ADAPT/DCU), and Marco Valentino (Neuro-Symbolic AI Group / Idiap Research Institute) confirmed as keynote speakers!
  • 24/07/2024: Based on multiple requests, we decided to update the deadline. The submission page is now open until Friday, July 26, 23.59 AoE. Good luck with your submissions!
  • 01/07/2024: We created a public Google group https://groups.google.com/g/public-d2t2024/ where you can follow the news and get technical feedback. For private matters see the /contacts page.
  • 28/06/2024: We now have a date! Practical D2T 2024 will be on 23 September @INLG 2024, Tokyo, Japan!
  • 20/06/2024: Call for paper is online!
  • 19/06/2024: the official Practical D2T 2024 website is online!

Background

Natural Language Generation (NLG) has been an active area of research for decades, both academically and industrially. Data-to-text (D2T) generation (Reiter and Dale, 1997; Gatt and Krahmer, 2018) is the NLG task where a system describes structured data in natural language. Traditionally, commercial D2T systems have been based on symbolic approaches (Dale, 2020; Leppänen et al., 2017), i.e. handcrafted rules or templates. More experimental approaches to D2T, such as E2E and Transformer-based systems (Dušek et al., 2020; Sharma et al., 2022) have been limited to research because of well-known issues like knowledge gaps, lack of factuality, and hallucination (Ji et al., 2023; Wang et al., 2023).

The recently introduced instruction-tuned, multi-task Large Language Models (LLMs) promise to become a viable alternative to rule-based D2T systems. They exhibit the ability to capture knowledge, follow instructions, and produce coherent text from various domains (Sanh et al., 2021; Ouyang et al., 2022). However, even the best LLMs still suffer from well-known issues of neural models, such as lack of controllability and risk of producing harmful text. Recent research thus proposed various approaches to improve the semantic accuracy of LLMs D2T, including prompt tuning (Su et al., 2022; Ye and Durrett, 2022), targeted fine-tuning (Zhang et al., 2024), Retrieval Augmented Generation (RAG) (Jiang et al., 2023; Chen et al., 2024), external tool integration (Wang et al., 2024), and neuro-symbolic approaches (Sarker et al, 2021; Hitzler et al., 2022).

Practical D2T 2024 aims to build a space for researchers to discuss and present innovative work on D2T systems using LLMs. Building upon the 2023 edition’s hackathon, Practical D2T 2024 opens up a broader range of activities, including a special track for neuro-symbolic D2T approaches and a hackathon focused on the evaluation and semantic accuracy of D2T using LLMs.

Important dates

Note: All deadlines are 23:59 UTC-12.

Main and special track:

  • First call for paper: 20 June
  • Regular paper submission: 22 July 26 July
  • Notification of acceptance: 19 August
  • Camera-ready: 28 August
  • Workshop: 23 September @INLG 2024, Tokyo, Japan

Invited Talks

Craig Thomson Marco Valentino
Craig Thomson,
ADAPT/DCU
Marco Valentino,
Neuro-Symbolic AI Group / Idiap Research Institute

Organizers

Simone Ballocu

Simone Balloccu

Charles University
Zdeněk Kasner

Zdeněk Kasner

Charles University
Ondřej Plátek

Ondřej Plátek

Charles University
Patricia Schmidtova

Patrícia Schmidtová

Charles University
Kristýna Onderková

Kristýna Onderková

Charles University
Mateusz Lango

Mateusz Lango

Charles University
Ondřej Dušek

Ondřej Dušek

Charles University
Lucie Flek

Lucie Flek

University of Bonn
Ehud Reiter

Ehud Reiter

University of Aberdeen
Dimitra Gkatzia

Dimitra Gkatzia

Edinburgh Napier University
Simon Mille

Simon Mille

ADAPT Centre

Previous Years

2023: 1st Workshop on Practical LLM-assisted Data-to-Text Generation

Acknowledgments

Funded by the European Union (ERC, NG-NLG, 101039303)

ERC