Fabio Barth
2026
SciLaD: A Large-Scale, Transparent, Reproducible Dataset for Natural Scientific Language Processing
Luca Foppiano | Sotaro Takeshita | Pedro Ortiz Suarez | Ekaterina Borisova | Raia Abu Ahmad | Malte Ostendorff | Fabio Barth | Julian Moreno-Schneider | Georg Rehm
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Luca Foppiano | Sotaro Takeshita | Pedro Ortiz Suarez | Ekaterina Borisova | Raia Abu Ahmad | Malte Ostendorff | Fabio Barth | Julian Moreno-Schneider | Georg Rehm
Proceedings of the Fifteenth Language Resources and Evaluation Conference
SciLaD is a novel, large-scale dataset of scientific language constructed entirely using open-source frameworks and publicly available data sources. It comprises a curated English split containing over 10 million scientific publications and a multilingual, unfiltered TEI XML split including more than 35 million publications. We also publish the extensible pipeline for generating SciLaD. The dataset construction and processing workflow demonstrates how open-source tools can enable large-scale, scientific data curation while maintaining high data quality. Finally, we pre-train a RoBERTa model on our dataset and evaluate it across a comprehensive set of benchmarks, achieving performance comparable to other scientific language models of similar size, validating the quality and utility of SciLaD. We publish the dataset and evaluation pipeline to promote reproducibility, transparency, and further research in natural scientific language processing and understanding including scholarly document processing.
PolyglotQL: A Pipeline for Multilingual Text-to-SPARQL Dataset Generation
Julio Perez | Fabio Barth | Georg Rehm
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Julio Perez | Fabio Barth | Georg Rehm
Proceedings of the Fifteenth Language Resources and Evaluation Conference
We present PolyglotQL, an open-source ETL (Extract, Transform, Load) pipeline for systematically creating multilingual text-to-SPARQL datasets, along with an accompanying framework for evaluating text-to-SPARQL generation models. PolyglotQL provides an extensible and modular architecture that aggregates, normalizes, and augments heterogeneous question–SPARQL pairs from established text-to-SPARQL datasets. With this pipeline, we automatically construct a bilingual English–German dataset featuring contextualized entity and relationship mappings as well as automatically translated and aligned question pairs. We also conduct an empirical evaluation using two multilingual open large language models under two distinct contextualization settings. The results show consistent performance improvements when explicit grounding information is provided, highlighting the benefits of structured context in multilingual semantic parsing.
2025
Familiarity: Better Evaluation of Zero-Shot Named Entity Recognition by Quantifying Label Shifts in Synthetic Training Data
Jonas Golde | Patrick Haller | Max Ploner | Fabio Barth | Nicolaas Jedema | Alan Akbik
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Jonas Golde | Patrick Haller | Max Ploner | Fabio Barth | Nicolaas Jedema | Alan Akbik
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Zero-shot named entity recognition (NER) is the task of detecting named entities of specific types (such as Person or Medicine) without any training examples. Current research increasingly relies on large synthetic datasets, automatically generated to cover tens of thousands of distinct entity types, to train zero-shot NER models. However, in this paper, we find that these synthetic datasets often contain entity types that are semantically highly similar to (or even the same as) those in standard evaluation benchmarks. Because of this overlap, we argue that reported F1 scores for zero-shot NER overestimate the true capabilities of these approaches. Further, we argue that current evaluation setups provide an incomplete picture of zero-shot abilities since they do not quantify the label shift (i.e., the similarity of labels) between training and evaluation datasets. To address these issues, we propose Familarity, a novel metric that captures both the semantic similarity between entity types in training and evaluation, as well as their frequency in the training data, to provide an estimate of label shift. It allows researchers to contextualize reported zero-shot NER scores when using custom synthetic training datasets. Further, it enables researchers to generate evaluation setups of various transfer difficulties for fine-grained analysis of zero-shot NER.
Table Understanding and (Multimodal) LLMs: A Cross-Domain Case Study on Scientific vs. Non-Scientific Data
Ekaterina Borisova | Fabio Barth | Nils Feldhus | Raia Abu Ahmad | Malte Ostendorff | Pedro Ortiz Suarez | Georg Rehm | Sebastian Möller
Proceedings of the 4th Table Representation Learning Workshop
Ekaterina Borisova | Fabio Barth | Nils Feldhus | Raia Abu Ahmad | Malte Ostendorff | Pedro Ortiz Suarez | Georg Rehm | Sebastian Möller
Proceedings of the 4th Table Representation Learning Workshop
Tables are among the most widely used tools for representing structured data in research, business, medicine, and education. Although LLMs demonstrate strong performance in downstream tasks, their efficiency in processing tabular data remains underexplored. In this paper, we investigate the effectiveness of both text-based and multimodal LLMs on table understanding tasks through a cross-domain and cross-modality evaluation. Specifically, we compare their performance on tables from scientific vs. non-scientific contexts and examine their robustness on tables represented as images vs. text. Additionally, we conduct an interpretability analysis to measure context usage and input relevance. We also introduce the TableEval benchmark, comprising 3017 tables from scholarly publications, Wikipedia, and financial reports, where each table is provided in five different formats: Image, Dictionary, HTML, XML, and LaTeX. Our findings indicate that while LLMs maintain robustness across table modalities, they face significant challenges when processing scientific tables.
2024
Occiglot at WMT24: European Open-source Large Language Models Evaluated on Translation
Eleftherios Avramidis | Annika Grützner-Zahn | Manuel Brack | Patrick Schramowski | Pedro Ortiz Suarez | Malte Ostendorff | Fabio Barth | Shushen Manakhimova | Vivien Macketanz | Georg Rehm | Kristian Kersting
Proceedings of the Ninth Conference on Machine Translation
Eleftherios Avramidis | Annika Grützner-Zahn | Manuel Brack | Patrick Schramowski | Pedro Ortiz Suarez | Malte Ostendorff | Fabio Barth | Shushen Manakhimova | Vivien Macketanz | Georg Rehm | Kristian Kersting
Proceedings of the Ninth Conference on Machine Translation
This document describes the submission of the very first version of the Occiglot open-source large language model to the General MT Shared Task of the 9th Conference of Machine Translation (WMT24). Occiglot is an open-source, community-based LLM based on Mistral-7B, which went through language-specific continual pre-training and subsequent instruction tuning, including instructions relevant to machine translation.We examine the automatic metric scores for translating the WMT24 test set and provide a detailed linguistically-motivated analysis.Despite Occiglot performing worse than many of the other system submissions, we observe that it performs better than Mistral7B, which has been based upon, which indicates the positive effect of the language specific continual-pretraining and instruction tuning. We see the submission of this very early version of the model as a motivation to unite community forces and pursue future LLM research on the translation task.
Search
Fix author
Co-authors
- Georg Rehm 4
- Pedro Ortiz Suarez 3
- Malte Ostendorff 3
- Raia Abu Ahmad 2
- Ekaterina Borisova 2
- Alan Akbik 1
- Eleftherios Avramidis 1
- Manuel Brack 1
- Nils Feldhus 1
- Luca Foppiano 1
- Jonas Golde 1
- Annika Grützner-Zahn 1
- Patrick Haller 1
- Nicolaas Jedema 1
- Kristian Kersting 1
- Vivien Macketanz 1
- Shushen Manakhimova 1
- Julian Moreno Schneider 1
- Sebastian Möller 1
- Julio Perez 1
- Max Ploner 1
- Patrick Schramowski 1
- Sotaro Takeshita 1