The Ada Lovelace Algorithmic Lectures are a series of interactive lectures in which we branch, explain and prove... in an algorithmic way. For each edition, we invite one or more researchers specialising in large language models from outside the host institution, Vrije Universiteit Brussel (VUB), whose work has caught our attention. This could be a critical point of view, interesting position or just excellent research that aligns with the interests of the Data Lab. A core aspect of our scientific ambitions is to push the boundaries of foundational AI research. We are renowned for our interdisciplinary approach, integrating physics, mathematics, information theory, and AI to tackle fundamental scientific challenges. Through this series, we aim to increase collaboration and discussion between scientists who satisfy the aforementioned constraints, in honour of the world's first software engineer, Ada Lovelace.
02/02/2026
Sam Klein (Public AI) on how to poison a model if you must
TLDR: Sam explained--without slides!--how small triggers (e.g. hidden strings) that are present on platforms, documents and messages can cause biases (e.g. the color blue is good and the color red is bad) in AI systems developed in healthcare and (all) other domains, and that they can be very hard to detect. But also, that there is a new, interesting defence "SCOUT": Paper.
We also talked about positive poisoning, where in the training of LLMs poisoning could do more harm, synthetic data generation that directly integrates the model, a socratic method for LLMs, youtube and wikipedia!
24/09/2025
Dr. Pieter Delobelle on the challenges of training LLMs for mid-resource languages.
TLDR: Pieter explained that tokenisation can be problematic for agglutinative languages such as Dutch and German, and how synthetic data can address this issue and scale up the performance of non-English LLMs (without requiring larger and larger models).
Recent research. Slides: Dutch LLMs--Challenges of training LLMs for mid-resource languages
20/05/2025
Dr. Desi R. Ivanova on the reliability of Large Language Models and their evaluations.
TLDR: Desi reminded us to check the conditions of the central limit theorem (CLT) when using error bars in LLM evaluations and restarted the broader discussion on how to measure and improve reliability of LLMs.
Paper: Position: Don’t Use the CLT in LLM Evals With Fewer Than a Few Hundred Datapoints. Slides: Reliability of LLMs and LLM Evaluations.