Uncertainty-Aware Agentic AI

Research Theme

Multi Agent Systems and Data Intelligence

Aim

To develop a principled framework for representing, communicating, and aggregating uncertainty in large language model (LLM)–based agent systems, thus improving coordination, calibration and safety in multi-agent reasoning.

Objectives

  1. Elicit Uncertainty: Develop and compare methods for extracting calibrated uncertainty estimates from individual LLMs.
  2. Represent & Communicate: Design message formats that carry uncertainty information between agents (e.g., probabilistic statements, confidence intervals, prediction sets).
  3. Propagate & Aggregate: Establish mathematically sound rules for combining and updating uncertainty as messages pass between agents.
  4. Evaluate Impact: Create benchmarks and metrics to quantify how uncertainty-aware communication affects accuracy, calibration, and robustness of multi-agent systems.

Description

State of the art solutions in many complex problem domains are often now “Agentic” systems in which collaborations of specialist LLM-based agents collectively solve problems. For example, in cyber-security, surveillance agents may monitor logs to detect potential exploits, and cooperate with patching agents that fix vulnerabilities, and testing agents that validate the results. In such systems agents communicate through natural-language messages that reflect single “best guesses.” In probabilistic terms, these are point estimates of a model’s belief. As messages are exchanged, uncertainty information is lost, leading to over-confidence, error cascades, and poor coordination under ambiguity.

This project investigates how to restore and maintain epistemic and aleatoric uncertainty throughout an agentic pipeline. The first stage will focus on discovering the most reliable way to elicit and perceive uncertainty estimates with single LLMs. The second stage will formalize uncertainty-aware message formats—for example, by attaching confidence scores, probabilistic distributions, or prediction sets to text outputs. The third stage will develop algorithms for propagating and aggregating uncertainties across agents (eg Bayesian, Dempster-Shafer, conformal fusion). Finally, we will build benchmarks that measure team-level uncertainty calibration. Expected outcomes include (i) theoretical understanding of uncertainty composition in LLM collectives, (ii) practical mechanisms for safer, more transparent AI collaboration, and (iii) open-source tools for uncertainty calibration and propagation in agentic frameworks.

Further information

https://arxiv.org/abs/2306.10193

https://aclanthology.org/2021.acl-long.84/

Closing date: 
Apply now

Principal Supervisor

Eligibility

Minimum entry qualification - an Honours degree at 2:1 or above (or International equivalent) in a relevant science or engineering discipline, possibly supported by an MSc Degree. Further information on English language requirements for EU/Overseas applicants.

Funding

Full funding is available for this position.