Research

My research sits at the intersection of probability and statistics, mostly focusing on hypothesis testing, statistical learning theory, and online learning. I've recently been working on sequential testing with e-values, the generalisation and expressiveness of over-parameterised models, and mean/parameter estimation. Here are some highlights of my recent work.

Sequential testing and e-values

Classical statistical hypothesis testing is based on p-values, which do not allow for changing plans mid-experiment. For example, you cannot peek at the results and stop early because the evidence looks strong, or decide on the significance level after having evaluated the p-value. Testing via e-values is a modern and more flexible approach, which is attracting growing interest. It is based on the concept of an e-variable: a non-negative random variable whose expectation under the null is at most one. Its realisation is an e-value. For a concise introduction see link, for a deeper (and often updated) treatment see the monograph by Ramdas and Wang (link). My recent work focuses on characterising the set of e-variables for constraint-defined hypotheses (e.g., testing for the mean), with a particular focus on minimal complete classes (what I call optimal e-classes) of e-variables, and the sequential analogue for e-processes.

Algorithmic statistics

Several recent works have have investigated ways to address classical statistical challenges by leveraging tools and techniques from online learning. A common pattern is to break the analysis into a stochastic and a worst-case part: the former is handled via standard martingale arguments, whilst the latter is controlled by the regret of an online learning game. I usually call this approach algorithmic statistics (this might not be a widely used term, I picked it up from Gergely Neu as I think it's a nice name). Some examples of algorithmic statitistics I am becoming rather familiar with are the online-to-PAC framework for generalisation bounds (link), sequential hypothesis testing via e-values (link), or obtaining confidence sequences via regret analysis (link , link). My work on the subject has mostly focused on introducing a general framework for confidence sets for the unknown parameter of generalised linear models, and on obtaining generalisation and concentration guarantees under data inter-dependencies.

Generalisation bounds

In machine learning, generalisation refers to a model's ability to perform well on unseen data, after training on a limited dataset. A neural network’s parameters are tuned by optimising a learning objective that ensures good accuracy on the training data. Upper bounds on the gap between performance on this training dataset and expected performance on new data are known as generalisation bounds. Beyond theoretical interest, these results have practical value, as they help assess the reliability of a model. My research has focused on information-theoretic (link) and PAC-Bayesian (link) approaches for generalisation bounds, both of which build on a notion of stability: a model generalises well if it is not overly sensitive to the particular dataset used for training. These frameworks capture this idea by comparing distributions over model parameters before and after training.