I wrote a long, detailed blog post on the state of AI and machine learning.
The post's purpose is to sharpen my thinking and help ensure I work on the right things over the next few years.
You might find parts of it interesting. Comments are welcome.
avt.im/blog/where-a...
02.02.2026 17:50
👍 9
🔁 0
💬 0
📌 0
After a delayed flight and missing my first poster, I have officially arrived at NeurIPS!
I’ll present another poster tomorrow at CDE 606 from 11-2. I’ll post more on this soon.
If you’re interested in meeting up, let’s get in touch!
05.12.2025 00:57
👍 6
🔁 0
💬 0
📌 0
Today, I gave a talk at the INFORMS Job Market Showcase!
If you're interested, here are the slides - link below!
presentations.avt.im/2025-10-26-A...
26.10.2025 20:16
👍 6
🔁 1
💬 0
📌 0
I am hiring a fully-funded #PhD in #ML to work at the University of Edinburgh on 𝐠𝐞𝐨𝐦𝐞𝐭𝐫𝐢𝐜 𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 and 𝐮𝐧𝐜𝐞𝐫𝐭𝐚𝐢𝐧𝐭𝐲 𝐪𝐮𝐚𝐧𝐭𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧.
Application deadline: 31 Dec '25. Starts May/Sep '26.
Details in the reply.
Pls RT and share with anyone interested!
24.10.2025 11:35
👍 16
🔁 7
💬 1
📌 3
So this sums up the work! If you followed along, thanks for the interest!
I think you'd agree that "Bayesian Algorithms for Adversarial Online Learning: from Finite to Infinite Action Spaces" is a much better title than before. The old one was much harder to pronounce.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
The Bayesian viewpoint proves useful for developing this analysis.
It allows us to guess what a good prior will be, and suggests ways to use probability as a tool to prove the algorithm works.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
We prove that the Bayesian approach works in this setting too.
To achieve this, we develop a new probabilistic analysis of correlated Gaussian follow-the-perturbed-leader algorithms, of which ours is a special case.
This has been an open challenge in the area.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
The second one is where X = [0,1]^d and Y is the space of bounded Lipschitz functions.
Here, you can't use a prior with independence across actions. You need to share information between actions.
We do this by using a Gaussian process, with correlations between actions.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
The first one is the classical discrete setting where standard algorithms such as exponential weights are studied.
You can use a Gaussian prior which is independent across actions.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
Okay, so we now know what "Bayesian Algorithms for Adversarial Online Learning" are.
What about "from Finite to Infinite Action Spaces"?
This covers the two settings we show the aforementioned results in.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
This approach appears to not make any sense: the Bayesian model is completely fake.
We're pretending to know a distribution for how the adversary will act in the future.
But, in reality, they can do anything.
And yet... we show that this works!
29.09.2025 19:55
👍 2
🔁 0
💬 1
📌 0
We show that this game secretly has a natural Bayesian strategy - one we show is strong.
What's the strategy?
It's really simple:
- Place a prior distribution of what the adversary will do in the future
- Condition on what the adversary has done
- Sample from the posterior
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
One observation about adversarial online learning is that it appears to have nothing to do with Bayesian learning.
There is a two-player zero-sum game, not a joint probability distribution.
So you can't just solve it by applying Bayes' Rule. Or can you?
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
Okay, so now we understand what "Adversarial Online Learning" is.
We propose "Bayesian Algorithms" for this.
What does that mean? Let's unpack.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
Online learning is therefore a good model for learning to explore by taking random actions.
In contrast to other approaches to resolving explore-exploit tradeoffs such as upper confidence bounds which produce purely deterministic strategies.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
So that's our setting. Why's it interesting?
Because many other hard decision problems can be reduced to online learning, including certain forms of reinforcement learning (via decision-estimation coefficients), equilibrium computation (via no-regret dynamics), and others.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
Specifically, their goal is to minimize regret
R(p,q) = E_{x_t~p_t, y_t~q_t} \sup_{x\in X} \sum_{t=1}^T y_t(x) - \sum_{t=1}^T y_t(x_t).
Meaning, the learner compares the sum of their rewards y_t(x_t) with the sum of y_t(x) for the best possible single non-time-dependent x.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
The learner's goal is to achieve the highest rewards possible. But at each time, the adversary can choose a different reward function.
So why is this game not impossible?
Because the learner only compares how well they do with the *sum* of the adversary's previous rewards.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
"Adversarial Online Learning" refers to a two-player zero-sum repeated game between a learner and adversary.
At each time point:
- The learner chooses a distribution of predictions p_t over an action space X.
- The adversary chooses a reward function y_t : X -> R.
29.09.2025 19:55
👍 1
🔁 0
💬 1
📌 0
Paper update: our recent work on Thompson sampling has a shiny new - and I hope much better - name!
This new name does much better job of emphasizing what we actually do.
Joint work with Jeff Negrea. Thread below!
29.09.2025 19:55
👍 10
🔁 3
💬 1
📌 0
Stochastic Poisson Surface Reconstruction with One Solve using Geometric Gaussian Processes
Project page: geospsr.github.io
Paper link: arxiv.org/abs/2503.19136
Link to my student's tweets on this work: x.com/sholalkere/s...
17.07.2025 18:24
👍 3
🔁 0
💬 0
📌 0
At ICML, we're presenting a paper on uncertainty-aware surface reconstruction!
Compared to previous approaches, we are able to completely remove the need for recursive linear solves for reconstruction and interpolation, using geometric GP machinery.
Check it out!
17.07.2025 18:24
👍 10
🔁 3
💬 1
📌 0
This week’s virtual seminar on Bayesian Decision-making and Uncertainty is happening now!
Noémie Jaquier (KTH Royal Institute of Technology)
On Riemannian Latent Variable Models and Pullback Metrics
Livestream link: www.youtube.com/watch?v=61Be...
02.06.2025 16:06
👍 4
🔁 0
💬 1
📌 0
We obtain a very simple condition which relates the prior covariance kernel with the adversary's function class in an easy-to-verify way that works in the bounded Lipschitz case.
I am very interested in extensions to more-general smoothness classes, and have ideas. Stay tuned!
30.05.2025 17:42
👍 1
🔁 0
💬 1
📌 0
This stands in contrast with prior arguments, which are linear-algebraic in flavor, involve bounding certain matrix norms by certain traces, and essentially-require independence in order to give sharp rates.
30.05.2025 17:42
👍 0
🔁 0
💬 1
📌 0
Specifically, we have a novel probabilistic argument which bounds Hessian-type terms which appear in the regret analysis of Gaussian follow-the-perturbed-leader algorithms, of which Thompson sampling is a special case.
Our argument works even with correlations!
30.05.2025 17:42
👍 0
🔁 0
💬 1
📌 0