Sebastian Lehner's Avatar

Sebastian Lehner

@sebaleh

Postdoc JKU Linz ELLIS: ML for sampling and optimization.

866
Followers
250
Following
6
Posts
06.02.2024
Joined
Posts Following

Latest posts by Sebastian Lehner @sebaleh

Post image

I'm excited to open the new year by sharing a new perspective paper.

I give a informal outline of MD and how it can interact with Generative AI. Then, I discuss how far the field has come since the seminal contributions, such as Boltzmann Generators, and what is still missing

16.01.2026 10:25 πŸ‘ 19 πŸ” 5 πŸ’¬ 1 πŸ“Œ 1
Post image

Measuring AI Progress in Drug Discovery - A NEW LEADERBOARD IN TOWN

2015-2025: turns out that there's hardly any improvement. AI bubble?

GPT is at 70% for this task, whereas the best methods get close to 85%.

Leaderboard: huggingface.co/spaces/ml-jk...
P: arxiv.org/abs/2511.14744

19.11.2025 06:52 πŸ‘ 12 πŸ” 7 πŸ’¬ 3 πŸ“Œ 3

thanks!

05.10.2025 18:05 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Posting a few nice importance sampling-related finds

"Value-aware Importance Weighting for Off-policy Reinforcement Learning"

proceedings.mlr.press/v232/de-asis...

04.10.2025 16:01 πŸ‘ 3 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Post image

Returning soon - stay tuned!

sites.google.com/view/monte-c...

18.09.2025 18:59 πŸ‘ 21 πŸ” 7 πŸ’¬ 0 πŸ“Œ 1
Post image

I am very happy to finally share something I have been working on and off for the past year:

"The Information Dynamics of Generative Diffusion"

This paper connects entropy production, divergence of vector fields and spontaneous symmetry breaking

link: arxiv.org/abs/2508.19897

02.09.2025 16:40 πŸ‘ 21 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Post image

xLSTM for multivariate time series anomaly detection: arxiv.org/abs/2506.22837

β€œIn our results, xLSTM showcases state-of-the-art accuracy, outperforming 23 popular anomaly detection baselines.”

Again, xLSTM excels in time series analysis.

01.07.2025 08:30 πŸ‘ 4 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

New paper on the generalization of Flow Matching www.arxiv.org/abs/2506.03719

🀯 Why does flow matching generalize? Did you know that the flow matching target you're trying to learn *can only generate training points*?

w @quentinbertrand.bsky.social @annegnx.bsky.social @remiemonet.bsky.social πŸ‘‡πŸ‘‡πŸ‘‡

18.06.2025 08:08 πŸ‘ 55 πŸ” 17 πŸ’¬ 2 πŸ“Œ 3
Post image

New preprint alert 🚨
How can you guide diffusion and flow-based generative models when data is scarce but you have domain knowledge? We introduce Minimum Excess Work, a physics-inspired method for efficiently integrating sparse constraints.
Thread below πŸ‘‡https://arxiv.org/abs/2505.13375

26.05.2025 09:13 πŸ‘ 26 πŸ” 6 πŸ’¬ 1 πŸ“Œ 0
Preview
MHNfs: Prompting In-Context Bioactivity Predictions for Low-Data Drug Discovery Today’s drug discovery increasingly relies on computational and machine learning approaches to identify novel candidates, yet data scarcity remains a significant challenge. To address this limitation,...

Need to predict bioactivity πŸ§ͺ but only have limited data ❌?

Try our interactive app for prompting MHNfs β€” a state-of-the-art model for few-shot molecule–property prediction. No coding or training needed. πŸš€

πŸ“„ Paper:
pubs.acs.org/doi/10.1021/...

πŸ–₯️ App:
huggingface.co/spaces/ml-jk...

13.05.2025 08:27 πŸ‘ 7 πŸ” 5 πŸ’¬ 0 πŸ“Œ 1
Preview
Optimal Transport for Machine Learners Optimal Transport is a foundational mathematical theory that connects optimization, partial differential equations, and probability. It offers a powerful framework for comparing probability distributi...

I have cleaned a bit my lecture notes on Optimal Transport for Machine Learners arxiv.org/abs/2505.06589

13.05.2025 05:18 πŸ‘ 119 πŸ” 29 πŸ’¬ 0 πŸ“Œ 0
Post image

Many recent posts on free energy. Here is a summary from my class β€œStatistical mechanics of learning and computation” on the many relations between free energy, KL divergence, large deviation theory, entropy, Boltzmann distribution, cumulants, Legendre duality, saddle points, fluctuation-response…

02.05.2025 19:22 πŸ‘ 63 πŸ” 9 πŸ’¬ 1 πŸ“Œ 0

I asked "on the other platform" what were the most important improvements to the original 2017 transformer.

That was quite popular and here is a synthesis of the responses:

28.04.2025 06:47 πŸ‘ 204 πŸ” 43 πŸ’¬ 4 πŸ“Œ 3
Post image

Come check out SDE Matching at the #ICLR2025 workshops, a new simulation-free framework for training fully general Latent/Neural SDEs (generalisation of diffusion and bridge models).

FPI: Morning poster session
DeLTa: Afternoon poster session

#SDE #Bayes #GenAI #Diffusion #Flow

27.04.2025 23:27 πŸ‘ 13 πŸ” 1 πŸ’¬ 1 πŸ“Œ 1
Post image

Excited to present our poster on Boltzmann priors for Implicit Transfer Operators tomorrow at @iclr-conf.bsky.social!
See you tomorrow at poster 13, 10-12:30.

24.04.2025 08:20 πŸ‘ 11 πŸ” 5 πŸ’¬ 0 πŸ“Œ 0
Post image

1/11 Excited to present our latest work "Scalable Discrete Diffusion Samplers: Combinatorial Optimization and Statistical Physics" at #ICLR2025 on Fri 25 Apr at 10 am!
#CombinatorialOptimization #StatisticalPhysics #DiffusionModels

24.04.2025 08:57 πŸ‘ 16 πŸ” 7 πŸ’¬ 1 πŸ“Œ 0
The Scientist Building an 'Artificial Scientist'
The Scientist Building an 'Artificial Scientist' YouTube video by Quanta Magazine

My video interview with @quantamagazine.bsky.social about AI-designed physics experiments, AI as a Muse for new ideas in Science, and Artificial Scientists: www.youtube.com/watch?v=T_2Z...

19.03.2025 10:46 πŸ‘ 18 πŸ” 7 πŸ’¬ 0 πŸ“Œ 2
Preview
Digital Discovery of Interferometric Gravitational Wave Detectors AI-driven design of gravitational wave detectors uncovers approaches that surpass current plans, potentially boosting sensitivity more than tenfold.

πŸ“’ AI-discovered Gravitational Wave Detectors

published in @apsphysics.bsky.social Phys.Rev.X, with Rana Adhikari & Yehonathan Drori @ligo.org @caltech.edu @mpi-scienceoflight.bsky.social

journals.aps.org/prx/abstract...

Extremely happy to see this paper online after 3.5 years of work.

🧡1/5

14.04.2025 17:32 πŸ‘ 11 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Post image

We have been reworking the Quickstart guide of POT to show multiple examples of OT with the unified API that facilitates access to OT value/plan/potentials. It allows to select regularization/unbalancedness/lowrank/Gaussian OT with just a few parameters. pythonot.github.io/master/auto_...

26.03.2025 07:39 πŸ‘ 32 πŸ” 11 πŸ’¬ 0 πŸ“Œ 0
Post image

xLSTM 7B: A Recurrent LLM for Fast and Efficient Inference

Meet the fastest 7B language model out there. Based on the mLSTM!

P: arxiv.org/abs/2503.13427

18.03.2025 06:33 πŸ‘ 1 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Post image

Tweedie's formula is super important in diffusion models & is also one of the cornerstones of empirical Bayes methods.

Given how easy it is to derive, it's surprising how recently it was discovered ('50s). It was published a while later when Tweedie wrote Stein about it

1/n

18.03.2025 06:12 πŸ‘ 65 πŸ” 14 πŸ’¬ 1 πŸ“Œ 0

Opportunity to work with @hochreitersepp.bsky.social , @jobrandstetter.bsky.social , and me!!

We have many open positions in machine learning, deep learning, LLMs!! Both for PostDocs and PhDs!

Join us!

14.03.2025 12:31 πŸ‘ 2 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Preview
Challenging the Quantum Advantage Frontier with Large-Scale Classical Simulations of Annealing Dynamics Recent demonstrations of D-Wave's annealing-based quantum simulators have established new benchmarks for quantum computational advantage [arXiv:2403.00910]. However, the precise location of the classi...

We provide a classical simulation of DWave quantum "s-word" paper.
Here it is arxiv.org/abs/2503.08247 , great work by Linda Mauron at the CQS Lab, check it out! (1/4)

12.03.2025 09:30 πŸ‘ 46 πŸ” 7 πŸ’¬ 4 πŸ“Œ 1
Post image

I shared a controversial take the other day at an event and I decided to write it down in a longer format: I’m afraid AI won't give us a "compressed 21st century"

Here: thomwolf.io/blog/scienti...

It's an extension of this interview discussion from the AI summit: youtu.be/AxBd3G0lFLs?...

06.03.2025 13:03 πŸ‘ 133 πŸ” 34 πŸ’¬ 11 πŸ“Œ 12
Post image

My new paper "Deep Learning is Not So Mysterious or Different": arxiv.org/abs/2503.02113. Generalization behaviours in deep learning can be intuitively understood through a notion of soft inductive biases, and formally characterized with countable hypothesis bounds! 1/12

05.03.2025 15:37 πŸ‘ 210 πŸ” 49 πŸ’¬ 6 πŸ“Œ 9

Thanks @zlatko-minev.bsky.social and hello bluesky world!

04.03.2025 15:47 πŸ‘ 13 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0

Luca (Martino) once told me (when I said "MCMC does not have weights") that this is incorrect (in his Sicilian style): When you reject in MCMC, you increase the weight of the current sample. Chains do have replicates, can be written like a weighted sample. High rejection rate *is* weight degeneracy.

28.02.2025 11:29 πŸ‘ 4 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

Excited to share our work with friends from MIT/Google on Learned Asynchronous Decoding! LLM responses often contain chunks of tokens that are semantically independent. What if we can train LLMs to identify such chunks and decode them in parallel, thereby speeding up inference? 1/N

27.02.2025 00:38 πŸ‘ 16 πŸ” 9 πŸ’¬ 1 πŸ“Œ 1
Post image

Excited about our progress in characterizing The Computational Advantage of Depth in Learning with Neural Networks. Check out the number of samples that can be saved when GD runs on a multi-layer rather than on a two-layer neural network. arxiv.org/pdf/2502.13961

22.02.2025 14:23 πŸ‘ 27 πŸ” 4 πŸ’¬ 1 πŸ“Œ 1

πŸ“’PSA: #NeurIPS2024 recordings are now publicly available!

The workshops always have tons of interesting things on at once, so the FOMO is realπŸ˜΅β€πŸ’« Luckily it's all recorded, so I've been catching up on what I missed.

Thread below with some personal highlights🧡

22.01.2025 21:06 πŸ‘ 128 πŸ” 33 πŸ’¬ 1 πŸ“Œ 1