Why do I have to pretend that I'm going to print something in order to save it as a PDF. Why do I have to engage in a little ruse.
Why do I have to pretend that I'm going to print something in order to save it as a PDF. Why do I have to engage in a little ruse.
...Maria Teodora Wetscherek, Klaus Maier-Hein, Panagiotis Korfiatis, @valesalvatelli.bsky.social, Javier Alvarez-Valle.
🧵12/12
Extra kudos to Tassilo Wald, who led the execution!
Thanks to everyone else involved in the project: @ibrahimethem.bsky.social, Yuan (William) Gao, @sambondtaylor.bsky.social, Harshita Sharma, @maxilse.bsky.social, Cynthia Lo, Olesya Melnichenko, @anton-sc.bsky.social, Noel Codella...
🧵11/12
COLIPRI is very easy to install and use! Just run `pip install colipri` and paste this snippet (from aka.ms/colipri) to get started.
I'm looking forward to seeing what the community will build on top of our model. Get in touch if you have questions or feedback!
🧵10/12
We used nnSSL (Wald et al., ICCV 2025) for training, TorchIO (@fepegar.com et al., CMPB 2021) for preprocessing and augmentation, nnU-Net (Isensee et al., Nature Methods 2021) for segmentation, and nifti-zarr-py for efficient patch loading from during cloud training.
🧵9/12
COLIPRI is generally superior to concurrent methods across all tasks. This is particularly clear when plugging an MLLM on top of our vision backbone. Our models are particularly stronger at clinical metrics, which are most relevant in practice.
🧵8/12
To overcome this domain shift, we introduce an Opposite Sentence Loss (OSL), a simple but effective mechanism that complementes the contrastive loss and improves our metrics substantially.
🧵7/12
Medical reports are often very long and most sentences describe what is *not* in the scan. However, for zero-shot classification, users tend to use very short prompts, such as "Lung nodules" and "No lung nodules".
🧵6/12
We generate reports during training to ensure that the vision encoder extracts from the image all the information that would be needed for reporting, similar to CapPa (@mtschannen.bsky.social et al., NeurIPS 2023).
🧵5/12
We resampled the volumes to 2-mm isotropic spacing using and used an input size of 160^3. We randomly shuffled and shortened sentences in the reports used for contrastive alignment. We initialised our encoder from CXR-BERT (Boecking, @naotous.bsky.social et al., ECCV 2022).
🧵4/12
We first pre-train our encoder only on images (no reports) sourced from different datasets, using a 3D MAE (Wald et al., CVPR 2025). This allows us to leverage more training data, as we did for Rᴀᴅ-DINO (@fepegar.com et al., Nature Machine Intelligence 2025).
🧵2/12
There is not a lot of paired 3D medical image–report data out there. A pioneering example is CT-RATE (@iethemhamamci, arxiv.org/abs/2403.17834), with samples from 21k patients, a number much smaller than what we see in the natural imaging domain.
🧵1/12
We are excited to release the weights of @msftresearch.bsky.social's COLIPRI, our 3D vision–language encoder for chest CT scans, on @hf.co 🤗
Model: aka.ms/colipri
Demo: aka.ms/colipri-demo
Paper: aka.ms/colipri-paper
Why does COLIPRI matter?
🧵0/12 👇
Grading and googling hallucinated citations, as one does nowadays, and now that LLMs have been around for a while, I've discovered new horrors: hallucinated journals are now appearing in Google Scholar with dozens of citations bc so many people are citing these fake things
This article frames the problem as [AI?] slop. Which is a problem, but not the main one here. This is an issue with authorship norms and practices. A single individual putting their name on hundreds of (workshop) papers they admitted they had little part in.
Opening slide for a presentation titled "From medical image interpretation to scientific discovery"
Excited to speak shortly at the Medical Imaging at @euripsconf.bsky.social workshop in Copenhagen, where I'll share some insights from our @msftresearch.bsky.social team's journey "From medical image interpretation to scientific discovery" over the past couple of years.
SpaceX is ready for its next Transporter mission! With 140 satellites onboard, this is the largest Transporter mission since Transporter-1 in 2021, which carried 143 satellites. Here's my identification attempt. Launch is scheduled for NET 18:19 UTC.
TLDR; The PSF has made the decision to put our community and our shared diversity, equity, and inclusion values ahead of seeking $1.5M in new revenue. Please read and share. pyfound.blogspot.com/2025/10/NSF-...
🧵
🩻Excited to share our latest preprint: “Data Scaling Laws for Radiology Foundation Models”
Foundation vision encoders like CLIP and DINOv2 have transformed general computer vision, but what happens when we scale them for medical imaging?
📄 Read the full preprint here: arxiv.org/abs/2509.12818
bsky.app/profile/fons...
Photo with Keir Starmer and copy that reads: RECOGNITION OF A PALESTINIAN STATE IS A HOLLOW GESTURE WITHOUT MEANINGFUL ACTION TO END ISRAEL'S GENOCIDE, APARTHEID & OCCUPATION
Recognition is no doubt significant but it will be a hollow gesture if the UK does not also seek to end Israel's genocide, illegal occupation, and system of apartheid against the Palestinian people.
🧵1/3
Yeah man we should really fight back by staying on X
Having some fun with DINOv3 and PCA! Although I'm not happy my nose has such a low foreground probability :D
Email addresses are very simple, and you will score highly in this quiz.
e-mail.wtf
Introducing DINOv3 🦕🦕🦕
A SotA-enabling vision foundation model, trained with pure self-supervised learning (SSL) at scale.
High quality dense features, combining unprecedented semantic and geometric scene understanding.
Three reasons why this matters👇
We’re not sure who needs to hear this, but ‘blueberry’ has two b’s.
title and abstract from https://arxiv.org/pdf/2507.19960
table 1 from https://arxiv.org/pdf/2507.19960
Boiling here at home in Cyprus but I put the finishing touches a couple of days ago on this preprint: What Does 'Human-Centred AI' Mean? doi.org/10.48550/arX...
Wherein I analyse HCAI & demonstrate through 3 triplets my new tripartite definition of AI (Table 1) that properly centres the human. 1/n
Scientists overwhelmingly recognize the value of sharing null results, but rarely publish them in the research literature
go.nature.com/450KElr
i love it when people make PRs really easy to review ❤️
feels really bad when there's a PR that's been open for a while, but i know i can't easily do a good job of reviewing it, so it's constantly "later"