SleepLM: Natural-Language Intelligence for Human Sleep
We present SleepLM, a family of sleep-language foundation models that enable human sleep alignment, interpretation, and interaction with natural language. Despite the critical role of sleep, learning-...
SleepLM points to a new direction for sleep AIπ. Read all about it!
β‘οΈPaper: arxiv.org/abs/2602.23605
Great work led by my students @ZongzheX2001, @ZitaoShuai, Eideen, and amazing collaborators @AysolaRavi and Rajesh!
More to comeπ
#AI #sleep #sensor #health #multimodal #LLMs
05.03.2026 17:18
π 1
π 0
π¬ 0
π 0
Finally, we wanted this to connect to real clinical workflows. π₯
SleepLM can combine its predictions across an entire night and produce useful full-night measures, while staying stable over long sequences. This matters as real sleep analysis is about understanding the whole night in a reliable way.
05.03.2026 17:18
π 0
π 0
π¬ 1
π 0
We also wanted the model to be more controllable. ποΈ
Instead of always generating one broad description, SleepLM can focus on a specific part of the physiology when asked. For example, it can emphasize π§ brain activity, π«breathing, β€οΈheart-related signals, or πͺbody movement.
05.03.2026 17:18
π 0
π 0
π¬ 1
π 0
SleepLM also learns when something happens, not just whether it happened. β±οΈ
Our results show that the model is sensitive to timing. The strongest match appears when the text and the signal line up at the correct moment, and that match weakens as the alignment moves away.
05.03.2026 17:18
π 1
π 0
π¬ 1
π 0
SleepLM learns a strong link between language and physiology. π
When we ask it to match text to signals, or signals to text, it performs much better than general-purpose baselines. It not only reads sleep signals well β but also learns a shared space where signal and language line up closely.
05.03.2026 17:18
π 0
π 0
π¬ 1
π 0
One clear takeaway: general LLMs are not enough. π
Even strong LLMs π€ are not built for dense physiology. They often work with summaries, but struggle when the task depends on subtle waveform structure.
π SleepLM is designed for that setting, and it shows clear gains on zero-shot sleep tasks.
05.03.2026 17:18
π 0
π 0
π¬ 1
π 0
At the core is ReCoCa ποΈ, our unified training framework.
It combines three signals in one objective:
π contrastive alignment
βοΈ caption generation
β»οΈ signal reconstruction
The result is a representation that stays both language-aware and signal-grounded.
05.03.2026 17:18
π 1
π 0
π¬ 1
π 0
Traditional sleep scoring compresses rich signals into a small set of labels. π§©
We built a multilevel strategy to turn sleep into layered text descriptions. This gives a much richer view of sleep, enabling us to curate the first sleep-language dataset:
ποΈ100K+ hours of data from >10,000 people! π
05.03.2026 17:18
π 0
π 0
π¬ 1
π 0
π What if your sleep signals could speak?
Introducing SleepLM β sleep-language foundation models that turns raw sleep into something we can describe, query, and localize with language. π£οΈ
πWebsite: yang-ai-lab.github.io/SleepLM
π΅οΈCode: github.com/yang-ai-lab/...
π€Models: hf.co/yang-ai-lab/...
π§΅π
05.03.2026 17:18
π 2
π 0
π¬ 1
π 0
π’ My lab at UCLA is hiring PhD students and postdocs!
Please apply to UCLA CS or CompMed and mention my name if you are interested in foundation models and (Gen)AI for health / medicine / science.
More info: cs.ucla.edu/~yuzhe
25.11.2025 07:27
π 2
π 1
π¬ 0
π 0
SensorLM: Learning the Language of Wearable Sensors
We present SensorLM, a family of sensor-language foundation models that enable wearable sensor data understanding with natural language. Despite its pervasive nature, aligning and interpreting sensor ...
Read all about it!
β‘οΈPaper: arxiv.org/abs/2506.09108
Huge team effort! Kudos to my intern Evelyn, amazing team @kmr_ayush, @aametwally1, @Orson_Xu, @timalthoff, @pushmeet, @cecim, @xliucs, @danmcduff, and other amazing co-authors!
#AI #wearable #sensor #health #multimodal
(8/8)
17.06.2025 15:38
π 1
π 0
π¬ 0
π 0
Beyond its discriminative power, SensorLM showcases compelling generative capabilities. It can produce hierarchical and realistic captions from input wearable data only, offering more coherent & correct descriptions compared to LLMs like Gemini 2.0 Flash. βοΈβ¨
(7/8)
17.06.2025 15:38
π 0
π 0
π¬ 1
π 0
SensorLM also demonstrates intriguing capabilities, including interesting scaling behavior over data size, model size, and compute. ππ‘
(6/8)
17.06.2025 15:38
π 0
π 0
π¬ 1
π 0
Experiments across real-world tasks in human activity analysis πββοΈ & healthcare βοΈ showcase its superior performance over SOTA models in:
β¨ Zero-shot recognition
β¨ Few-shot learning
β¨ Cross-modal retrieval
(5/8)
17.06.2025 15:38
π 0
π 0
π¬ 1
π 0
SensorLM extends prominent multimodal pretraining architectures (e.g., contrastive, generative) unifying their principles for sensor data. It extends prior approaches, recovering them as specific configurations within a single architecture. ποΈπ
(4/8)
17.06.2025 15:38
π 0
π 0
π¬ 1
π 0
This enabled us to curate the largest sensor-language dataset to date: over 59.7 million hours of data from >103,000 people. That's orders of magnitude larger than prior studies! ππΎ
(3/8)
17.06.2025 15:38
π 0
π 0
π¬ 1
π 0
Despite its pervasiveness, aligning & interpreting sensor data with language remains challenging π due to the lack of richly annotated sensor-text descriptions. π«
Our solution? A hierarchical pipeline captures statisticalπ, structuralποΈ, and semanticπ§ sensor info.
(2/8)
17.06.2025 15:38
π 1
π 0
π¬ 1
π 0
π¨ Let your wearable data "speak" for themselves! βοΈπ£οΈ
Introducing *SensorLM*, a family of sensor-language foundation models, trained on ~60 million hours of data from >103K people, enabling robust wearable sensor data understanding with natural language. π§΅
17.06.2025 15:38
π 6
π 1
π¬ 1
π 0
Demographic bias of expert-level vision-language foundation models in medical imaging
Compared to certified radiologists, expert-level AI models show notable and consistent demographic biases across pathologies.
π©»βοΈ AI underdiagnoses Black female patients
A new study found that expert-level vision-language models for chest X-rays systematically underdiagnose marginalised groups β especially Black women β more than radiologists.
π doi.org/10.1126/sciadv.adq0305
#SciComm #AI #HealthEquity π§ͺ
03.04.2025 08:28
π 17
π 8
π¬ 1
π 1
AI models miss disease in Black and female patients
Analysis of chest x-rays underscores need for monitoring artificial intelligence tools for bias, experts say
Science News provides a great cover of our paper: www.science.org/content/arti...
Started in 2023, delayed but finally out! Huge congrats & thanks to amazing collaborators: Yujia, @xliucs, @Avanti0609, @Mastrodicasa_MD, Vivi, @ejaywang, @sahani_dushyant, Shwetak π
(6/6)
#AI #health #fairness
28.03.2025 20:01
π 2
π 0
π¬ 0
π 0
Why the gap? These foundation models in medical imaging encode demographic info (age, sex, race) from X-raysβmore than humans do! Fascinating, but a challenge for fair healthcare βοΈ.
(5/)
28.03.2025 20:01
π 3
π 0
π¬ 1
π 0
This fairness disparity also holds for unseen pathologies during training, as well as for differential diagnoses across 50+ pathologies. βοΈ
(4/)
28.03.2025 20:01
π 2
π 0
π¬ 1
π 0
While expert-level VLMs can achieve _overall_ diagnosis accuracy on par with clinicians, they show significant underdiagnosis disparity over (intersectional) subpopulations vs. Radiologists π¨
(3/)
28.03.2025 20:01
π 2
π 0
π¬ 1
π 0
We tested top vision-language models like CheXzero on 5 global datasets π. Result? They consistently show disparities in diagnosis based on race, sex, and ageβesp. across marginalized groupsβcompared to certified radiologists π·
(2/)
28.03.2025 20:01
π 2
π 0
π¬ 1
π 0
Do foundation models in medical imaging see everyone fairly?π€
Excited to share our new Science Advances paper uncovering & auditing demographic biases of expert-level VLMs, and comparing to board-certified radiologistsπ§ββοΈ
πscience.org/doi/10.1126/sciadv.adq0305
π»github.com/YyzHarry/vlm-fairness
(1/)
28.03.2025 20:01
π 28
π 7
π¬ 1
π 0
Automated loss of pulse detection on a consumer smartwatch - Nature
Nature - Automated loss of pulse detection on a consumer smartwatch
How multimodal A.I. of real time smartwatch data can automatically detect a person's loss of pulseβsudden cardiac deathβand notify emergency services
www.nature.com/articles/s41...
@jakesunshine.bsky.social @nature.com
26.02.2025 17:46
π 186
π 36
π¬ 12
π 2
Deep profiling of gene expression across 18 human cancers - Nature Biomedical Engineering
Using unsupervised deep learning to generate low-dimensional latent spaces for gene-expression data can unveil biological insight across cancers.
Just published in Nature Biomedical Engineering! Working with the incredible PhD student Wei Qiu and our brilliant collaborator Kamila Naxerova at Harvard was a great pleasure. Our deep profiling framework enables us to view 18 human cancers through the lens of AI!
www.nature.com/articles/s41...
18.12.2024 01:02
π 25
π 3
π¬ 0
π 0
A neurologist with 2 APOE4 copies tells us about his experience with #Alzheimers disease
washingtonpost.com/wellness/202...
17.12.2024 15:32
π 377
π 100
π¬ 17
π 7
Seven years ago, Scott Lundberg, presented our SHAP framework at the NeurIPS 2017 conference. Since then, SHAP has become one of the most widely used feature attribution methods, with our paper receiving approximately 30,000 citations. It's wonderful that SHAP's birthday aligns perfectly with mine!π
06.12.2024 03:20
π 47
π 5
π¬ 0
π 1
I will be at #NeurIPS and #ML4H all next week β let me know if you would like to catch up in person!
π’ I am also recruiting PhD students! Drop me an email if you're attending NeurIPS and would like to chat or learn more π
04.12.2024 18:56
π 2
π 0
π¬ 0
π 0