Yuzhe Yang's Avatar

Yuzhe Yang

@yuzheyang

AI Prof @ UCLA | RS @ Google | PhD @ MIT #ML, #AI, #health https://www.cs.ucla.edu/~yuzhe

1,672
Followers
70
Following
32
Posts
19.11.2024
Joined
Posts Following

Latest posts by Yuzhe Yang @yuzheyang

Preview
SleepLM: Natural-Language Intelligence for Human Sleep We present SleepLM, a family of sleep-language foundation models that enable human sleep alignment, interpretation, and interaction with natural language. Despite the critical role of sleep, learning-...

SleepLM points to a new direction for sleep AIπŸš€. Read all about it!
➑️Paper: arxiv.org/abs/2602.23605

Great work led by my students @ZongzheX2001, @ZitaoShuai, Eideen, and amazing collaborators @AysolaRavi and Rajesh!

More to comeπŸŒ™

#AI #sleep #sensor #health #multimodal #LLMs

05.03.2026 17:18 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Finally, we wanted this to connect to real clinical workflows. πŸ₯

SleepLM can combine its predictions across an entire night and produce useful full-night measures, while staying stable over long sequences. This matters as real sleep analysis is about understanding the whole night in a reliable way.

05.03.2026 17:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We also wanted the model to be more controllable. πŸŽ›οΈ

Instead of always generating one broad description, SleepLM can focus on a specific part of the physiology when asked. For example, it can emphasize 🧠brain activity, 🫁breathing, ❀️heart-related signals, or πŸ’ͺbody movement.

05.03.2026 17:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

SleepLM also learns when something happens, not just whether it happened. ⏱️

Our results show that the model is sensitive to timing. The strongest match appears when the text and the signal line up at the correct moment, and that match weakens as the alignment moves away.

05.03.2026 17:18 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

SleepLM learns a strong link between language and physiology. πŸ”„

When we ask it to match text to signals, or signals to text, it performs much better than general-purpose baselines. It not only reads sleep signals well β€” but also learns a shared space where signal and language line up closely.

05.03.2026 17:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

One clear takeaway: general LLMs are not enough. πŸ“Š

Even strong LLMs πŸ€– are not built for dense physiology. They often work with summaries, but struggle when the task depends on subtle waveform structure.

πŸ›Œ SleepLM is designed for that setting, and it shows clear gains on zero-shot sleep tasks.

05.03.2026 17:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

At the core is ReCoCa πŸ—οΈ, our unified training framework.

It combines three signals in one objective:
πŸ”— contrastive alignment
✍️ caption generation
♻️ signal reconstruction

The result is a representation that stays both language-aware and signal-grounded.

05.03.2026 17:18 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

Traditional sleep scoring compresses rich signals into a small set of labels. 🧩

We built a multilevel strategy to turn sleep into layered text descriptions. This gives a much richer view of sleep, enabling us to curate the first sleep-language dataset:

πŸ—‚οΈ100K+ hours of data from >10,000 people! πŸš€

05.03.2026 17:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

πŸŒ™ What if your sleep signals could speak?

Introducing SleepLM β€” sleep-language foundation models that turns raw sleep into something we can describe, query, and localize with language. πŸ—£οΈ

🌐Website: yang-ai-lab.github.io/SleepLM
πŸ•΅οΈCode: github.com/yang-ai-lab/...
πŸ€—Models: hf.co/yang-ai-lab/...

πŸ§΅πŸ‘‡

05.03.2026 17:18 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

πŸ“’ My lab at UCLA is hiring PhD students and postdocs!

Please apply to UCLA CS or CompMed and mention my name if you are interested in foundation models and (Gen)AI for health / medicine / science.

More info: cs.ucla.edu/~yuzhe

25.11.2025 07:27 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Preview
SensorLM: Learning the Language of Wearable Sensors We present SensorLM, a family of sensor-language foundation models that enable wearable sensor data understanding with natural language. Despite its pervasive nature, aligning and interpreting sensor ...

Read all about it!
➑️Paper: arxiv.org/abs/2506.09108

Huge team effort! Kudos to my intern Evelyn, amazing team @kmr_ayush, @aametwally1, @Orson_Xu, @timalthoff, @pushmeet, @cecim, @xliucs, @danmcduff, and other amazing co-authors!

#AI #wearable #sensor #health #multimodal
(8/8)

17.06.2025 15:38 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Beyond its discriminative power, SensorLM showcases compelling generative capabilities. It can produce hierarchical and realistic captions from input wearable data only, offering more coherent & correct descriptions compared to LLMs like Gemini 2.0 Flash. ✍️✨

(7/8)

17.06.2025 15:38 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

SensorLM also demonstrates intriguing capabilities, including interesting scaling behavior over data size, model size, and compute. πŸ“ˆπŸ’‘

(6/8)

17.06.2025 15:38 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Experiments across real-world tasks in human activity analysis πŸƒβ€β™€οΈ & healthcare βš•οΈ showcase its superior performance over SOTA models in:
✨ Zero-shot recognition
✨ Few-shot learning
✨ Cross-modal retrieval

(5/8)

17.06.2025 15:38 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

SensorLM extends prominent multimodal pretraining architectures (e.g., contrastive, generative) unifying their principles for sensor data. It extends prior approaches, recovering them as specific configurations within a single architecture. πŸ—οΈπŸ”—

(4/8)

17.06.2025 15:38 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

This enabled us to curate the largest sensor-language dataset to date: over 59.7 million hours of data from >103,000 people. That's orders of magnitude larger than prior studies! πŸš€πŸ’Ύ

(3/8)

17.06.2025 15:38 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Despite its pervasiveness, aligning & interpreting sensor data with language remains challenging πŸ“ˆ due to the lack of richly annotated sensor-text descriptions. 🚫

Our solution? A hierarchical pipeline captures statisticalπŸ“Š, structuralπŸ—οΈ, and semantic🧠 sensor info.

(2/8)

17.06.2025 15:38 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

🚨 Let your wearable data "speak" for themselves! βŒšοΈπŸ—£οΈ

Introducing *SensorLM*, a family of sensor-language foundation models, trained on ~60 million hours of data from >103K people, enabling robust wearable sensor data understanding with natural language. 🧡

17.06.2025 15:38 πŸ‘ 6 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
Demographic bias of expert-level vision-language foundation models in medical imaging Compared to certified radiologists, expert-level AI models show notable and consistent demographic biases across pathologies.

πŸ©»βš–οΈ AI underdiagnoses Black female patients

A new study found that expert-level vision-language models for chest X-rays systematically underdiagnose marginalised groups – especially Black women – more than radiologists.

πŸ”— doi.org/10.1126/sciadv.adq0305

#SciComm #AI #HealthEquity πŸ§ͺ

03.04.2025 08:28 πŸ‘ 17 πŸ” 8 πŸ’¬ 1 πŸ“Œ 1
Preview
AI models miss disease in Black and female patients Analysis of chest x-rays underscores need for monitoring artificial intelligence tools for bias, experts say

Science News provides a great cover of our paper: www.science.org/content/arti...

Started in 2023, delayed but finally out! Huge congrats & thanks to amazing collaborators: Yujia, @xliucs, @Avanti0609, @Mastrodicasa_MD, Vivi, @ejaywang, @sahani_dushyant, Shwetak πŸŽ‰

(6/6)
#AI #health #fairness

28.03.2025 20:01 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Why the gap? These foundation models in medical imaging encode demographic info (age, sex, race) from X-raysβ€”more than humans do! Fascinating, but a challenge for fair healthcare βš–οΈ.

(5/)

28.03.2025 20:01 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

This fairness disparity also holds for unseen pathologies during training, as well as for differential diagnoses across 50+ pathologies. βš•οΈ

(4/)

28.03.2025 20:01 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

While expert-level VLMs can achieve _overall_ diagnosis accuracy on par with clinicians, they show significant underdiagnosis disparity over (intersectional) subpopulations vs. Radiologists 🚨

(3/)

28.03.2025 20:01 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We tested top vision-language models like CheXzero on 5 global datasets 🌍. Result? They consistently show disparities in diagnosis based on race, sex, and ageβ€”esp. across marginalized groupsβ€”compared to certified radiologists πŸ“·

(2/)

28.03.2025 20:01 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Do foundation models in medical imaging see everyone fairly?πŸ€”

Excited to share our new Science Advances paper uncovering & auditing demographic biases of expert-level VLMs, and comparing to board-certified radiologistsπŸ§‘β€βš•οΈ

πŸ“„science.org/doi/10.1126/sciadv.adq0305
πŸ’»github.com/YyzHarry/vlm-fairness
(1/)

28.03.2025 20:01 πŸ‘ 28 πŸ” 7 πŸ’¬ 1 πŸ“Œ 0
Automated loss of pulse detection on a consumer smartwatch - Nature Nature - Automated loss of pulse detection on a consumer smartwatch

How multimodal A.I. of real time smartwatch data can automatically detect a person's loss of pulseβ€”sudden cardiac deathβ€”and notify emergency services
www.nature.com/articles/s41...
@jakesunshine.bsky.social @nature.com

26.02.2025 17:46 πŸ‘ 186 πŸ” 36 πŸ’¬ 12 πŸ“Œ 2
Preview
Deep profiling of gene expression across 18 human cancers - Nature Biomedical Engineering Using unsupervised deep learning to generate low-dimensional latent spaces for gene-expression data can unveil biological insight across cancers.

Just published in Nature Biomedical Engineering! Working with the incredible PhD student Wei Qiu and our brilliant collaborator Kamila Naxerova at Harvard was a great pleasure. Our deep profiling framework enables us to view 18 human cancers through the lens of AI!

www.nature.com/articles/s41...

18.12.2024 01:02 πŸ‘ 25 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Post image

A neurologist with 2 APOE4 copies tells us about his experience with #Alzheimers disease
washingtonpost.com/wellness/202...

17.12.2024 15:32 πŸ‘ 377 πŸ” 100 πŸ’¬ 17 πŸ“Œ 7
Post image

Seven years ago, Scott Lundberg, presented our SHAP framework at the NeurIPS 2017 conference. Since then, SHAP has become one of the most widely used feature attribution methods, with our paper receiving approximately 30,000 citations. It's wonderful that SHAP's birthday aligns perfectly with mine!😊

06.12.2024 03:20 πŸ‘ 47 πŸ” 5 πŸ’¬ 0 πŸ“Œ 1

I will be at #NeurIPS and #ML4H all next week β€” let me know if you would like to catch up in person!

πŸ“’ I am also recruiting PhD students! Drop me an email if you're attending NeurIPS and would like to chat or learn more πŸ˜€

04.12.2024 18:56 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0