Ilyass Moummad's Avatar

Ilyass Moummad

@ilyassmoummad

Postdoctoral Researcher @ Inria Montpellier (IROKO, Pl@ntNet) SSL for plant images Interested in Computer Vision, Natural Language Processing, Machine Listening, and Biodiversity Monitoring Website: ilyassmoummad.github.io

392
Followers
352
Following
68
Posts
18.11.2024
Joined
Posts Following

Latest posts by Ilyass Moummad @ilyassmoummad

Loria hiring Master internship in Nancy, Grand Est, France | LinkedIn Posted 5:06:23 PM. Master 2 Research Internship – Acoustic Aware Speech Enhancement in Distributed Microphone Arrays…See this and similar jobs on LinkedIn.

We have a master internship position with possible PhD extension at @labo-loria.bsky.social
The work focuses on on speech enhancement with distributed microphone array.
www.linkedin.com/jobs/view/43...

18.12.2025 17:15 👍 1 🔁 1 💬 0 📌 0

*Darkthrone played in the background* 😂

04.12.2025 08:58 👍 1 🔁 0 💬 1 📌 0
Preview
DBD-research-group/Bird-MAE-Base · Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

bonus point (the feature extractor is available on HF): huggingface.co/DBD-research...

21.11.2025 09:37 👍 0 🔁 0 💬 0 📌 0

Check out the amazing work by my collaborators Lukas and René, who trained a ViT-based masked autoencoder to learn patterns in mel-spectrograms of bird vocalizations without annotations, and then cleverly aggregated the learned features to solve downstream bird species classification tasks. 🐦

21.11.2025 09:28 👍 5 🔁 1 💬 1 📌 0

[10/10] Wrap-up 🎯
🔹 Unified supervised + unsupervised hashing
🔹 Flexible: works via probing or LoRA
🔹 SOTA hashing in minutes on a single GPU

📄 Paper: arxiv.org/abs/2510.27584
💻 Code: github.com/ilyassmoumma...

Shoutout to my wonderful co-authors Kawtar, Hervé, and Alexis.

03.11.2025 14:31 👍 2 🔁 0 💬 0 📌 0

[9/10] Strong generalization 🌍
CroVCA produces compact codes that transfer efficiently:
✅ Single HashCoder trained on ImageNet-1k works on downstream datasets without retraining (More experiments and ablations in the paper)

03.11.2025 14:31 👍 0 🔁 0 💬 1 📌 0
Post image

[8/10] Semantically consistent retrieval 🔍
CroVCA retrieves correct classes even for fine-grained or ambiguous queries (e.g., indigo bird, grey langur).
✅ Outperforms Hashing-Baseline
✅ Works with only 16 bits and without supervision

03.11.2025 14:31 👍 0 🔁 0 💬 1 📌 0
Post image

[7/10] Compact yet meaningful codes 💾
Even with just 16 bits, CroVCA preserves class structure.
t-SNE on CIFAR-10 shows clear, separable clusters — almost identical to the original 768-dim embeddings.

03.11.2025 14:31 👍 0 🔁 0 💬 1 📌 0
Post image

[6/10] Strong performance across encoders 💪
Tested on multiple vision encoders (SimDINOv2, DINOv2, DFN…), CroVCA achieves SOTA unsupervised hashing:

03.11.2025 14:30 👍 0 🔁 0 💬 1 📌 0

[5/10] Fast convergence 🚀
CroVCA trains in just ~5 epochs:
✅ COCO (unsupervised) <2 min
✅ ImageNet100 (supervised) ~3 min
✅ Single GPU
Despite simplicity, it achieves state-of-the-art retrieval performance.

03.11.2025 14:30 👍 1 🔁 0 💬 1 📌 0
Post image

[4/10] HashCoder 🛠️
A lightweight MLP with final BatchNorm for balanced bits (inspired by OrthoHash). Can be used as:
🔹 Probe on frozen features
🔹 LoRA-based fine-tuning for efficient encoder adaptation

03.11.2025 14:30 👍 1 🔁 0 💬 1 📌 0
Post image

[3/10] Unifying hashing 🔄
Can supervised + unsupervised hashing be done in one framework?

CroVCA aligns binary codes across semantically consistent views:
Augmentations → unsupervised
Class-consistent samples → supervised

🧩 One BCE loss + coding-rate regularizer

03.11.2025 14:29 👍 1 🔁 1 💬 1 📌 0

[2/10] The challenge ⚡
Foundation models (DINOv3, DFN, SWAG…) produce rich embeddings, but similarity search in high-dimensional spaces is expensive.
Hashing provides fast Hamming-distance search, yet most deep hashing methods are complex, slow, and tied to a single paradigm.

03.11.2025 14:29 👍 1 🔁 0 💬 1 📌 0

[1/10] Introducing CroVCA ✨
A simple, unified framework for supervised and unsupervised hashing that converts foundation model embeddings into compact binary codes.
✅ Preserves semantic structure
✅ Trains in just a few iterations

03.11.2025 14:29 👍 3 🔁 0 💬 1 📌 0
BioDCASE Workshop - BioDCASE Join us for the BioDCASE Workshop held in Barcelona, Spain on the 29th of October! The workshop will be held at the Campus del Poblenou of Universitat Pompeu Fabra. The BioDCASE workshop will be hosted the day before the DCASE workshop on the 30-31st of October at the same venue …

BioDCASE workshop - registration closes next week Oct 10th https://biodcase.github.io/workshop2025/ - Hope to see you there! #bioacoustics

03.10.2025 10:17 👍 8 🔁 10 💬 1 📌 0

I heard that the Linux client is buggy, I use it on the browser and it's working ok.

09.09.2025 07:03 👍 1 🔁 0 💬 1 📌 0

for the curious, the code, slides and the article are on Github: github.com/BastienPasde...

29.08.2025 11:44 👍 4 🔁 1 💬 0 📌 0

love it haha wish I were there to hear Prostitute Disfigurement in an amphitheater

29.08.2025 11:40 👍 1 🔁 0 💬 2 📌 0
Preview
NAVIGU: a powerful image collection explorer. NAVIGU lets you dive into the ocean of images. Drag the image sphere or double-click on an image you like to browse large collections.

A website to visually browse and explore the ImageNet-1k dataset (there are other supported datasets: IN-12M, WikiMedia, ETH Images, Pixabay, Fashion) navigu.net#imagenet
(Maybe this is already known, but I was happy to discover it this morning)

27.08.2025 07:39 👍 1 🔁 0 💬 0 📌 0

Im interested in the quantum and footnotesize, how much params should they have 😂

23.08.2025 06:31 👍 0 🔁 0 💬 0 📌 0
Learning Deep Representations of Data Distributions Landing page for the book Learning Deep Representations of Data Distributions.

Learning Deep Representations of Data Distributions
Sam Buchanan · Druv Pai · Peng Wang · Yi Ma

ma-lab-berkeley.github.io/deep-represe...

The best Deep Learning book is out, I've been waiting for its release for more than a year. Let's learn how to build intelligent systems via compression.

23.08.2025 06:27 👍 4 🔁 0 💬 0 📌 0

It feels like we can now fit more noise with more model capacity 🤔 (Figure 6), maybe we need newer architectures and/or newer training losses.

19.08.2025 21:36 👍 0 🔁 0 💬 0 📌 0
Post image

1/ Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research.

21.07.2025 14:47 👍 83 🔁 21 💬 2 📌 3

👋 I worked on bioacoustics during my PhD, but I post mostly about AI

18.07.2025 20:56 👍 0 🔁 0 💬 0 📌 0
Research Scientist Build tabular foundation models and shape how the world works with its most valuable data. Opportunity to work on fundamental breakthroughs such as multimodal, causality and specialized architectures.

🏹 Job alert: Research Scientist at Prior Labs

📍Freiburg or Berlin 🇩🇪
📅 Apply by Dec 31 - preferably earlier
🔗 More info: https://bit.ly/4kqn5rY

04.07.2025 06:45 👍 6 🔁 3 💬 0 📌 0

Congratz! 👏

03.07.2025 10:19 👍 1 🔁 0 💬 0 📌 0
Of Petrichor Weaves Black Noise
Of Petrichor Weaves Black Noise YouTube video by Ne Obliviscaris - Topic

my new addiction today: youtu.be/dSyJqwN36ow
I can't wait to see them this summer in Motocultor Festival

19.06.2025 09:54 👍 0 🔁 0 💬 0 📌 0

the best discovery I've had in recent years, I'm addicted to it now as well 😁

19.06.2025 07:12 👍 1 🔁 0 💬 1 📌 0

Thank you for making this accessible to everyone! I've read some sections, it is very instructive.

16.06.2025 10:08 👍 0 🔁 0 💬 0 📌 0
Preview
Foundations of Computer Vision The print version was published by

Our computer vision textbook is now available for free online here:
visionbook.mit.edu

We are working on adding some interactive components like search and (beta) integration with LLMs.

Hope this is useful and feel free to submit Github issues to help us improve the text!

15.06.2025 15:45 👍 117 🔁 32 💬 3 📌 1