Lukas Galke's Avatar

Lukas Galke

@lukasgalke

Assistant Professor @SDU tracing connectionist mechanisms. https://lgalke.github.io

123
Followers
367
Following
10
Posts
07.02.2024
Joined
Posts Following

Latest posts by Lukas Galke @lukasgalke

Preview
Isolating Culture Neurons in Multilingual Large Language Models Language and culture are deeply intertwined, yet it is so far unclear how and where multilingual large language models encode culture. Here, we extend upon an established methodology for identifying l...

πŸš€ Excited to share our latest work, "Isolating Culture Neurons in Multilingual Large Language Models".

πŸ’» Data & code: github.com/namazifard/C...
πŸ“„ Preprint: arxiv.org/abs/2508.02241

05.09.2025 13:20 πŸ‘ 5 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

RAG is a powerful way to improve LLMs' answering abilities across many languages. But how do LLMs deal with multilingual contexts? Do they answer consistently when the retrieved info is provided to them in different languages?

Joint work w/ @jiruiqi.bsky.social & Raquel_FernΓ‘ndez
See thread! ‡️

18.04.2025 10:01 πŸ‘ 6 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Preview
Isotropy Matters: Soft-ZCA Whitening of Embeddings for Semantic Code Search Low isotropy in an embedding space impairs performance on tasks involving semantic inference. Our study investigates the impact of isotropy on semantic code search performance and explores post-proces...

πŸ—žοΈ A simple trick improves embedding retrieval performance even without further training.

ZCA whitening increases isotropy of the embedding space and thereby helps retrieval

Paper by Andor Diera and with @lukasgalke.bsky.social at ESANN 2025.

Preprint: arxiv.org/abs/2411.17538

20.02.2025 21:42 πŸ‘ 1 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image Post image

Thrilled to share our #ICLR2025 work on Meta-Causal States! 🌟 Causal graphs evolve with dynamic systems & agent actions. We show how to cluster causal models by qualitative behavior, revealing hidden dynamics & emergent relationships πŸš€ #Causality #ML

https://arxiv.org/abs/2410.13054

24.01.2025 19:34 πŸ‘ 12 πŸ” 6 πŸ’¬ 0 πŸ“Œ 0
Preview
Deep neural networks and humans both benefit from compositional language structure - Nature Communications This study demonstrates that deep neural networks, like humans, show a learnability advantage when trained on languages with more structured linguistic input, resulting in closer alignment with human ...

Deep neural networks and humans both benefit from compositional language structure. New paper by @lukasgalke.bsky.social, Yoav Ram, and @limorraviv.bsky.social. doi.org/10.1038/s414....

03.01.2025 10:19 πŸ‘ 8 πŸ” 5 πŸ’¬ 1 πŸ“Œ 0

What can we conclude? Humans and deep nets are not so different after all when learning a new language. The simplicity bias of overparameterized models seems to guide them towards learning compositional structures, even though they could easily memorize all different combinations.

30.12.2024 18:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

When analyzing the learning trajectory of RNNs throughout training, we make several other interesting observations: medium-structured languages have an learnability advantage early in training (likely due to same word being used for multiple meanings) but fall behind high-structured languages later.

30.12.2024 18:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We find a similar effect when looking at memorization errors. In the memorization test, the task for in-context LLMs boils down to copying a word that is present earlier in the prompt. But even here, we can see an advantage of language structure.

30.12.2024 18:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image Post image

All these learning systems, small RNNs, pre-trained LLMs, and humans, show *very* similar memorization and generalization behavior -- with more structured languages leading to generalizations that are more systematic generalization and more similar to the generalization of human participants.

30.12.2024 18:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Investigating the relationship between language learning and language structure, we find striking similarities between humans and language models: small recurrent neural networks trained from scratch and large pre-trained language models via in-context learning.

30.12.2024 18:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Deep neural networks and humans both benefit from compositional language structure Nature Communications - This study demonstrates that deep neural networks, like humans, show a learnability advantage when trained on languages with more structured linguistic input, resulting in...

πŸ”₯ Now finally out in Nature Communications:

Deep neural networks and humans both benefit from compositional structure

with Yoav Ram and @limorraviv.bsky.social

Paper link right away: rdcu.be/d5f2e

πŸ§΅β¬‡οΈ

30.12.2024 18:34 πŸ‘ 11 πŸ” 3 πŸ’¬ 1 πŸ“Œ 1
Several PhD scholarships in Artificial Intelligence Application deadline: 19 December 2024 at 23:59 hours local Danish time

Two more days left to apply for PhD positions on training multilingual language models at the Centre for Machine Learning in the Department of Mathematics and Computer Science (IMADA), University of Southern Denmark (SDU).

tinyurl.com/dfm2025phd

Application deadline: Dec 19, 2024

17.12.2024 10:13 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 1

tell me about LLMs tool use best practices. I know the high level, and want to learn about implementation/prompting details, e.g.:
- how do you best feed in the tool specs or DSL to the LLM?
- how do you ask it to indicate a tool use (which wrapper / indicator)
- how do you ask for nested calls
etc

24.11.2024 08:07 πŸ‘ 38 πŸ” 4 πŸ’¬ 9 πŸ“Œ 0

We have some openings for PhD/Postdoc positions on multilingual language modeling at SDU's Centre for Machine Learning, Denmark. Topics go down to the core of pre-training and instruction tuning and adjacent topics such as efficient language modeling. Please consider to apply :)

22.11.2024 12:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Several PhD scholarships in Artificial Intelligence Application deadline: 19 December 2024 at 23:59 hours local Danish time

Research positions on LLMs and available at the SDU Centre for ML:

tinyurl.com/dfm2025phd
tinyurl.com/dfm2025postdoc

22.11.2024 06:39 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 1

Thanks!

19.11.2024 22:55 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

I'm Lukas, working on machine learning and natural language processing. I'm particularly interested in interpretability of language models, efficient language models, continual learning, ood generalization, and machine communication.

I hope to find a community like the ex-twitter ML community here.

19.11.2024 22:04 πŸ‘ 6 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0