Karolina Staล„czak's Avatar

Karolina Staล„czak

@karstanczak

#NLP Postdoc at Mila - Quebec AI Institute and McGill University | Former PhD @ University of Copenhagen (CopeNLU) ๐ŸŒ karstanczak.github.io

949
Followers
398
Following
13
Posts
18.11.2024
Joined
Posts Following

Latest posts by Karolina Staล„czak @karstanczak

Preview
Feedback form for the Workshop on Gender Bias in NLP (GeBNLP) @ ACL Thank you for participating in the 6th Workshop on Gender Bias in NLP! Your feedback is incredibly valuable and will help us improve future events. This anonymous form should take less than 5 minutes ...

Thank you to everyone who joined the GeBNLP Workshop at #ACL2025! โœจ

To help us improve future events, please share your feedback in our anonymous 5-minute survey. Your perspective is needed!

๐Ÿ“ Survey Link: forms.gle/mXtZ4mWUGSBN...

#GeBNLP #NLP

11.08.2025 11:07 ๐Ÿ‘ 1 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

The Gender Bias in NLP workshop #ACL2025 is officially over, and we're buzzing with energy! โœจ

A huge thank you to all our participants for a day packed with incredible discussions. You are the community we need ๐Ÿ™Œ

See you next time!

#GeBNLP #NLP

04.08.2025 11:36 ๐Ÿ‘ 2 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

It's time! โฐ Organizers are ready to kick off the Gender Bias in NLP Workshop #ACL2025. Join us for a day of talks, posters, and discussions!

๐Ÿ—“๏ธ Today, August 1
๐Ÿ•˜ 9 AM - 6:15 PM
๐Ÿ“ Hall C

#GeBNLP #NLP

01.08.2025 06:54 ๐Ÿ‘ 4 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

We are thrilled to announce our keynote speakers for the 6th Workshop on Gender Bias in NLP @ #ACL2025!

Please join us in welcoming:
๐Ÿ”นAnne Lauscher @a-lauscher.bsky.social
๐Ÿ”นMaarten Sap @maartensap.bsky.social

Full details: gebnlp-workshop.github.io/keynotes.html
See you on August 1! โ˜€๏ธ
#NLP #GeBNLP

16.07.2025 13:36 ๐Ÿ‘ 15 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Our VLMs4All workshop is taking place today!
๐Ÿ“… on Thursday, June 12
โฒ๏ธ from 9AM CDT
๐Ÿ›๏ธin Room 104E

Join us today at @cvprconference.bsky.social for amazing speakers, posters, and a panel discussion on making VLMs more geo-diverse and culturally aware!

#CVPR2025

12.06.2025 11:22 ๐Ÿ‘ 4 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 1
Post image

Do LLMs hallucinate randomly? Not quite.

Our #ACL2025 (Main) paper shows that hallucinations under irrelevant contexts follow a systematic failure mode โ€” revealing how LLMs generalize using abstract classes + context cues, albeit unreliably.

๐Ÿ“Ž Paper: arxiv.org/abs/2505.22630 1/n

06.06.2025 18:09 ๐Ÿ‘ 46 ๐Ÿ” 18 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 3
Post image

๐Ÿ—“๏ธ Save the date! It's official: The VLMs4All Workshop at #CVPR2025 will be held on June 12th!

Get ready for a full day of speakers, posters, and a panel discussion on making VLMs more geo-diverse and culturally aware ๐ŸŒ

Check out the schedule below!

06.06.2025 09:19 ๐Ÿ‘ 4 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 1

๐Ÿšจ Deadline Extension Alert for #VLMs4All Challenges! ๐Ÿšจ

We have extended the challenge submission deadline
๐Ÿ› ๏ธ New challenge deadline: Apr 22

Show your stuff in the CulturalVQA and GlobalRG challenges!
๐Ÿ‘‰ sites.google.com/view/vlms4al...

Spread the word and keep those submissions coming! ๐ŸŒโœจ

17.04.2025 13:58 ๐Ÿ‘ 2 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Exciting release! AgentRewardBench offers that much-needed closer look at evaluating agent capabilities: automatic vs. human eval. Important findings here, especially on the popular LLM judges. Amazing work by @xhluca.bsky.social & team!

15.04.2025 19:11 ๐Ÿ‘ 3 ๐Ÿ” 1 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

๐Ÿ”” Reminder & Call for #VLMs4All @ #CVPR2025!
Help shape the future of culturally aware & geo-diverse VLMs:
โš”๏ธ Challenges: Deadline: Apr 15 ๐Ÿ”—https://sites.google.com/view/vlms4all/challenges
๐Ÿ“„ Papers (4pg): Deadline: Apr 22 ๐Ÿ”—https://sites.google.com/view/vlms4all/call-for-papers
Join us!

09.04.2025 11:20 ๐Ÿ‘ 4 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Reviewers needed! ๐Ÿ“ข The 6th Workshop on Gender Bias in NLP at #ACL2025 (Vienna, Aug 1st) is looking for you! Sign up to review:

forms.gle/VkPU4vS4EacE... #NLP

07.04.2025 08:33 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Excited to be organizing the VLMs4All workshop at #CVPR2025! ๐ŸŽ‰
The workshop features fantastic speakers, a short-paper track, and two challenges, including one based on CulturalVQA. Donโ€™t miss it!

14.03.2025 17:21 ๐Ÿ‘ 10 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

6/ ๐Ÿค Thanks to our steering committees and co-organizers for their hard work in making the VLMs4All Workshop possible!

@meharbhatia.bsky.social @rabiul.bsky.social @spandanagella.bsky.social @sivareddyg.bsky.social @svansteenkiste.bsky.social @karstanczak.bsky.social

14.03.2025 15:55 ๐Ÿ‘ 5 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Exploiting Instruction-Following Retrievers for Malicious Information Retrieval Parishad BehnamGhader, Nicholas Meade, Siva Reddy

Instruction-following retrievers can efficiently and accurately search for harmful and sensitive information on the internet! ๐ŸŒ๐Ÿ’ฃ

Retrievers need to be aligned too! ๐Ÿšจ๐Ÿšจ๐Ÿšจ

Work done with the wonderful Nick and @sivareddyg.bsky.social

๐Ÿ”— mcgill-nlp.github.io/malicious-ir/
Thread: ๐Ÿงต๐Ÿ‘‡

12.03.2025 16:15 ๐Ÿ‘ 12 ๐Ÿ” 8 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

The potential for malicious misuse of LLM agents is a serious threat.

That's why we created SafeArena, a safety benchmark for web agents. See the thread and our paper for details: arxiv.org/abs/2503.04957 ๐Ÿ‘‡

10.03.2025 18:20 ๐Ÿ‘ 9 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Llamas browsing the web look cute, but they are capable of causing a lot of harm!

Check out our new Web Agents โˆฉ Safety benchmark: SafeArena!

Paper: arxiv.org/abs/2503.04957

10.03.2025 17:50 ๐Ÿ‘ 9 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

6/ @meharbhatia.bsky.social @hattie.bsky.social @nicolaspapernot.bsky.social @nicolaschapados.bsky.social @anamarasovic.bsky.social @ghadfield.bsky.social @sivareddyg.bsky.social
#AIAlignment #LLM #ResponsibleAI #NLP

04.03.2025 16:08 ๐Ÿ‘ 6 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Societal Alignment Frameworks Can Improve LLM Alignment Recent progress in large language models (LLMs) has focused on producing responses that meet human expectations and align with shared values - a process coined alignment. However, aligning LLMs remain...

5/ Read our full paper here: arxiv.org/abs/2503.00069
Letโ€™s discuss! How should AI align with society? ๐Ÿค๐Ÿ’ก

04.03.2025 16:08 ๐Ÿ‘ 7 ๐Ÿ” 1 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

4/ We also discuss the role of participatory alignment, where diverse stakeholders help shape AI behavior rather than deferring solely to designers.

04.03.2025 16:08 ๐Ÿ‘ 5 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

3/ Instead of perfecting rigid alignment objectives, we explore how LLMs can navigate uncertaintyโ€”a feature, not a flaw!

04.03.2025 16:08 ๐Ÿ‘ 5 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

2/ We propose leveraging societal alignment frameworks to guide LLM alignment:
๐Ÿ”น Social alignment: Modeling norms, values & cultural competence
๐Ÿ”น Economic alignment: Fair reward mechanisms & collective decision-making
๐Ÿ”น Contractual alignment: Legal principles for LLMs

04.03.2025 16:08 ๐Ÿ‘ 5 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

1/ LLM alignment remains a challenge because human values are complex, dynamic, and often conflict with narrow optimization goals.

Existing methods like RLHF struggle with misspecified objectives.

04.03.2025 16:08 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

๐Ÿ“ขNew Paper Alert!๐Ÿš€

Human alignment balances social expectations, economic incentives, and legal frameworks. What if LLM alignment worked the same way?๐Ÿค”

Our latest work explores how social, economic, and contractual alignment can address incomplete contracts in LLM alignment๐Ÿงต

04.03.2025 16:08 ๐Ÿ‘ 28 ๐Ÿ” 13 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 3
Post image

The 6th edition of the Gender Bias in Natural Language Processing (GeBNLP) workshop is happening at #ACL2025 in Vienna!

๐Ÿ“ข Check out our Call for Papers! Find all the details on our website: gebnlp-workshop.github.io

We look forward to your submissions!

28.02.2025 18:04 ๐Ÿ‘ 14 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 1
Call For Papers The 6th Workshop on Gender Bias in Natural Language Processing at ACL 2025.

Great news! We've extended the direct submission deadline to *April 18*, giving you more time to prepare your submissions ๐Ÿš€

Check the new deadlines on our webpage: gebnlp-workshop.github.io/cfp.html

28.02.2025 18:08 ๐Ÿ‘ 8 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

Presenting โœจ ๐‚๐‡๐€๐’๐„: ๐†๐ž๐ง๐ž๐ซ๐š๐ญ๐ข๐ง๐  ๐œ๐ก๐š๐ฅ๐ฅ๐ž๐ง๐ ๐ข๐ง๐  ๐ฌ๐ฒ๐ง๐ญ๐ก๐ž๐ญ๐ข๐œ ๐๐š๐ญ๐š ๐Ÿ๐จ๐ซ ๐ž๐ฏ๐š๐ฅ๐ฎ๐š๐ญ๐ข๐จ๐ง โœจ

Work w/ fantastic advisors Dima Bahdanau and @sivareddyg.bsky.social

Thread ๐Ÿงต:

21.02.2025 16:28 ๐Ÿ‘ 17 ๐Ÿ” 8 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 1
Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.'
SHADES is in multiple grey colors (shades).

Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.' SHADES is in multiple grey colors (shades).

โšซโšช It's coming...SHADES. โšชโšซ
The first ever resource of multilingual, multicultural, and multigeographical stereotypes, built to support nuanced LLM evaluation and bias mitigation. We have been working on this around the world for almost **4 years** and I am thrilled to share it with you all soon.

10.02.2025 08:28 ๐Ÿ‘ 128 ๐Ÿ” 23 ๐Ÿ’ฌ 6 ๐Ÿ“Œ 3
Preview
International Association for Safe & Ethical AI Conference โ€” IASEAI 2025 The International Association for Safe and Ethical AI will host its inaugural conference (IASEAI โ€˜25) on Feb 6-7, 2025 at the OECD La Muette Headquarters and Conference Centre in Paris, ahead of the P...

Excited to present today a poster at @OECD in Paris @IASEAIorg based on our upcoming paper "Societal Alignment Frameworks Can Improve LLM Alignment" (stay tuned for the pre-print soon!๐ŸŽŠ). Today (Fri) at 1pm CET. Conference livestream: iaseai.org/conference

07.02.2025 09:13 ๐Ÿ‘ 8 ๐Ÿ” 2 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

It turns out we had even more papers at EMNLP!

Let's complete the list with three more๐Ÿงต

24.11.2024 02:17 ๐Ÿ‘ 14 ๐Ÿ” 4 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 1

๐Ÿ™‹โ€โ™€๏ธ๐Ÿ‘‹

22.11.2024 14:25 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0