Julia Kreutzer
NLP & ML research @cohereforai.bsky.social 🇨🇦
- At #Neurips2025 this week with @cohereforai.bsky.social 🤩 This is what brought me into research: the ✨Network Effect ✨ Let's build the next breakthrough together! #LabLegends
- Reposted by Julia KreutzerWe’re thrilled to announce that some of our research will be presented at @emnlpmeeting.bsky.social next week! 🥳 If you’re attending the conference, don’t miss the chance to explore our work and connect with our team.
- Reposted by Julia KreutzerHow well do LLMs handle multilinguality? 🌍🤖 🔬We brought the rigor from Machine Translation evaluation to multilingual LLM benchmarking and organized the WMT25 Multilingual Instruction Shared Task spanning 30 languages and 5 subtasks.
- Reposted by Julia Kreutzer🌍Most multilingual instruction data starts as English and translation can’t capture cultural nuance or linguistic richness What if we optimized prompts instead of completions? That’s the focus of our most recent work on prompt space optimization for multilingual synthetic data🗣️
- Reposted by Julia KreutzerThe next generation of open LLMs should be inclusive, compliant, and multilingual by design. That’s why we @icepfl.bsky.social @ethz.ch @cscsch.bsky.social ) built Apertus.
- EPFL, ETH Zurich & CSCS just released Apertus, Switzerland’s first fully open-source large language model. Trained on 15T tokens in 1,000+ languages, it’s built for transparency, responsibility & the public good. Read more: actu.epfl.ch/news/apertus...
- Let's do the venue justice. Very excited for today's multilingual workshops at #COLM2025 💙
- Today at COLM, Cohere Labs Sr Research Scientist, @juliakreutzer.bsky.social will be presenting at 2 workshops. First, the Multilingual Data Quality Signals workshop, bringing together researchers across disciplines to discuss & present research on data quality signals in multilingual data.
- Looking forward to tomorrow's #COLM2025 workshop on multilingual data quality! 🤩
- In collaboration with @commoncrawl.bsky.social, MLCommons, and @eleutherai.bsky.social, the first edition of WMDQS at @colmweb.org starts tomorrow in Room 520A! We have an updated schedule on our website, including a list of all accepted papers.
- Ready for our poster today at #COLM2025! 💭This paper has had an interesting journey, come find out and discuss with us! @swetaagrawal.bsky.social @kocmitom.bsky.social Side note: being a parent in research does have its perks, poster transportation solved ✅
- Today at COLM, we are excited to share our work Déjà Vu: Multilingual LLM Evaluation through the Lens of Machine Translation Evaluation, during Poster Session 4, 4:30 - 6:30pm. Come connect with paper authors @juliakreutzer.bsky.social and @kocmitom.bsky.social.
- Reposted by Julia KreutzerWe’re not your average lab. We’re a hybrid research environment dedicated to revolutionizing the ML space. And we’re hiring a Senior Research Scientist to co-create with us. If you believe in research as a shared, global effort — this is your chance.
- 💡A collaborative➕diverse team is key. In real life as in the LLM world 💪🦾 Check out our latest work that builds on this insight. 👇
- Reposted by Julia Kreutzer[Not loaded yet]
- Reposted by Julia Kreutzer[Not loaded yet]
- Reposted by Julia Kreutzer[Not loaded yet]
- 🍋 Squeezing the most of few samples - check out our LLMonade recipe for few-sample test-time scaling in multitask environments. Turns out that standard methods miss out on gains on non-English languages. We propose more robust alternatives. Very proud of this work that our scholar Ammar led! 🚀
- 🚨LLM safety research needs to be at least as multilingual as our models. What's the current stage and how to progress from here? This work led by @yongzx.bsky.social has answers! 👇
- 🚧No LLM safety without multilingual safety - what is missing to closing the language gap? And where does this gap actually originate from? Answers 👇
- Multilingual 🤝reasoning 🤝 test-time scaling 🔥🔥🔥 New preprint! @yongzx.bsky.social has all the details 👇
- Reposted by Julia Kreutzer[Not loaded yet]
- 🤓MT eyes on multilingual LLM benchmarks 👉 Here's a bunch of simple techniques that we could adopt easily, and in total get a much richer understanding of where we are with multilingual LLMs. 🍬Bonus question: how can we spur research on evaluation of evaluations?
- Reposted by Julia Kreutzer[Not loaded yet]
- 📖New preprint with Eleftheria Briakou @swetaagrawal.bsky.social @mziizm.bsky.social @kocmitom.bsky.social! arxiv.org/abs/2504.11829 🌍It reflects experiences from my personal research journey: coming from MT into multilingual LLM research I missed reliable evaluations and evaluation research…
- Reposted by Julia Kreutzer[Not loaded yet]
- Reposted by Julia Kreutzer[Not loaded yet]
- Reposted by Julia Kreutzer[Not loaded yet]
- Reposted by Julia Kreutzer[Not loaded yet]
- 💬The first Q&A starts in a few hours. 🔔Also, a reminder to create your Open review profile if you haven't already. Non-institutional accounts require a verification process that can take time. One week till the abstract deadline!
- Reposted by Julia Kreutzer[Not loaded yet]
- ✨ Multilingual language modeling meets WMT✨ very exciting opportunity to get WMT-style evaluations for MLLMs: unseen tests, human evaluation, meta-evaluation, and that for multiple languages and tasks. Almost too good to be true! 🤩