Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile
Gintare Karolina Dziugaite

@gkdziugaite

Sr Research Scientist at Google DeepMind, Toronto. Member, Mila. Adjunct, McGill CS. PhD Machine Learning & MASt Applied Math (Cambridge), BSc Math (Warwick).

ID: 954436574468624384

linkhttps://gkdz.org calendar_today19-01-2018 19:33:06

79 Tweet

3,3K Followers

110 Following

Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

Looking forward to participating and talking at this #ICML2023 workshop on PAC-Bayes and interactive learning. Working on related topics? Consider submitting! The deadline is in on May 31st!

Eleni Triantafillou (@eleni30fillou) 's Twitter Profile Photo

We’re thrilled to announce the first competition on unlearning, as part of the #neurips competition track! Joint work w Fabian Pedregosa, Vincent Dumoulin, Gintare Karolina Dziugaite, Ioannis Mitliagkas, @KurmanjiMeghdad, Peter Triantafillou, Isabelle Guyon and others. Read our blog post!

SOUVIK KUNDU (@thisissouvikk) 's Twitter Profile Photo

We are 1.5 months away from the deadline, consider submitting your recent findings in "sparse" learning, data, modality, signal processing. Introducing Conference on Parsimony and Learning (CPAL) 2023 at Hong Kong University youtu.be/pGbjiZOR63I via YouTube

Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

Sharing this fantastic opportunity for high school students interested in STEM! It’s a unique mentoring program enabling young people to engage in research and advanced topics in math. Applications are due in September πŸ—“οΈ

UniReps (@unireps) 's Twitter Profile Photo

We're excited to announce the first edition of πŸ”΅πŸ”΄ UniReps: the Workshop on Unifying Representations in Neural Models! 🧠 To be held at NeurIPS Conference 2023! SUBMISSION DEADLINE: 4 October Check out our Call for Papers, lineup of speakers and schedule at: unireps.org

We're excited to announce the first edition of πŸ”΅πŸ”΄ UniReps: the Workshop on Unifying Representations in Neural Models! 🧠
To be held at <a href="/NeurIPSConf/">NeurIPS Conference</a> 2023!

SUBMISSION DEADLINE: 4 October

Check out our Call for Papers, lineup of speakers and schedule at: unireps.org
Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

Deep learning may be hard, but deep un-learning is even harder. πŸ’ͺ How do we efficiently remove the influence of specific training examples while maintaining good performance on the remainder? Announcing NeurIPS Unlearning Competition πŸ“’ Submit your best ideas!πŸ†

Conference on Parsimony and Learning (CPAL) (@cpalconf) 's Twitter Profile Photo

CONGRATS to our CPAL Rising Star Awardees!! An incredibly talented and impressive bunch πŸ’― See who they are and what they do: cpal.cc/rising_stars_a…

Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

It was a great experience to give an in-person talk at #MLinPL to such an active audience with lots of thoughtful questions! Thanks again to the fantastic hosts who made me feel really welcomed, and made sure my visit was really well planned.

Yu Yang (@yuyang_ucla) 's Twitter Profile Photo

πŸŽ‰ Two of my papers have been accepted this week at #ICLR2024 & #AISTATS! Big thanks and congrats to co-authors Xuxi Chen & Eric Gan, mentors Atlas Wang & Gintare Karolina Dziugaite, and especially my advisor Baharan Mirzasoleiman! πŸ™ More details on both papers after the ICML deadline!

πŸŽ‰ Two of my papers have been accepted this week at #ICLR2024 &amp; #AISTATS! 
Big thanks and congrats to co-authors <a href="/xxchenxx_ut/">Xuxi Chen</a> &amp; Eric Gan, mentors Atlas Wang &amp; <a href="/gkdziugaite/">Gintare Karolina Dziugaite</a>, and especially my advisor <a href="/baharanm/">Baharan Mirzasoleiman</a>! πŸ™
More details on both papers after the ICML deadline!
Jacob Austin (@jacobaustin132) 's Twitter Profile Photo

We've finally put out a detailed IEEE/ACM paper on Google's multi-year effort to ease the burden of code review with ML. Google engineers now resolve 7.5% of all code review comments with an ML-suggested edit. But the path to that number has been a fun ML and UX journey!

We've finally put out a detailed IEEE/ACM paper on <a href="/Google/">Google</a>'s multi-year effort to ease the burden of code review with ML. Google engineers now resolve 7.5% of all code review comments with an ML-suggested edit. But the path to that number has been a fun ML and UX journey!
Pablo Samuel Castro (@pcastr) 's Twitter Profile Photo

πŸ“’Mixtures of Experts unlock parameter scaling for deep RL! Adding MoEs, and in particular Soft MoEs, to value-based deep RL agents results in more parameter-scalable models. Performance keeps increasing as we increase number of experts (green line below)! 1/9

πŸ“’Mixtures of Experts unlock parameter scaling for deep RL!

Adding MoEs, and in particular Soft MoEs, to value-based deep RL agents results in more parameter-scalable models.

Performance keeps increasing as we increase number of experts (green line below)!
1/9
Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

In deep nets, we observe good generalization together with memorization. In this new work, we show that, in stochastic convex optimization, memorization of most of the training data is a necessary feature of optimal learning.

EEML (@eemlcommunity) 's Twitter Profile Photo

DEADLINE March 29: prepare and submit your application for EEML 2024, Novi Sad, Serbia eeml.eu πŸ‡·πŸ‡Έ. Topics: Basics of ML, Multimodal learning, NLP, Advanced DL architectures, Generative models, AI for Science. Check our stellar speakers! Scholarships available! πŸŽ‰

DEADLINE March 29: prepare and submit your application for EEML 2024, Novi Sad, Serbia eeml.eu πŸ‡·πŸ‡Έ. Topics: Basics of ML, Multimodal learning, NLP, Advanced DL architectures, Generative models, AI for Science. Check our stellar speakers! Scholarships available! πŸŽ‰
Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

Google presents Mixture-of-Depths Dynamically allocating compute in transformer-based language models Same performance w/ a fraction of the FLOPs per forward pass arxiv.org/abs/2404.02258

Google presents Mixture-of-Depths

Dynamically allocating compute in transformer-based language models

Same performance w/ a fraction of the FLOPs per forward pass

arxiv.org/abs/2404.02258
Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

How are LLM capabilities affected by pruning? Checkout our ICLR 2025 paper showing that ICL is preserved until high levels of sparsity, in contrast to fact recall which quickly deteriorates. Our analysis reveals which part of the network is more prunable for a given capability.

Tian Jin (@tjingrant) 's Twitter Profile Photo

See u tmrw ICLR 2025 Sess 1 #133! When we down-scale LLMs (e.g.pruning), what happens to their capabilities? We studied complementary skills of memory recall & in-context learning and consistently found that memory recall deteriorates much quicker than ICL when down-scaling.

See u tmrw <a href="/iclr_conf/">ICLR 2025</a> Sess 1 #133!
When we down-scale LLMs (e.g.pruning), what happens to their capabilities? We studied complementary skills of memory recall &amp; in-context learning and consistently found that  memory recall deteriorates much quicker than ICL when down-scaling.
Gintare Karolina Dziugaite (@gkdziugaite) 's Twitter Profile Photo

We've seen memorization in NNets, despite good generalization. But can we generalize without memorizing? Come hear about our best #ICML2024 paper award work on showing that in stochastic convex opt, optimal learners memorize! Talk today at 11.15am, poster at 11.30am.