NLP Newsletter

Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com

🥇Top ML Papers of the Week

The top ML Papers of the Week (Feb 13 - Feb 19)

elvis
Feb 20
2
Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com

In this issue, we cover the top ML Papers of the Week (Feb 13 - Feb 19).


1). Lion (EvoLved Sign Momentum) - a simple and effective optimization algorithm that’s more memory-efficient than Adam. (paper)

Twitter avatar for @DrJimFan
Jim Fan @DrJimFan
The Adam optimizer is at the heart of modern AI. Researchers have been trying to dethrone Adam for years. How about we ask a machine to do a better job? @GoogleAI uses evolution to discover a simpler & efficient algorithm with remarkable features. It’s just 8 lines of code: 🧵
Image
6:11 PM ∙ Feb 15, 2023
3,732Likes606Retweets

2). Transformer models: an introduction and catalog. (paper)

Twitter avatar for @xamat
Xavier(Xavi) Amatriain @xamat
My Transformers Catalog has become one of my most popular posts ever. Some of you told me that you turned into a pdf for easier reading. I thought I should make it into an arXiv preprint. Here you go: 60 Transformers in 36 pages 🤖 🎉 arxiv.org/abs/2302.07730
Image
4:52 AM ∙ Feb 16, 2023
837Likes184Retweets

3). pix2pix3D - a 3D-aware conditional generative model extended with neural radiance fields for controllable photorealistic image synthesis. (paper)

Twitter avatar for @_akhaliq
AK @_akhaliq
3D-aware Conditional Image Synthesis abs: arxiv.org/abs/2302.08509 project page: cs.cmu.edu/~pix2pix3D/
2:37 AM ∙ Feb 17, 2023
3,116Likes702Retweets

4). Moral Self-Correction in Large Language Models - finds strong evidence that language models trained with RLHF have the capacity for moral self-correction. The capability emerges at 22B model parameters and typically improves with scale. (paper)

Twitter avatar for @AnthropicAI
Anthropic @AnthropicAI
Language models (LMs) exhibit harmful biases that can get worse with size. Reinforcement learning from human feedback (RLHF) helps, but not always enough. We show that simple prompting approaches can help LMs trained with RLHF produce less harmful outputs. arxiv.org/abs/2302.07459
Image
4:43 PM ∙ Feb 16, 2023
574Likes97Retweets

5). Vision meets RL - uses reinforcement learning to align computer vision models with task rewards; observes large performance boost across multiple CV tasks such as object detection and colorization. (paper)

Twitter avatar for @__kolesnikov__
Alexander Kolesnikov @__kolesnikov__
Vision meets RL! We reveal that policy gradient can be used for tuning vision models to optimize complex metrics, such as mAP, PQ or “color diversity”, observing large performance boosts on tasks like object detection, panoptic segmentation, etc. arxiv.org/abs/2302.08242
Image
11:36 AM ∙ Feb 17, 2023
566Likes111Retweets

6). Language Quantized AutoEncoders (LQAE) - an unsupervised method for text-image alignment that leverages pretrained language models; it enables few-shot image classification with LLMs. (paper)

Twitter avatar for @haoliuhl
Hao Liu @haoliuhl
We introduce an unsupervised method to align text and image. Language Quantized AutoEncoders (LQAE) enables few-shot image classification with GPT3 and linear classification of images based on RoBERTa text features. paper: arxiv.org/abs/2302.00902 code: github.com/lhao499/lqae
Image
11:20 PM ∙ Feb 13, 2023
368Likes76Retweets

7). Augmented Language Models - a survey of language models that are augmented with reasoning skills and the capability to use tools. (paper)

Twitter avatar for @omarsar0
elvis @omarsar0
Really nice survey on augmenting language models with reasoning skills and the ability to use tools. Includes prompting techniques, use cases, and applications. A must-read! arxiv.org/abs/2302.07842
Image
2:16 AM ∙ Feb 16, 2023
237Likes51Retweets

8). Geometric Clifford Algebra Networks (GCANs) - an approach to incorporate geometry-guided transformations into neural networks using geometric algebra. (paper)

Twitter avatar for @djjruhe
David Ruhe @djjruhe
New work on Geometric Clifford Algebra Networks (GCANs). We propose geometric templates for modeling dynamical systems. A 🧵on geometric / Clifford algebras, and symmetry group transformations in neural networks. 📜arxiv.org/abs/2302.06594
10:48 AM ∙ Feb 14, 2023
381Likes91Retweets

9) Auditing large language models - proposes a policy framework for auditing LLMs. (paper)

Twitter avatar for @hannahrosekirk
Hannah Rose Kirk @hannahrosekirk
✨New preprint (w/ Jakob Mökander, @jonasschuett and @Floridi) ✨ In this paper, we propose a policy framework for auditing LLMs by breaking down responsibilities at the governance-, model- and application-level. arxiv.org/abs/2302.08500 🧵
Image
4:45 PM ∙ Feb 17, 2023
142Likes23Retweets

10). Energy Transformer - a transformer architecture that replaces the sequence of feedforward transformer blocks with a single large Associate Memory model; this follows the popularity that Hopfield Networks have gained in the field of ML. (paper)

Twitter avatar for @arankomatsuzaki
Aran Komatsuzaki @arankomatsuzaki
Energy Transformer Replaces the sequence of feedforward transformer blocks with a single large Associative Memory model. arxiv.org/abs/2302.07253
Image
3:58 PM ∙ Feb 15, 2023
224Likes34Retweets

See you next week for another round of awesome ML papers!

Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com
Comments
TopNewCommunity

No posts

Ready for more?

© 2023 elvis
Privacy ∙ Terms ∙ Collection notice
Start WritingGet the app
Substack is the home for great writing