NLP Newsletter

Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com

🥇Top ML Papers of the Week

The top ML Papers of the Week (Feb 20 - Feb 26)

elvis
Feb 26
6
Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com

This issue highlights the top ML Papers of the Week (Feb 20 - Feb 26).


1). LLaMA - a 65B parameter foundation model released by Meta AI; relies on publicly available data and outperforms GPT-3 on most benchmarks despite being 10x smaller. (paper)

Twitter avatar for @GuillaumeLample
Guillaume Lample @GuillaumeLample
Today we release LLaMA, 4 foundation models ranging from 7B to 65B parameters. LLaMA-13B outperforms OPT and GPT-3 175B on most benchmarks. LLaMA-65B is competitive with Chinchilla 70B and PaLM 540B. The weights for all models are open and available at research.facebook.com/publications/l… 1/n
Image
Image
Image
Image
4:08 PM ∙ Feb 24, 2023
4,119Likes877Retweets

2) Composer - a 5B parameter creative and controllable diffusion model trained on billions (text, image) pairs. (paper)

Twitter avatar for @_akhaliq
AK @_akhaliq
Composer is a large (5 billion parameters) controllable diffusion model trained on billions of (text, image) pairs github: github.com/damo-vilab/com… paper: arxiv.org/abs/2302.09778 project page: damo-vilab.github.io/composer-page/
Image
2:20 AM ∙ Feb 25, 2023
1,099Likes236Retweets

3) Hindsight Instruction Relabeling - an alternative algorithm to train LLMs from feedback; the feedback is converted to instruction by relabeling the original one and training the model, in a supervised way, for better alignment. (paper)

Twitter avatar for @tianjun_zhang
Tianjun Zhang @NeurIPS 2022 @tianjun_zhang
Can we use purely supervised learning for RLHF using large language models? We introduce HIR (Hindsight Instruction Relabeling), which achieves impressive results using FLAN-T5 on hard BigBench tasks!
Image
11:52 PM ∙ Feb 21, 2023
401Likes58Retweets

4). Active-Prompt - a prompting technique to adapt LLMs to different task-specific example prompts (annotated with human-designed chain-of-thought reasoning); this process involves finding where the LLM is most uncertain and annotating those. (paper)

Twitter avatar for @johnjnay
John Nay @johnjnay
Active Prompting for LLMs -Most Chain-of-Thought examples are pulled from a fixed set -Instead, to adapt to diff tasks 1) Find where LLM is most uncertain 2) Annotate those -State-of- the-art on complex reasoning tasks Paper arxiv.org/abs/2302.12246 Code github.com/shizhediao/act…
Image
12:20 PM ∙ Feb 24, 2023
291Likes51Retweets

5). Modular Deep Learning - a survey offering a unified view of the building blocks of modular neural networks; it also includes a discussion about modularity in the context of scaling LMs, causal inference, and other key topics in ML. (paper)

Twitter avatar for @seb_ruder
Sebastian Ruder @seb_ruder
In our new survey “Modular Deep Learning”, we provide a unified taxonomy of the building blocks of modular neural nets and connect disparate threads of research. 📄 arxiv.org/abs/2302.11529 📢 ruder.io/modular-deep-l… 🌐 modulardeeplearning.com w/ @PfeiffJo @licwu @PontiEdoardo
Image
11:40 AM ∙ Feb 23, 2023
332Likes72Retweets

6). Recitation-Augmented LMs - an approach that recites passages from the LLM’s own memory to produce final answers; shows high performance on knowledge-intensive tasks. (paper)

Twitter avatar for @EdwardSun0909
Zhiqing Sun @EdwardSun0909
How can LLMs such as GPT-3 and ChatGPT achieve greater factual accuracy without relying on an external retrieval search engine? Our #ICLR2023 paper shows that recitation can help - like humans! Recitation-Augmented Language Models arxiv.org/abs/2210.01296 1/N
Image
8:37 PM ∙ Feb 22, 2023
331Likes79Retweets

7). LLMs to Optimize Code - an approach that uses LLMs to suggest functionally correct, performance-improving code edits. (paper)

Twitter avatar for @mathemagic1an
Jay Hack @mathemagic1an
AI systems can optimize their own code (!) "Learning Performance-Improving Code Edits" arxiv.org/pdf/2302.07867… Introduces a dataset of (before, after) code optimizations + describes methods for building code optimizing LLMs My takeaways 👇
Image
11:40 PM ∙ Feb 19, 2023
437Likes58Retweets

8). Prompt Injection Threats - a comprehensive analysis of novel prompt injection threats to application-integrated LLMs. (paper)

Twitter avatar for @omarsar0
elvis @omarsar0
If you are building with LLMs, it's good to know about novel adversarial prompting techniques. This paper presents an analysis of the topic. It also discusses attack vectors when augmenting LLMs with retrieval and API calling abilities. arxiv.org/abs/2302.12173 https://t.co/wqcW02Ccoe
Image
2:26 AM ∙ Feb 24, 2023
207Likes58Retweets

9). Aligning Text-to-Image Models using Human Feedback - proposes a fine-tuning method to align generative models using human feedback. (paper)

Twitter avatar for @kimin_le2
Kimin @kimin_le2
📄Can "learning from human feedback" improve text-to-image models? I'm excited to share "Aligning Text-to-Image Models using Human Feedback" 📝 arxiv.org/abs/2302.12192 1/N
Image
4:37 PM ∙ Feb 24, 2023
200Likes57Retweets

10). MERF - a memory-efficient radiance field representation for real-time view synthesis of large scenes in a browser. (paper)

Twitter avatar for @_akhaliq
AK @_akhaliq
MERF: Memory-Efficient Radiance Fields for Real-time View Synthesis in Unbounded Scenes abs: arxiv.org/abs/2302.12249 project page: merf42.github.io
4:21 AM ∙ Feb 24, 2023
394Likes72Retweets

See you next week for another round of awesome ML papers!

Share this post

🥇Top ML Papers of the Week

nlpnews.substack.com
Comments
TopNewCommunity

No posts

Ready for more?

© 2023 elvis
Privacy ∙ Terms ∙ Collection notice
Start WritingGet the app
Substack is the home for great writing