jzb

57 posts

jzb

jzb

@rucjzb

Katılım Mayıs 2017
373 Takip Edilen36 Takipçiler
jzb retweetledi
Graham Neubig
Graham Neubig@gneubig·
How can we vibe code while still maintaining code quality? Over the past year, I've shifted 95% of my development from manually writing code to using coding agents. I wrote this blog on some tricks I learned to work successfully with agents: all-hands.dev/blog/vibe-codi…
Graham Neubig tweet media
English
4
36
180
67.9K
jzb retweetledi
Graham Neubig
Graham Neubig@gneubig·
How far are we from having competent AI co-workers that can perform tasks as varied as software development, project management, administration, and data science? In our new paper, we introduce TheAgentCompany, a benchmark for AI agents on consequential real-world tasks.
Graham Neubig tweet media
English
19
144
829
126.6K
jzb retweetledi
Mustafa Suleyman
Mustafa Suleyman@mustafasuleyman·
Today we’re launching our new Copilot experience. I truly believe we can deliver a calmer, more helpful and supportive era of technology, with a Copilot that is now more intuitive, more personalized, and secure. Learn more, download, and enjoy. At Microsoft AI, we are creating an AI companion for everyone.  This is the first step.
English
81
160
906
309.9K
jzb retweetledi
Zhiqing Sun
Zhiqing Sun@EdwardSun0909·
🌟Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision 🌟 arxiv.org/abs/2403.09472 How can we keep improving AI systems when their capabilities surpass those of human supervisors? (1/n)
Zhiqing Sun tweet media
English
7
58
256
106.7K
jzb retweetledi
Srini Iyer
Srini Iyer@sriniiyer88·
New paper! How to train LLMs to effectively answer questions on new documents? Introducing *pre-instruction-tuning* - instruction-tuning *before* continued pre-training — significantly more effective than traditional instruction-tuning after PT. arxiv.org/abs/2402.12847
Srini Iyer tweet media
English
1
33
145
17.3K
jzb
jzb@rucjzb·
@saizhang0 My question is "Is it normal to have clusters with >100 8*A100 nodes in academia"? 😀
English
1
0
4
5.5K
Sai Zhang
Sai Zhang@saizhang0·
Someone is running jobs on 128 GPU nodes (~1000 A100) during holidays. Is this a normal PhD life?
Sai Zhang tweet media
English
85
56
1K
624.7K
jzb retweetledi
Zora Wang
Zora Wang@ZhiruoW·
Everyone is using RAG, but most of the retrieved context is noisy! 🚨 Introducing FilCo: “Learning to Filter Context for Retrieval-Augmented Generation” TL;DR: Get rid of the irrelevant content using FilCo, and you'll get better outputs. Preprint: arxiv.org/abs/2311.08377
English
6
56
284
51.1K
jzb retweetledi
Zhiqing Sun
Zhiqing Sun@EdwardSun0909·
🚀 Can RLAIF fully replace RLHF to align language models from scratch, enhancing both their alignment and capabilities? SALMON introduces a principle-following reward model in the realm of self-alignment, using just 6 ICL exemplars and 31 principles to outperform LLaMA-2-Chat!
Zhiqing Sun tweet media
English
5
90
296
100K
jzb retweetledi
Chunting Zhou
Chunting Zhou@violet_zct·
How do you turn a language model into a chatbot without any user interactions? We introduce LIMA: a LLaMa-based model fine-tuned on only 1,000 curated prompts and responses, which produces shockingly good responses. * No user data * No mode distillation * No RLHF
Chunting Zhou tweet media
English
27
222
1K
179.8K
jzb retweetledi
Luyu Gao
Luyu Gao@luyu_gao·
[1/4] Large language models (LLMs) tend to hallucinate, especially when generating long outputs. We present active retrieval augmented generation, in which an LLM actively decides when and what to retrieve throughout the generation process.
GIF
English
8
75
484
131K
jzb retweetledi
John Nay
John Nay@johnjnay·
Active LLM Retrieval Augmented Generation -Iteratively uses a prediction of upcoming sentence to anticipate future content which is used as query to retrieve relevant docs to regenerate sentence -On 4 long-form generation tasks: superior / competitive arxiv.org/abs/2305.06983
John Nay tweet media
English
6
72
393
133K
jzb retweetledi
Qian Liu
Qian Liu@sivil_taram·
🔥If data is the oil, have we exhausted the mine? Introducing our latest work, 🎉. "From Zero to Hero: Examining the Power of Symbolic Tasks in Instruction Tuning"💪 💡TL;DR: Generate tons of train examples via symbolic tasks to boost data quantity for instruction tuning🚀 1/3
Qian Liu tweet media
English
2
21
90
14.4K
jzb retweetledi
Jinlan Fu
Jinlan Fu@JinlanFu·
Can the text evaluator be customized for different/new evaluation aspects without training? Our GPTScore achieves customized, multifaceted, and training-free using emergent abilities of PLM, i.g., instruction and in-context learning. Paper: arxiv.org/pdf/2302.04166…
GIF
English
2
10
47
11.8K
jzb retweetledi
Graham Neubig
Graham Neubig@gneubig·
Retrieving information for QA is critical for reliability, but training separate retriever/reader models is cumbersome. At #EMNLP2022 we present Retrieval as Attention (ReAtt), a single retrieve/read model that is effective, generalizable, and adaptable arxiv.org/abs/2212.02027 🧵
Graham Neubig tweet media
English
1
26
127
0
jzb retweetledi
Uri Alon
Uri Alon@urialon1·
📢 New Paper: Program-aided Language models Prompting methods such as chain-of-thought (@_jasonwei) employ LLM for decomposing the problem into steps *and* solving each step. Instead, PaL decomposes the problem into *programmatic* steps and solves using a Python interpreter. 1/4
Uri Alon tweet media
English
13
99
497
0
jzb retweetledi
AI at Meta
AI at Meta@AIatMeta·
(1/4) Introducing EditEval, an instruction-based evaluation suite leveraging high-quality existing & new datasets for automatic evaluation of editing capabilities. At present, comprehensive eval of editing capabilities (i.e., fixing wrong info or reorganizing text) is lacking.
AI at Meta tweet media
English
6
88
524
0
jzb retweetledi
Chunting Zhou
Chunting Zhou@violet_zct·
I'm excited to share our work on a new sequence modeling architecture called Mega: Moving Average Equipped Gated Attention. Mega achieves SOTA results on multiple benchmarks, including NMT, Long Range Arena, language modeling, ImageNet and raw speech classification.
Chunting Zhou tweet media
English
8
56
364
0
jzb retweetledi
Graham Neubig
Graham Neubig@gneubig·
MEGA is a new method for modeling long sequences based on the surprisingly simple technique of taking the moving average of embeddings. Excellent results, outperforming strong competitors such as S4 on most tasks! Strongly recommend that you check it out: arxiv.org/abs/2209.10655
Graham Neubig tweet mediaGraham Neubig tweet media
Chunting Zhou@violet_zct

I'm excited to share our work on a new sequence modeling architecture called Mega: Moving Average Equipped Gated Attention. Mega achieves SOTA results on multiple benchmarks, including NMT, Long Range Arena, language modeling, ImageNet and raw speech classification.

English
2
29
230
0
jzb retweetledi
Google DeepMind
Google DeepMind@GoogleDeepMind·
Internship applications are now open! This year we have opportunities across various teams and offices 🌎🌍 Apply today via dpmd.ai/internships and learn more about the experience below ⬇️ #DeepMindInterns
Google DeepMind@GoogleDeepMind

Former intern turned intern mentor, @reverettai, describes his journey to DeepMind, sharing tips and advice for aspiring DeepMinders. Learn more about #internships and #LifeAtDeepMind on our blog: dpmd.ai/RichardQA-TW

English
24
119
408
0
jzb retweetledi
Graham Neubig
Graham Neubig@gneubig·
Happy to announce that I've formed a company, Inspired Cognition (inspiredco.ai) together with @stefan_fee and @odashi_en! Our goal is to make it easier and more efficient to build AI systems (particularly NLP) through our tools and expertise. 1/2
Graham Neubig tweet media
English
12
47
438
0