Robert L. Logan IV

102 posts

Robert L. Logan IV

Robert L. Logan IV

@rloganiv

UC Irvine Katılım Ocak 2017
513 Takip Edilen227 Takipçiler
Robert L. Logan IV retweetledi
Yoshitomo Matsubara
Yoshitomo Matsubara@yoshitomo_cs·
🚨Please repost🔃 Actively exploring new opportunities that align with my passions and skills, ideally a research scientist position with a spirit of open source/science DM me if you know such opportunities! About me yoshitomo-matsubara.net Long ver. linkedin.com/posts/yoshitom…
Yoshitomo Matsubara tweet mediaYoshitomo Matsubara tweet media
English
0
5
5
3.8K
Robert L. Logan IV retweetledi
Royi Rassin
Royi Rassin@RoyiRassin·
How diverse are the outputs of text-to-image models and how can we measure that? In our new work, we propose a measure based on LLMs and Visual-QA (VQA), and show NONE of the 12 models we experiment with are diverse. 🧵 1/11
Royi Rassin tweet media
English
5
29
98
18.5K
Robert L. Logan IV retweetledi
Cory Doctorow NO LONGER ON TWIT TER
Retiring the US debt would retire the US dollar: We don't tax billionaires to fight the national debt. We tax billionaires to fight BILLIONAIRES. twitter.com/doctorow/statu… 3/
Cory Doctorow NO LONGER ON TWIT TER tweet media
Cory Doctorow NO LONGER ON TWIT TER@doctorow

One of the most consequential series of investigative journalism of this decade was the @Propublica series that @eisingerj helmed, in which Eisinger and colleagues analyzed a trove of leaked IRS tax returns for the richest people in America: propublica.org/series/the-sec… 1/

English
1
2
2
539
Robert L. Logan IV retweetledi
Yu Fei
Yu Fei@Walter_Fei·
Alignment is necessary for LLMs, but do we need to train aligned versions for all model sizes in every model family? 🧐 We introduce 🚀Nudging, a training-free approach that aligns any base model by injecting a few nudging tokens at inference time. 🌐fywalter.github.io/nudging/ 📜arxiv.org/pdf/2410.09300 1/7
GIF
English
6
24
136
18.9K
Robert L. Logan IV retweetledi
dr. jack morris
dr. jack morris@jxmnop·
prompt optimization is such a hard problem, and the algorithms are dumb. the fact that they work at all is baffling, and people don't talk about it enough for those who don't know, the prompt optimization problem is argmax_{x} ℓ(y | x; θ) where x is some prompt and ℓ is the loss for output y under language model θ. solving this problem exactly requires enumerating all possible prompts x... ...and the space of possible 10-token prompts in a 50k vocab is 50,000^10. yet people regularly find solutions to these problems with 100-token prompts. how can this be possible? the most popular algorithm (AutoPrompt/GCG) is basically "while true: rank a bunch of token-swaps, and then take the top one" and yet this works somehow. in fact, this process is "good enough" that everyone uses it and has been cited thousands of times. truly boggles the mind
English
43
60
834
121.1K
Robert L. Logan IV retweetledi
Simran Arora
Simran Arora@simran_s_arora·
Excited to share Just read twice: going beyond causal language modeling to close quality gaps between efficient recurrent models and attention-based models!! There’s so much recent progress on recurrent architectures, which are dramatically more memory efficient and asymptotically faster than attention 💨 But there’s no free lunch 🥪 these models can’t fit all the information from long contexts into the limited memory, degrading in-context learning quality. Is all lost?
Simran Arora tweet media
English
7
58
299
93K
Robert L. Logan IV retweetledi
Sasha Rush
Sasha Rush@srush_nlp·
Video: An attempt to explain Diffusion LMs in 15 minutes. Several people have noted that they have bounced off this topic before because the math felt a bit perplexing. Think it's a good time to revisit. youtu.be/WjAUX23vgfg
YouTube video
YouTube
English
3
56
405
41.7K
Robert L. Logan IV retweetledi
Anthony Chen
Anthony Chen@_anthonychen·
Lots of discourse around long-context language models (LCLMs) subsuming RAG and retrieval but how close are we to this paradigm shift? Introducing LOFT a 1 million token benchmark spanning 6 tasks & 35 datasets to test LCLMs’ ability to do in-context retrieval & reasoning [1/10]
Jinhyuk Lee@leejnhk

Can long-context language models (LCLMs) subsume retrieval, RAG, SQL, and more? Introducing LOFT: a benchmark stress-testing LCLMs on million-token tasks like retrieval, RAG, and SQL. Surprisingly, LCLMs rival specialized models trained for these tasks! arxiv.org/abs/2406.13121

English
1
4
22
2.3K
Robert L. Logan IV
Robert L. Logan IV@rloganiv·
@zouharvi 90% of the text GPT5 generates will be written at 1 am on nights of major conference deadlines.
English
5
0
4
129
Robert L. Logan IV retweetledi
Kolby Nottingham
Kolby Nottingham@kolbytn·
LLMs can have a huge impact on gaming🎮, but not in the way you may think! Number 10 may surprise you!😱 I wrote up some of my thoughts and ideas about the use of LLMs in games. There's a lot more to it than just generative AI NPCs. 📖kolbynottingham.com/llms-and-games
English
0
3
11
881
Robert L. Logan IV retweetledi
Yoshitomo Matsubara
Yoshitomo Matsubara@yoshitomo_cs·
🔃Please RT🔃 My 1st TODO item this year is getting a new exciting research job! Please DM me if open source/science matters to your team e.g., open-sourcing research projects & publishing papers Long ver. linkedin.com/posts/yoshitom… Webpage: yoshitomo-matsubara.net Thanks🙏
English
0
12
21
8.2K
Robert L. Logan IV retweetledi
Shivanshu Gupta
Shivanshu Gupta@shivanshug11·
En route to Singapore 🇸🇬 for #EMNLP2023! I'll present (virtual poster Sun @ 9am) our work on training-free example selection for In-Context Learning. Come say hi if interested in chatting about ICL, reasoning, and generalization with LLMs or even exploring Singapore!
Shivanshu Gupta tweet media
Shivanshu Gupta@shivanshug11

(1/6) 🚀🚀 Thrilled that our paper arxiv.org/abs/2305.14907 has been accepted to #EMNLP2023 findings! 🎉 tl;dr: Selecting in-context examples that together cover all the salient aspects of the test input yields training-free methods that beat even trained SoTA methods! 💪🔥

English
0
1
12
1.3K
Robert L. Logan IV retweetledi
Shivanshu Gupta
Shivanshu Gupta@shivanshug11·
(1/6) 🚀🚀 Thrilled that our paper arxiv.org/abs/2305.14907 has been accepted to #EMNLP2023 findings! 🎉 tl;dr: Selecting in-context examples that together cover all the salient aspects of the test input yields training-free methods that beat even trained SoTA methods! 💪🔥
English
1
13
53
8K
Robert L. Logan IV retweetledi
ACL 2026
ACL 2026@aclmeeting·
Dear ACL community, ACL is considering multiple proposals to change its anonymity period policy. It seeks immediate feedback from the community about the proposed changes. Please add your voice until Friday, September 22nd (AOE): aclweb.org/portal/content… #NLProc
English
5
192
272
115.9K
Robert L. Logan IV retweetledi
Aaron Reichlin-Melnick
Aaron Reichlin-Melnick@ReichlinMelnick·
New! DHS has added 8 new "fields of study" to the STEM OPT eligible list: - Landscape architecture - Institutional research - Mechatronics - Composite materials tech - Linguistics and comp sci - Developmental and adolescent psych - Geospatial int. - Demography and pop studies.
Aaron Reichlin-Melnick tweet mediaAaron Reichlin-Melnick tweet mediaAaron Reichlin-Melnick tweet media
English
4
15
51
10.2K