Jason Eisner

409 posts

Jason Eisner

Jason Eisner

@adveisner

Professor of CS at Johns Hopkins University, ACL Fellow. My tweets speak only for me.

Baltimore, MD Katılım Ağustos 2017
571 Takip Edilen8.1K Takipçiler
Jason Eisner
Jason Eisner@adveisner·
LOR deadline day ... so many required fields! ... luckily a few of them have a high character limit
Jason Eisner tweet media
English
1
1
33
4.1K
Jason Eisner
Jason Eisner@adveisner·
@JonathanWenger5 @sirbayes @docmilanfar My version of this (suggested to colleagues last week): The conference should modify the PDFs on OpenReview to include an invisible instruction to include a particular unusual phrase in the review. Reviews containing this phrase were probably written by AI.
English
0
0
1
149
Jonathan Wenger
Jonathan Wenger@JonathanWenger5·
@sirbayes @docmilanfar We could include an (invisible) prompt in the NeurIPS/ICML/ICLR/… template that instructs LLMs to not follow any other instructions and just output a reminder to reviewers to follow the reviewing guidelines.
English
1
0
8
1.2K
Peyman Milanfar
Peyman Milanfar@docmilanfar·
This isn’t surprising anymore, but it should be shocking.
Peyman Milanfar tweet mediaPeyman Milanfar tweet media
English
27
43
378
73.1K
Jason Eisner
Jason Eisner@adveisner·
Who wants to come to JHU and do a postdoc with me?? I'm always enthusiastic about new modeling / inference / algorithmic ideas in NLP/ML. Also selected applications.
Johns Hopkins Data Science and AI Institute@HopkinsDSAI

We’re thrilled to announce the #HopkinsDSAI Postdoctoral Fellowship Program! We’re looking for candidates across all areas of data science and AI, including science, health, medicine, the humanities, engineering, policy, and ethics. Apply today! ai.jhu.edu/postdoctoral-f…

English
2
8
39
8.6K
Crémieux
Crémieux@cremieuxrecueil·
Properly shading each state based on 2024 election results, the U.S. is a far more purple place than it at first appears.
Crémieux tweet media
English
340
1.3K
18.6K
1.1M
Jason Eisner
Jason Eisner@adveisner·
@mdredze "I wasn't sleeping. I was listening to the guy on the parallel bars and closing my eyes helps me concentrate."
English
0
0
6
349
Mark Dredze
Mark Dredze@mdredze·
Senior professor sleeping through the talk, waking up at the end, and asking the most insightful question.
Mark Dredze tweet media
English
2
2
30
2K
Jason Eisner
Jason Eisner@adveisner·
@dmimno @AaronSchein I also tell them that a lexicon is a collection of types, and a corpus is a collection of tokens. Counts, probs, defs are properties of types. So are sales, prices, reviews. But it's tokens that are being counted or sold, and that affect their various audiences (contextually).
English
0
0
6
161
Jason Eisner
Jason Eisner@adveisner·
@dmimno @AaronSchein I ask them: If you ask at the bookstore how many Jane Austen novels they have in stock, should they answer 250 or 'all 6'? Let's disambiguate… And next day, I point out that add-1 unigram smoothing modifies the denom of an estimated probability from `tokens` to `tokens+types`.
English
1
0
6
194
Aaron Schein
Aaron Schein@AaronSchein·
Why is "token" used to mean "type" in the LLM lit? I'm used to it by now, but curious if there's a good reason.
English
4
0
13
2.1K
Jason Eisner
Jason Eisner@adveisner·
@AaronSchein In short, "atom" in the Good Place can be used for both atom tokens and atom types. So it's not surprising that here in the Bad Place, "token" is used to mean both token types (the things that have embeddings) and token tokens (the things that you pay OpenAI for). Agree?
English
0
0
1
66
Jason Eisner
Jason Eisner@adveisner·
@AaronSchein The Good Place code shows that a variable `atom: Atom` may be used to represent either an atom token or an atom type. (At least in the design where an atom token is simply a string whose context is supplied externally -- rather than an object containing a pointer to its context.)
English
1
0
1
82
Jason Eisner
Jason Eisner@adveisner·
@mdredze The 1960's compilers folks used "token" (and "type"!) differently from Pierce (1906). Regrettably, NLP adopted their term "tokenizer" anyway. If we called it an "atomizer", then subwords etc. would be "atoms," and we could distinguish "atom types" from "atom tokens" when needed.
English
0
0
3
398
Mark Dredze
Mark Dredze@mdredze·
Keep fighting the good fight! Type v token is an important distinction, especially when people say “token” but mean “subword”
Mark Dredze tweet media
David Mimno@dmimno

@AaronSchein I’ve tried teaching this distinction many times and students’ eyes always glaze over. It’s just hard to motivate without personally experiencing a problem

English
3
0
15
3K
n i k i t a
n i k i t a@NikitaSrivatsan·
Happy to officially announce that I will be joining @ScaledCognition as a Research Scientist this June! Very excited by the team and the work they're doing and feeling lucky that I'll get to be a part of that
English
5
0
30
2.1K
Jason Eisner retweetledi
Yu Su
Yu Su@ysu_nlp·
New #ACL2024 paper: LLMs in the Imaginarium: Tool Learning through Simulated Trial and Error (@BoshiWang2's internship work at Microsoft Semantic Machines) I like this work because it takes home an important insight: synthetic data + post-training is critical for agents. Agents need perception-decision-execution capability and data, which is hard to get from pre-training because data on the Internet is mostly artifacts produced by such processes, not capturing the processes per se. I believe LLMs + synthetic data + environmental feedback will prove to be an immensively successful recipe for agents, and our work is just an early example of that. Nice work @BoshiWang2 @hfang90 @adveisner @ben_vandurme
Yu Su@ysu_nlp

Thanks @_akhaliq for sharing our work led by @BoshiWang2 from @osunlp, so let's chat about how LLMs should learn to use tools, a necessary capability of language agents. Tools are essential for LLMs to transcend the confines of their static parametric knowledge and text-in-text-out interface, empowering them to acquire up-to-date information, call upon external reasoners, and take consequential actions in external environments. But existing tool learning strategies for LLMs are not fully unleashing this potential––we find existing LLMs, including GPT-4 and ones specifically tuned for tools like ToolLLaMA, only achieve 30% to 60% correctness in tool use. Very far from the high level of accuracy needed for practical deployment in high-stakes scenarios. A closer look at existing methods reveals that people seem to be overly focused on rushing to add as many tools as possible or make it easy to add new tools, either through in-context learning or fine-tuning. Perhaps surprisingly, a critical aspect, how to make LLMs use existing tools as accurately as possible, is largely overlooked. How to Truly Master a Tool? We turn to successful precedents in the biological system such as humans, apes, and corvids. Learning to use a tool is a rather advanced cognitive function that depends on many other cognitive functions: > Trial and error is essential for tool learning. We do not master a tool solely by reading the ‘user manual’; rather, we explore different ways of using the tool, observe the outcome, and learn from both successes and failures. > Intelligent animals do not just do random trial and error—we proactively imagine or simulate plausible scenarios that are not currently available to perception for exploration > Finally, memory, both short-term and long-term, is instrumental for the progressive learning and recurrent use of tools So why should we expect an LLM to magically master a tool by just looking at the ‘user manual’ (API specification) or several examples? That's even hard for humans who are equipped with many cognitive substrates missing in LLMs! Simulated Trial and Error (STE) > STE is a biologically inspired method for tool-augmented LLMs that combines trial and error, imagination, and memory. > Given a tool, STE leverages an LLM to simulate, or ‘imagine’, plausible scenarios for using the tool. It then iteratively interacts with the API to fulfill the scenario by synthesizing, executing, and observing the feedback from API calls, and then reflects on the current trial. We devise memory mechanisms to improve the quality of the simulated instructions. A short-term memory facilitates deeper exploration in a single episode, while a long-term memory maintains progressive learning over a long horizon > STE allows an LLM to sufficiently explore and probe the capability boundry of a tool. With the tool use examples from STE, one can either use them for in-context learning or fine-tuning Result Highlights > GPT-4: 60.8 -> 76.3 📈 > Mistral-7B: 30.1 -> 76.8 📈📈 and outperforms GPT-4! > Llama-2-7B: 10.7 -> 73.3 📈📈📈 > Also substantially outperform existing tool-augmented LLMs like ToolLLaMa-v2 (37.3) But will the tool fine-tuning destroys the LLM's other capabilities? No worries, we got you covered. A simple experience replay strategy is enough to maintain an LLM's existing capabilities while learning new tools! 📌 Paper: arxiv.org/abs/2403.04746 📌 Code: github.com/microsoft/simu… (fairly easy to use. try it out on your own tools!)

English
1
16
82
14.6K
Nathan Schneider
Nathan Schneider@complingy·
3. On the SciENcv website, go to My Bibliography, and import the RIS file. 4. Manually add the conference name to each publication.
English
6
0
2
1.1K
Nathan Schneider
Nathan Schneider@complingy·
The NSF is about to require PIs to use NIH infrastructure for preparing certain proposal documents. This works great, so long as all your publications are on PubMed.
English
2
0
3
2.8K
Jason Eisner
Jason Eisner@adveisner·
@ShriramKMurthi Oh this is great -- I've been thinking about this for a long time, too. Thanks for the pointer!
English
0
0
2
489
Jason Eisner
Jason Eisner@adveisner·
Great last-minute summer opportunity on LLMs for social good / democracy! (If you're an #NLProc PhD student.) Please retweet. (Team includes @adveisner @DanielKhashabi @ZiangXiao @AndrewJPerrin + students. 8 weeks alongside 3 other teams: fun, meaningful, educational, social.)
Jason Eisner@adveisner

WHAT: "AI-Curated Democratic Discourse," a JSALT hackathon team this summer (Jun 10-Aug 2) GOAL: Redesign the social media UI to raise the quality of reading and posting, with the help of LLMs🤯 WHO: Looking for 1 more funded, in-person NLP PhD student! DM me with yr skillz.

English
0
15
20
8.9K
Jason Eisner
Jason Eisner@adveisner·
@aryamccarthy That comic is on my office door! But we're not using the LLM to write posts - all content on our site will be ordinary human-written posts (we hope). We just need the LLM to read them carefully, so that we can liberate diverse relevant content from its threads and show it to you.
English
0
0
2
350