Ishaan Gulrajani

1K posts

Ishaan Gulrajani banner
Ishaan Gulrajani

Ishaan Gulrajani

@__ishaan

Hi! I’m a machine learning researcher @openai. Previously @stanford @facebook @google @mila_quebec

San Francisco Katılım Kasım 2010
512 Takip Edilen4.2K Takipçiler
Aidan Clark
Aidan Clark@_aidan_clark_·
Absolutely crazy to me there is a whole generation of people who know what a transformer is but might not know what an RNN is (not meant to be a comment on Helen). Is this what getting old is like?
Helen Toner@hlntnr

Never forget @karpathy training a recurrent neural net (precursor to transformers) to imitate @paulg in 2015—a thing of syntactic and semantic beauty:

English
14
7
185
24.1K
Ishaan Gulrajani
Ishaan Gulrajani@__ishaan·
@azmythalauris and one day you’ll say your last goodbye without knowing it at the time, and one day you’ll long for them, and one day …
English
1
0
6
241
🌸🕷𝔞𝔷𝔪𝔶𝔱𝔥🕷🌸
I often get giddy thinking about the fact that there are people out there I don’t know, who I have yet to meet and will become close to and love. I don’t know who they are but eventually our paths will cross and I’m SO EXCITED ABOUT IT.
English
5
4
92
4.5K
Shanthi
Shanthi@evodevotee·
Why am I suppressing the big gigantic crush I have on MY OWN LIFE?
English
3
5
41
2K
Sebastien Bubeck
Sebastien Bubeck@SebastienBubeck·
Just started at @OpenAI and I couldn't be more excited to join at this pivotal moment of safe AGI development! Met so many old friends already, talent density of this place is just insane!! Thank you all for the warm welcome, and in particular @sama. Now let the unicorns fly!
Sebastien Bubeck tweet media
English
70
28
905
104K
Ari Weinstein
Ari Weinstein@AriX·
OK Apple Mail I know I’m not on top of my inbox but you don’t have to call me out like this
Ari Weinstein tweet media
English
8
3
110
11.8K
Durk Kingma
Durk Kingma@dpkingma·
@sirbayes @ylecun @srush_nlp This is a non-issue but I've never understood why scientists prefer to minimize/descent. Maximization/ascent feels so much better. Much better vibes IMO 😎
English
4
0
24
4.8K
Sasha Rush
Sasha Rush@srush_nlp·
Updating all my NeurIPS papers.
Sasha Rush tweet media
English
37
163
2.4K
250.4K
Mimee // smart casual dark and academic
I realize that one of the hindrances to my making fast progress on the PhD is that I don’t have family here, and I would really appreciate it if my friends could sign up to attend my defense if I ever have one.
English
4
0
11
928
Dave Font
Dave Font@davefontenot·
Just got GPT5 access. Holy fucking shit. Reply with your GitHub if you want to come try it out.
English
46
4
136
66.9K
Anna Huang
Anna Huang@huangcza·
I’m excited I’ll be joining MIT next fall, for a shared interdisciplinary faculty position between Music (@MIT_MTA @MIT_SHASS) & EECS. I’m recruiting PhD students @MITEECS for Fall 2024 (apply by Dec 15), and also Postdocs. See czhuang.github.io for details. Come join us!
English
47
77
758
114.1K
Chenlin Meng
Chenlin Meng@chenlin_meng·
Super excited for the launch of Pika 1.0! I am extremely grateful to be working with such an amazing and talented team on this journey! ❤️ I am also very thankful for the support from our incredible investors, advisors, friends, and community members. We couldn't have achieved this without you! :)
Pika@pika_labs

Introducing Pika 1.0, the idea-to-video platform that brings your creativity to life. Create and edit your videos with AI. Rolling out to new users on web and discord, starting today. Sign up at pika.art

English
52
32
605
168.5K
Ishaan Gulrajani
Ishaan Gulrajani@__ishaan·
OpenAI is nothing without its people.
English
5
16
295
23.5K
Ishaan Gulrajani
Ishaan Gulrajani@__ishaan·
@jwthickstun @LucaAmb @sedielem @vlastelicap By a change-of-basis, Gaussian diffusion on learned embeddings is exactly equivalent to Gaussian diffusion on one-hots with a learned noise covariance. I really wanted one-hots to win, but reality had other plans.
English
0
0
3
99
John Thickstun
John Thickstun@jwthickstun·
@LucaAmb @sedielem @vlastelicap Ishaan has run experiments on one-hots (e.g., in the Diffusion-LM paper last year: Appendix F). I'm not sure whether a direct ablation of one-hots vs. embeddings has made it into any published papers. Maybe @__ishaan can comment.
English
1
0
2
103
Sander Dieleman
Sander Dieleman@sedielem·
Making diffusion language models work as well as autoregressive ones will be a challenge (see my earlier blog post: sander.ai/2023/01/09/dif…). This paper quantifies this and finds a 64x efficiency disadvantage across all scales 👀 a big gap, but at least it's a constant factor!
Ishaan Gulrajani@__ishaan

New paper with @tatsu_hashimoto! Likelihood-Based Diffusion Language Models: arxiv.org/abs/2305.18619 Likelihood-based training is a key ingredient of current LLMs. Despite this, diffusion LMs haven't shown any nontrivial likelihoods on standard LM benchmarks. We fix this!🧵

English
4
15
84
31.1K
Ishaan Gulrajani
Ishaan Gulrajani@__ishaan·
@sedielem Fun fact that didn’t make it into the paper: when we started the project, the gap was 10,000x 😳. It took stacking many 10% improvements to make it this far.
English
2
1
17
435