Akash Srivastava

643 posts

Akash Srivastava banner
Akash Srivastava

Akash Srivastava

@variational_i

Director, Core AI, IBM. Chief Architect https://t.co/lo3JOcuDFA . Founder, Red Hat AI Innovation Team. PI @MITIBMLab. ❤️ Density Ratios.

Cambridge, MA انضم Temmuz 2009
1.2K يتبع1.3K المتابعون
Akash Srivastava أُعيد تغريده
Isha Puri
Isha Puri@ishapuri101·
come check out poster #5518 at NeurIPS morning session today to learn about how you can encourage diversity / prevent early-pruning during inference-time scaling and boost the performance of any model without additional training!
Isha Puri@ishapuri101

[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint @MIT_CSAIL / @RedHat AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io

English
1
1
14
1.5K
Anna Goldie
Anna Goldie@annadgoldie·
Excited to announce that @Azaliamirh and I are launching @RicursiveAI, a frontier AI lab creating a recursive self-improving loop between AI and the hardware that fuels it. Today, chip design takes 2-3 years and requires thousands of human experts. We will reduce that to weeks. This will be incredibly hard. For context: Go has a search space of 10^360. A simplified version of chip placement—only one part of the design process—has a search space of 10^9000. But we are the right team to solve it. We co-founded the Machine Learning for Systems team at Google Brain. There, we built AlphaChip—an RL agent for chip placement. AlphaChip has been used to design four generations of TPUs, data center CPUs, autonomous vehicle chips, and mobile phone chips. These chips are running in data centers and devices all over the world. Our immediate goal is to dramatically accelerate chip design. Next, we plan to design chips end-to-end given an ML workload, unlocking a Cambrian explosion of custom silicon. Finally, we will close the recursive loop. We will build our own chips, train our own models, and co-evolve them on the path to superintelligence. AI designs better chips 🔄chips train better AI We sat down with @WSJ’s @Berber_Jin1 to discuss Ricursive: wsj.com/tech/this-ai-s…
Ricursive Intelligence@RicursiveAI

Introducing Ricursive Intelligence, a frontier AI lab enabling a recursive self-improvement loop between AI and the chips that fuel it. Learn more at ricursive.com

English
50
42
417
58.9K
Akash Srivastava
Akash Srivastava@variational_i·
🚀 How is generative AI transforming the way we design cars, planes, and entire systems? In Ep 2 of No Math AI, @ishapuri101 and I chat with Dr. @_faezahmed (@MIT DeCoDE Lab) about how AI boosts creativity, cuts design time, and works with engineers—not against them.
Red Hat AI@RedHat_AI

How is generative AI reshaping engineering design? In Episode 2 of No Math AI, hosts Dr. Akash Srivastava (@variational_i) and MIT PhD student Isha Puri (@ishapuri101) sit down with Dr. Faez Ahmed (@_faezahmed) from MIT DeCoDE Lab to explore just that. 👇

English
0
1
3
1K
Akash Srivastava أُعيد تغريده
Red Hat AI
Red Hat AI@RedHat_AI·
Excited to share our preliminary work on customizing reasoning models using Red Hat AI Innovation’s Synthetic Data Generation (SDG) package! 📄 Turn your documents into training data for LLMs. 🧵👇
English
2
5
11
1.3K
Akash Srivastava أُعيد تغريده
Isha Puri
Isha Puri@ishapuri101·
had a great time giving a talk about probabilistic inference scaling and the power of small models at the IBM Research ML Seminar Series - the best talks end with tons of questions, and it was great to see everyone so engaged : ) youtube.com/watch?v=--3rsQ…
YouTube video
YouTube
English
2
21
141
14.9K
Akash Srivastava
Akash Srivastava@variational_i·
Excited to share our latest work with @ishapuri101 et al.! 🚀 We introduce a probabilistic inference approach for inference-time scaling of LLMs using particle-based Monte Carlo methods—achieving 4–16x better scaling on math reasoning tasks and O1-level performance on MATH500.
Isha Puri@ishapuri101

[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint @MIT_CSAIL / @RedHat AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io

English
0
0
5
426
Akash Srivastava أُعيد تغريده
Seungwook Han
Seungwook Han@seungwookh·
🧩 Why do task vectors exist in pretrained LLMs? Our new research uncovers how transformers form internal abstractions and the mechanisms behind in-context learning(ICL).
Seungwook Han tweet media
English
6
30
190
21.8K
Akash Srivastava أُعيد تغريده
Cole Hurwitz
Cole Hurwitz@cole_hurwitz·
Neural activity is correlated among animals performing the same task and across sequential trials. Led by @zhang_yizi and @hl3616, we develop an reduced-rank model that exploits shared structure across animals to improve neural decoding. biorxiv.org/content/10.110…
Cole Hurwitz tweet media
English
1
35
189
14.8K
Akash Srivastava أُعيد تغريده
Cole Hurwitz
Cole Hurwitz@cole_hurwitz·
What will a foundation model for the brain look like? We argue that it must be able to solve a diverse set of tasks across multiple brain regions and animals. Check out our preprint where we introduce a multi-region, multi-animal, multi-task model (MtM): arxiv.org/abs/2407.14668
English
5
62
256
36.4K
Akash Srivastava أُعيد تغريده
Seungwook Han
Seungwook Han@seungwookh·
🚀 Stronger, simpler, and better! 🚀 Introducing Value Augmented Sampling (VAS) - our new algorithm for LLM alignment and personalization that outperforms existing methods!
Seungwook Han tweet media
English
4
33
128
25.5K
Akash Srivastava أُعيد تغريده
Seungwook Han
Seungwook Han@seungwookh·
Excited to give a talk on our hottest, newest work “Value Augmented Sampling for Language Model Alignment and Personalization” at 2:30p Halle A3 in #ICLR2024 Reliable and Responsible Foundation Models Workshop 🥳🥳
Huaxiu Yao@HuaxiuYaoML

📢Workshop on Reliable and Responsible Foundation Models will happen today (8:50am - 5:00pm). Join us at #ICLR2024 room Halle A 3 for a wonderful lineup of speakers, along with 63 amazing posters and 4 contributed talks! Schedule: #program" target="_blank" rel="nofollow noopener">iclr-r2fm.github.io/#program.

English
1
2
12
1.5K
Akash Srivastava
Akash Srivastava@variational_i·
Attending #ICLR2024, interested in continual learning and like probabilistic modeling? Lazar from the @MITIBMLab, will be presenting our latest work that takes a probabilistic approach to modular continual learning on Tuesday, 7 May, Halle B #222 (iclr.cc/virtual/2024/p…).
Lazar Valkov@lazarvalkov

I’ll be presenting our #ICLR2024 paper on a probabilistic approach to scaling modular continual learning algorithms while achieving different types of knowledge transfer. (arxiv.org/abs/2306.06545, in collaboration with @variational_i @swarat @RandomlyWalking ). A tldr (1/8):

English
0
1
11
1.1K
Akash Srivastava أُعيد تغريده
Faez Ahmed
Faez Ahmed@_faezahmed·
Check out our work titled "From Automation to Augmentation: Redefining Engineering Design and Manufacturing in the Age of NextGen-AI", where we highlight the requirements for NextGenAI suitable for design, engineering, and manufacturing. mit-genai.pubpub.org/pub/9s6690gd/r…
MIT Stone Center on Inequality & Shaping Work@MITshapingwork

Instead of continuing to emphasize automation, a human-centric approach to the next generation of #AI technologies in #manufacturing could enhance workers' skills and boost productivity. mit-genai.pubpub.org/pub/9s6690gd/r… @AustinLentsch @DAcemogluMIT @baselinescene @_faezahmed @MITMechE

English
1
3
13
3.1K
Akash Srivastava أُعيد تغريده
Mathieu
Mathieu@miniapeur·
Mathieu tweet media
ZXX
27
230
2.7K
235.6K
Jeethu Rao
Jeethu Rao@jeethu·
@Geronimo_AI They seem to be comparing it against the older Mistral-7B-Instruct-v0.1 and not the newer Mistral-7B-Instruct-v0.2 model.
English
1
0
2
233