labml.ai

654 posts

labml.ai banner
labml.ai

labml.ai

@labmlai

📝 Annotated paper implementations https://t.co/qeO4UTbrJ3

Katılım Aralık 2020
9 Takip Edilen12.6K Takipçiler
labml.ai retweetledi
Himanshu
Himanshu@nothiingf4·
As an ML engineer, implementation >>>>everything. Knowing is theory. Implementation is understanding. Few outstanding topics it has: 1. Reinforcement Learning - ppo, dqn 2. Transformer - classical to Retro, switch, gpt models 3. Diffusion models - stable, DDPM, DDIM, UNET 4. GANs - cycle, wasserstein, stylegan & few more 5. Graph neural networks - GAT, GATv2 Skip the tutorial hell & learn about various models Learn implementations in this GitHub repo. I’ll share more resources later. Link in comments 👇
Himanshu tweet media
English
17
54
596
36.2K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
We've open-sourced our internal AI coding IDE. We built this IDE to help with coding and to experiment with custom AI workflows. It's based on a flexible extension system, making it easy to develop, test, and tweak new ideas quickly. Each extension is a Python script that runs locally. (Links in replies) 🧶👇
NOTBAD AI tweet mediaNOTBAD AI tweet media
English
2
5
11
3.1K
labml.ai retweetledi
vpj
vpj@vpj·
GEPA appears to be an effective method for enhancing LLM performance, requiring significantly fewer rollouts than reinforcement learning (RL). It maintains a pool of system prompts. It uses an the LLM to improve them by reflecting on the generated answers and the scores/feedback for a minibatch of problems. GEPA keeps the Pareto frontier of system prompts. That is, if system prompt A performs worse on every validation problem compared to system prompt B, A is filtered out from the pool. 👇
vpj tweet media
English
2
4
17
2.7K
labml.ai retweetledi
vpj
vpj@vpj·
Wrote an annotated Triton implementation of Flash Attention 2. (Links in reply) This is based on the flash attention implementation by the Triton team. Changed it to support GQA and cleaned up a little bit. Check it out to read the code for forward and backward passes along with the math and derivations. Hope this helps understand transformer attention and flash attention better. There's about 60 more annotated deep learning paper implementations on this website.
vpj tweet media
English
3
9
48
4.3K
labml.ai retweetledi
vpj
vpj@vpj·
The following scripts stalls and times out on B200 x 8. Seems like we are having problems with NCCL. Anyone else experiencing this? @PyTorch
vpj tweet media
English
1
1
4
986
labml.ai retweetledi
vpj
vpj@vpj·
The new training also improved GPQA from 64.2% to 67.3% and MMLU Pro from 64.2% to 67.3%. This model was also trained with the same reasoning datasets we used to train the v1.0 model. We mixed more general instruction data with answers sampled from the Mistral-Small-24B-Instruct-2501 model during the SFT to reduce the degradation of IFEval, which seems to have resulted in generalization of reasoning to non math and coding problems. The datasets and the models are available on @huggingface. Follow @notbadai for updates.
NOTBAD AI@notbadai

We are releasing an updated reasoning model with improvements on IFEval scores (77.9%) than our previous model (only 51.4%). 👇 Links to try the model and to download weights below

English
1
6
7
3.1K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
We are releasing an updated reasoning model with improvements on IFEval scores (77.9%) than our previous model (only 51.4%). 👇 Links to try the model and to download weights below
NOTBAD AI tweet mediaNOTBAD AI tweet media
English
1
6
10
5.5K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
We just released a Python coding reasoning dataset with 200k samples on @huggingface This was generated by our RL-based self-improved Mistral 24B 2501 model. This dataset was used to train train Notbad v1.0 Mistral 24B. 🤗 Links in replies 👇
NOTBAD AI tweet media
English
2
7
19
4.6K
labml.ai retweetledi
vpj
vpj@vpj·
Uploaded the dataset of 270k math reasoning samples that we used to finetune Notbad v1.0 Mistral 24B (MATH-500=77.52% GSM8k Platinum=97.55%) to @huggingface (link in reply) Follow @notbadai for updates
English
9
13
62
14.5K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
We're open-sourcing a math reasoning dataset with 270k samples, generated by our RL-based self-improved Mistral 24B 2501 model and used to train Notbad v1.0 Mistral 24B. Available on Hugging Face: huggingface.co/datasets/notba…
English
0
4
8
1.6K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
📢 We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is built upon the @MistralAI Small 24B 2501 and has been further trained with reinforcement learning on math and coding.
NOTBAD AI tweet mediaNOTBAD AI tweet mediaNOTBAD AI tweet media
English
1
8
22
9.1K
labml.ai
labml.ai@labmlai·
We added token visualization to Inspectus. It lets you visualize metrics associated with tokens such as loss, entropy, KL div, etc. It works on notebooks and pretty easy to use. 👇
English
1
4
27
5.2K
labml.ai
labml.ai@labmlai·
Our open source deep learning experiment monitoring library now has 2000 stars! Thank you
labml.ai tweet media
English
1
3
18
2.9K
labml.ai retweetledi
NOTBAD AI
NOTBAD AI@notbadai·
We’ve been training @nvidia Mistral-NeMo-Minitron-8B-Base for math reasoning on the GSM8K-Aug dataset, and we have a version with a 70.2% gsm8k score, up from a 58.5% cot score (reported in the paper LLM Pruning and distillation). 👇
English
1
9
18
5.6K