Maxime De Bruyn

204 posts

Maxime De Bruyn banner
Maxime De Bruyn

Maxime De Bruyn

@_maxime_db

Data Scientist @ BNP

Brussels, Belgium Entrou em Haziran 2014
648 Seguindo86 Seguidores
Lila Rest
Lila Rest@LilaRest·
Introducing 𝐆𝐞𝐦𝐦𝐚 𝟒 𝟑𝟏𝐁 𝐓𝐮𝐫𝐛𝐨 ⚡️ It runs on a 𝘴𝘪𝘯𝘨𝘭𝘦 RTX 5090, at 51 tok/s (single) and 1244 tok/s (batched). And prefills up to 15359 tok/s. It's 𝟔𝟖% 𝐬𝐦𝐚𝐥𝐥𝐞𝐫 in GPU memory and ~𝟐.𝟓𝐱 𝐟𝐚𝐬𝐭𝐞𝐫 than the base model, and retains nearly 𝐢𝐝𝐞𝐧𝐭𝐢𝐜𝐚𝐥 𝐪𝐮𝐚𝐥𝐢𝐭𝐲 on benchmarks (1-3% loss). Turbo is a derivative of the NVFP4 quant that NVIDIA released a few days ago. It fully leverages NVIDIA Blackwell FP4 tensor cores for ~𝟐× 𝐡𝐢𝐠𝐡𝐞𝐫 𝐜𝐨𝐧𝐜𝐮𝐫𝐫𝐞𝐧𝐭 𝐭𝐡𝐫𝐨𝐮𝐠𝐡𝐩𝐮𝐭 𝐭𝐡𝐚𝐧 𝐨𝐭𝐡𝐞𝐫 𝐪𝐮𝐚𝐧𝐭𝐬. I'm using it for hard classification tasks — on internal benchmarks it showed 𝐒𝐨𝐧𝐧𝐞𝐭-𝟒.𝟓-𝐥𝐞𝐯𝐞𝐥 𝐢𝐧𝐭𝐞𝐥𝐥𝐢𝐠𝐞𝐧𝐜𝐞 (scored well above Haiku 4.5), at a 600𝘵𝘩 of the cost. A single RTX 5090 scales up to 18 req/s at 1000in/20out 🥵. Model card and benchmark in comments 👇 I'd love to hear your use cases.
Lila Rest tweet media
English
12
10
97
12.5K
Alisa Liu
Alisa Liu@alisawuffles·
We created SuperBPE🚀, a *superword* tokenizer that includes tokens spanning multiple words. When pretraining at 8B scale, SuperBPE models consistently outperform the BPE baseline on 30 downstream tasks (+8% MMLU), while also being 27% more efficient at inference time.🧵
Alisa Liu tweet media
English
93
323
2.8K
368.8K
Maxime De Bruyn
Maxime De Bruyn@_maxime_db·
@SNCB @SrFanon Bonjour. Qu’en sera-t-il pour demain? Faut-il prévoir autant de désagréments sur cette ligne?
Français
1
0
1
64
SNCB
SNCB@SNCB·
@SrFanon Bonjour Mariela, il y a malheureusement toujours un problème de disponibilité du matériel pour ce train, le service en charge fait le maximum afin de résoudre cela au plus vite, désolé pour les désagréments occasionnés. ^Seb
Français
1
0
0
105
🎙Jean-Louis Queguiner
🎙Jean-Louis Queguiner@JiliJeanlouis·
Our new Real-Time STT engine is out! 🔥It offers the best of both worlds: batch-level quality with real-time transcription speed. With < 300 ms latency, support in 100+ languages, code-switching and RT add-ons, @gladia_io is here to set a new standard for real-time AI — with $16M Series A in new funding!
English
13
25
236
32.9K
Lewis Tunstall
Lewis Tunstall@_lewtun·
Here's a simple recipe to train a 7B model that outperforms Llama2 70B on MT Bench 🥇 1. SFT Mistral 7B on the UltraChat dataset 2. Align the SFT model to the UltraFeedback dataset with "direct preference optimisation" (DPO) Demo: huggingfaceh4-zephyr-chat.hf.space More details in the 🧵
Lewis Tunstall tweet media
English
18
200
948
463.1K
Yuxiang Wu
Yuxiang Wu@yuxiangwu_·
If you are at #EMNLP2022, I am looking forward to meeting you in Poster Session 9 on Dec 10, 11:00-12:00 (Abu Dhabi time). I am staying in Abu Dhabi from Dec 6-12, and would love to meet and chat!
English
1
0
4
0
Yuxiang Wu
Yuxiang Wu@yuxiangwu_·
#EMNLP2022 If you’re interested in incorporating large-scale knowledge into LMs, don't forget to check-out our work “An Efficient Memory-Augmented Transformer for Knowledge-Intensive NLP Tasks”. Work w/ @zhaoyuhitsz, Baotian, @PMinervini, Pontus, @riedelcastro 1/N
GIF
English
2
9
81
0
Maxime De Bruyn retweetou
CLiPS
CLiPS@clipsua·
The 16th Belgium NLP Meetup (bit.ly/3CgGPdw) will be organized in Leuven on November 17 2022. Among others, Maxime De Bruyn (@_maxime_db) from CLiPS will talk about whether language models know what they don't know.
CLiPS tweet media
English
0
2
4
0
Raj Dabre
Raj Dabre@prajdabre·
If you had submitted a paper to AACL (@aaclmeeting) and if you see a final camera submission link in your softconf then it's likely that the paper has been accepted. However, this is subject to change since the notification ddl is a few days away. #NLProc
English
2
1
2
0
Maxime De Bruyn retweetou
clem 🤗
clem 🤗@ClementDelangue·
To me, huge ML models are to machine learning what formula 1 is to the car industry!
English
24
66
857
0
Maxime De Bruyn
Maxime De Bruyn@_maxime_db·
@ikazban C’est votre résistance au changement qui est hallucinante.
Français
1
2
5
0
Jamal Ikazban
Jamal Ikazban@ikazban·
Hallucinant d’observer dans notre démocratie qu’une multinationale condamnée par la justice pour contournement de la loi et montage frauduleux arrive encore à faire porter son projet de modification de la loi par des parlementaires belges !
Français
7
10
26
0
Zach Waterfield
Zach Waterfield@zlwaterfield·
Any alternatives to Grammarly that are open source or do all the computation locally so ALL my data isn’t all being sent to their servers?
English
22
14
232
0
Maxime De Bruyn
Maxime De Bruyn@_maxime_db·
@stefan_it_ Hi Stefan. Just wanted to mention that we did something similar with MQA here: huggingface.co/datasets/clips… The FAQ part is still messy (lots of close duplicates about travel), but the CQA part is not affected by these problems.
English
0
0
0
0
.
.@jmnollet·
D’aucuns s'interrogent sur l’attitude des ministres @ecolo. Nos actes seront bien entendu posés en pleine cohérence avec nos propos et nous l’avons fait savoir dès hier au Premier ministre. Nous continuons à travailler à des solutions plus qu’urgentes au vu de la situation.
. tweet media
Français
58
22
69
0
Andrew Ng
Andrew Ng@AndrewYNg·
Tell me you work in AI without telling me you work in AI.
English
1.1K
257
3.4K
0