Jade Copet

69 posts

Jade Copet banner
Jade Copet

Jade Copet

@jadecopet

AI Research @ FAIR, Meta AI

Katılım Eylül 2015
357 Takip Edilen677 Takipçiler
Sabitlenmiş Tweet
Jade Copet retweetledi
Summer Yue
Summer Yue@summeryue0·
We’re excited to share our preparedness report on Code World Model (CWM), FAIR’s latest open-weight model for code generation and reasoning. This report was developed by the SEAL team and the AI Security team, marking our first external publication since part of SEAL joined Meta just 1.5 months ago. It reflects our ongoing commitment to safety and alignment as we look ahead to future open-source and frontier models. We’re actively working to broaden our evaluation coverage and improve our processes, and we welcome collaboration and feedback from the broader research community. For more details, please read the full report: ai.meta.com/research/publi…
English
6
16
141
53K
Jade Copet retweetledi
Jonas Gehring
Jonas Gehring@jnsgehring·
LLMs for code should do much better if they can iterate on tests -- but they don't. Our new work (RLEF) addresses this with execution feedback at RL *training time* to use execution feedback at *inference time*. arxiv.org/abs/2410.02089 is just out! 1/6
Jonas Gehring tweet media
English
7
86
335
139.2K
Jade Copet retweetledi
Robin San Roman
Robin San Roman@RobinSanroman·
AudioSeal training code is now available inside the beautiful audiocraft repo 🚀 github.com/facebookresear… You can now train your own audio watermarking models and define your own tradeoff between fidelity, robustness and message capacity based on your needs.
English
1
9
53
5.4K
Jade Copet retweetledi
AI at Meta
AI at Meta@AIatMeta·
Introducing Meta Llama 3: the most capable openly available LLM to date. Today we’re releasing 8B & 70B models that deliver on new capabilities such as improved reasoning and set a new state-of-the-art for models of their sizes. Today's release includes the first two Llama 3 models — in the coming months we expect to introduce new capabilities, longer context windows, additional model sizes and enhanced performance + the Llama 3 research paper for the community to learn from our work. More details ➡️ go.fb.me/i2y41n Download Llama 3 ➡️ go.fb.me/ct2xko
English
340
1.4K
5.7K
1.1M
Jade Copet retweetledi
Robin San Roman
Robin San Roman@RobinSanroman·
Today we present AudioSeal, a proactive solution for the detection of voice cloning based on localised watermarking. It relies on 2 jointly trained models: an imperceptible watermark generator and a detector with sample level precision. 1/n
Robin San Roman tweet media
English
3
13
45
3.3K
Jade Copet retweetledi
Baptiste Rozière
Baptiste Rozière@b_roziere·
We released a 70B version of CodeLlama today! Trained on 1T tokens, it is a much stronger base model for coding tasks. I look forward to seeing what the community will do with it! :)
AI at Meta@AIatMeta

Today we’re releasing Code Llama 70B: a new, more performant version of our LLM for code generation — available under the same license as previous Code Llama models. Download the models ➡️ bit.ly/3Oil6bQ • CodeLlama-70B • CodeLlama-70B-Python • CodeLlama-70B-Instruct

English
3
31
143
28.5K
Jade Copet retweetledi
AK
AK@_akhaliq·
MAGNeT by Meta text-to-music and text-to-audio huggingface.co/papers/2401.04… MAGNeT, a masked generative sequence modeling method that operates directly over several streams of audio tokens. Unlike prior work, MAGNeT is comprised of a single-stage, non-autoregressive transformer. During training, we predict spans of masked tokens obtained from a masking scheduler, while during inference we gradually construct the output sequence using several decoding steps. To further enhance the quality of the generated audio, we introduce a novel rescoring method in which, we leverage an external pre-trained model to rescore and rank predictions from MAGNeT, which will be then used for later decoding steps. Lastly, we explore a hybrid version of MAGNeT, in which we fuse between autoregressive and non-autoregressive models to generate the first few seconds in an autoregressive manner while the rest of the sequence is being decoded in parallel. We demonstrate the efficiency of MAGNeT for the task of text-to-music and text-to-audio generation and conduct an extensive empirical evaluation, considering both objective metrics and human studies. The proposed approach is comparable to the evaluated baselines, while being significantly faster (x7 faster than the autoregressive baseline). Through ablation studies and analysis, we shed light on the importance of each of the components comprising MAGNeT, together with pointing to the trade-offs between autoregressive and non-autoregressive modeling, considering latency, throughput, and generation quality.
English
9
119
517
52.6K
Jade Copet retweetledi
Yann LeCun
Yann LeCun@ylecun·
How to make machines understand the world, reason, plan, and learn as efficiently as animals humans? Meta-FAIR is hiring scientists to work on this question. We call this Advanced Machine Intelligence (AMI). Help us build the next generation of AI. Open positions are available in all our locations in NORAM (Bay Area, Seattle, Montreal, New York, Pittsburgh) and EMEA (Paris, London, Tel Aviv). FAIR believes in open research. If you are at NeurIPS, come see us at the Meta booth. Just click here: metacareers.com/jobs/872265983…
English
70
121
976
210.6K
Jade Copet
Jade Copet@jadecopet·
We present MusicGen at #NeurIPS2023 on Thu at 10.45am. Poster #603 "Simple and Controllable Music Generation”. MusicGen is a state-of-the-art text-to-music generation model with optional control over the melody. It can generate high-quality samples, both mono and stereo. 🧵[1/6]
Jade Copet tweet media
English
1
27
126
35.4K
Jade Copet retweetledi
Alexandre Défossez
Alexandre Défossez@honualx·
Curious about using NLP pretrained LLMs for speech? Turns out you can just fine tune them. Poster number #543, 5pm to 7pm local time with @MichaelHassid. See you there ;) See @adiyossLC great recap for more 👇
Yossi Adi@adiyossLC

Unfortunately I could not attend @NeurIPSConf this year :( However, in case you are attending and interesting in SpeechLMs, checkout our work (Wed. 3-5 p.m. PST) presented by @MichaelHassid!! Paper: arxiv.org/abs/2305.13009 Demo, code & models: pages.cs.huji.ac.il/adiyoss-lab/tw…

English
0
2
9
1.3K