Preston Badeer

2.2K posts

Preston Badeer banner
Preston Badeer

Preston Badeer

@pbadeer

I post about the intersection of 🦾AI, 🤖LLMs, 📊data products, and 📈data engineering.

Turning data into products👉🏻 เข้าร่วม Mart 2015
1.6K กำลังติดตาม525 ผู้ติดตาม
Preston Badeer
Preston Badeer@pbadeer·
@DataChaz The output files: The Video: A compact file containing all the text encoded as QR code frames. The FAISS Index: The binary index for lightning-fast semantic similarity search. The JSON Index: Links the search results from the FAISS index to the correct frame in the video.
English
0
0
0
73
Preston Badeer
Preston Badeer@pbadeer·
@DataChaz To enable fast searching, it creates a search index that maps the semantic meaning of the text to the frame number in the video where the corresponding QR code is. The core process can be visualized as: Text -> Chunking -> JSON Payload -> QR Code Image -> Video Frame ...
English
1
0
0
68
Charly Wargnier
Charly Wargnier@DataChaz·
Crazy. Vector databases for AI memory just got replaced by... MP4 files! You can now store millions of text chunks inside an MP4 file with Memvid, and still get lightning-fast semantic search! 🤯 No database needed. 100% Opensource! repo below ↓
Charly Wargnier tweet media
English
94
250
1.8K
329.4K
Preston Badeer
Preston Badeer@pbadeer·
Chain of Continuous Thought looks dope, very excited to try models trained this way.
English
1
0
1
79
Preston Badeer
Preston Badeer@pbadeer·
It's finally $AMD AI time.
Preston Badeer tweet media
English
1
0
1
99
AI at Meta
AI at Meta@AIatMeta·
New research from Meta FAIR: Large Concept Models (LCM) is a fundamentally different paradigm for language modeling that decouples reasoning from language representation, inspired by how humans can plan high-level thoughts to communicate.
English
59
546
3.1K
531.2K
Preston Badeer
Preston Badeer@pbadeer·
🔥 This is sick. Using code to run simulations is way too uncommon IMO. So many amazing discoveries can be made by developing a simple simulation framework (even without LLMs).
Matthew Berman@MatthewBerman

.@Microsoft just dropped TinyTroupe! Described as "an experimental Python library that allows the simulation of people with specific personalities, interests, and goals." These agents can listen, reply back, and go about their lives in simulated TinyWorld environments.

English
0
0
1
59
Preston Badeer
Preston Badeer@pbadeer·
Swarm is cool but definitely a tutorial. Explicitly not for production and not a library, just an example.
Philipp Schmid@_philschmid

This came unexpected! @OpenAI released Swarm, a lightweight library for building multi-agent systems. Swarm provides a stateless abstraction to manage interactions and handoffs between multiple agents and does not use the Assistants API. 🤔 How it works: 1️⃣ Define Agents, each with its own instructions, role (e.g., "Sales Agent"), and available functions (will be converted to JSON structures). 2️⃣ Define logic for transferring control to another agent based on conversation flow or specific criteria within agent functions. This handoff is achieved by simply returning the next agent to call within the function. 3️⃣ Context Variables provide initial context and update them throughout the conversation to maintain state and share information between agents. 4️⃣ Client run() initiate and manage the multi-agent conversation. It needs an initial agent, user messages, and context and returns a response containing updated messages, context variables, and the last active agent. Insights: 🔄 Swarm manages a loop of agent interactions, function calls, and potential handoffs. 🧩 Agents encapsulate instructions, available functions (tools), and handoff logic. 🔌 The framework is stateless between calls, offering transparency and fine-grained control. 🛠️ Swarm supports direct Python function calling within agents. 📊 Context variables enable state management across agent interactions. 🔄 Agent handoffs allow for dynamic switching between specialized agents. 📡 Streaming responses are supported for real-time interaction. 🧪 The framework is experimental. Maybe to collect feedback? 🔧 Flexible and works with any OpenAI client, e.g., Hugging Face TGI or vLLM-hosted models.

English
1
0
1
98
Preston Badeer
Preston Badeer@pbadeer·
FINALLY got access to @cerebras. They ain't kidding, it's even faster than @GroqInc. 🤯 I'm getting 447/ts on Llama 3.1 70B with JSON parsing. 0.95s round trip!
English
3
1
15
2.2K
Preston Badeer
Preston Badeer@pbadeer·
This is specific to the Instruct models: #llama-3.1-instruct" target="_blank" rel="nofollow noopener">llama.meta.com/docs/model-car…. However, if you're having trouble with Llama 3.1 Instruct 8B on any JSON-mode tasks, I recommend trying 70B before increasing complexity of your pipeline or changing models entirely.
English
0
0
0
38
Preston Badeer
Preston Badeer@pbadeer·
Struggling with Llama 3.1 8B? I wish I had seen this sooner. Meta: "We recommend using Llama 70B-instruct or Llama 405B-instruct for applications that combine conversation and tool calling. Llama 8B-Instruct can not reliably maintain a conversation alongside tool calling definitions. It can be used for zero-shot tool calling, but tool instructions should be removed for regular conversations between the model and the user." (Emphasis added.) Link in reply below
English
1
0
0
70
Preston Badeer
Preston Badeer@pbadeer·
30% is a huge improvement over all the previous hype that was around 13% (Devin), but still not something I would consider even close to production ready. Fast progress though! I hope we get some solid open source options in this space as the commercial ones improve.
Alistair@AlistairPullen

I'm excited to share that we've built the world's most capable AI software engineer, achieving 30.08% on SWE-Bench – ahead of Amazon and Cognition. This model is so much more than a benchmark score: it was trained from the start to think and behave like a human SWE.

English
0
0
0
89
Preston Badeer
Preston Badeer@pbadeer·
Great update on open source, with expanded details in the replies 👇
Vaibhav (VB) Srivastav@reach_vb

What a massive week for Open Source AI: We finally managed to beat closed source fair and square! 1. Meta Llama 3.1 405B, 70B & 8B—The latest in the llama series, this version (base + instruct) comes with multilingual (8 languages) support, a 128K context, and an even more commercially permissive license. The best part: 405B beats GPT4o/ mini fair and square! Bonus: Meta posted a banger of a tech report with quite a lot of details also on upcoming (?) multi-modal (image/ audio/ video) 2. Mistral dropped Large 123B—Dense, multilingual (12 languages), and 128K context. Comes as instruct-only model checkpoint, with performance less than 405B but higher than L3.1 70B. Released under non-commercial license. 3. Nvidia released Minitron distilled 4B & 8B - apache 2.0 license, 256K vocab, with student beating the teacher by 16% on MMLU. Uses iterative pruning and distilling to achieve SoTA! The real question: Who is distilling 405B right now? ;) 4. InternLM shared Step Prover 7B—SoTA on the Lean, which was trained on Github repos with large-scale formal data. Achieves 48.8 pass@1, 54.5 pass@64. They release the dataset, tech report and the fine-tuned InternLM math plus model checkpoint 5. CofeAI dropped Chonky TeleFM 1T - A one trillion parameter dense model trained on 2T tokens, bilingual - Chinese and English, apache 2.0 licensed and tech report. They use a novel progressive upsampling approach. Stability dropped Sv4D, Nvidia released MambaVision, SakanaLabs with Evo (merging + stable diffusion), and more. This was a landmark week, and I'm personally quite happy with the direction of open source AI/ ML! Did I miss anything interesting drop them in comments! 🤗

English
0
0
1
61
Preston Badeer
Preston Badeer@pbadeer·
In the commercial AI use case space, open source models are everything. This details how much of a leader Llama/meta is in this space.
Andrej Karpathy@karpathy

Huge congrats to @AIatMeta on the Llama 3.1 release! Few notes: Today, with the 405B model release, is the first time that a frontier-capability LLM is available to everyone to work with and build on. The model appears to be GPT-4 / Claude 3.5 Sonnet grade and the weights are open and permissively licensed, including commercial use, synthetic data generation, distillation and finetuning. This is an actual, open, frontier-capability LLM release from Meta. The release includes a lot more, e.g. including a 92-page PDF with a lot of detail about the model: ai.meta.com/research/publi… The philosophy underlying this release is in this longread from Zuck, well worth reading as it nicely covers all the major points and arguments in favor of the open AI ecosystem worldview: "Open Source AI is the Path Forward" facebook.com/4/posts/101157… I like to say that it is still very early days, that we are back in the ~1980s of computing all over again, that LLMs are a next major computing paradigm, and Meta is clearly positioning itself to be the open ecosystem leader of it. - People will prompt and RAG the models. - People will finetune the models. - People will distill them into smaller expert models for narrow tasks and applications. - People will study, benchmark, optimize. Open ecosystems also self-organize in modular ways into products apps and services, where each party can contribute their own unique expertise. One example from this morning is @GroqInc , who built a new chip that inferences LLMs *really fast*. They've already integrated Llama 3.1 models and appear to be able to inference the 8B model ~instantly: x.com/karpathy/statu… And (I can't seem to try it due to server pressure) the 405B running on Groq is probably the highest capability, fastest LLM today (?). Early model evaluations look good: ai.meta.com/blog/meta-llam… x.com/alexandr_wang/… Pending still is the "vibe check", look out for that on X / r/LocalLlama over the next few days (hours?). I expect the closed model players (which imo have a role in the ecosystem too) to give chase soon, and I'm looking forward to that. There's a lot to like on the technical side too, w.r.t. multilingual, context lengths, function calling, multimodal, etc. I'll post about some of the technical notes a bit later, once I make it through all the 92 pages of the paper :)

English
0
0
0
53