Xiaofei Du 杜笑菲 🦄

204 posts

Xiaofei Du 杜笑菲 🦄 banner
Xiaofei Du 杜笑菲 🦄

Xiaofei Du 杜笑菲 🦄

@xiaofei_instill

Founder & COO @instill_tech | Building AI, trusted by knowledge workers | ➡️ Join our early access: https://t.co/Re7LdfPOrq

London Katılım Mayıs 2012
21 Takip Edilen51 Takipçiler
Xiaofei Du 杜笑菲 🦄 retweetledi
Instill AI
Instill AI@instill_tech·
How is Instill AI different from other #AI chat assistants? - Persistent knowledge - AI workflow automation - Customizable analysis Beta release this week, try it for free! Instill AI slideshow: gamma.app/docs/Instill-A…
English
0
1
4
202
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
LLMs lack real opinions—no self-awareness, subjectivity, or growth. Their “insights” are just patterns & context guesses, not true understanding. Even with perfect memory, they’d mimic, not judge. So, without needing LLMs to have “souls,” what memory features can we build? - How do they spot key insights vs. noise in a chat? - How can long chats build deeper, abstract memories? Tried marathon chats with an LLM? Share your discoveries! #AIThoughts #FutureOfAI
English
0
0
0
39
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
Andrej says “New Chat” forces LLMs to clear their context to patch these flaws—it’s a compromise for tech limits. But we need better memory, not resets. I’m all in 🙌 Think about a friend who’s known you 3 years—they reflect, grow, & evolve from your chats. LLMs? They might recall bits—your name, pet’s name, projects—but don’t truly “get” you as a whole. Why? #Innovation #AIResearch
English
1
0
0
43
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
Why do ChatGPT, Grok, & Perplexity all have a “New Chat” design? Is it necessary or just lazy? Let’s dive into LLMs & memory! #AI #LLM
Andrej Karpathy@karpathy

When working with LLMs I am used to starting "New Conversation" for each request. But there is also the polar opposite approach of keeping one giant conversation going forever. The standard approach can still choose to use a Memory tool to write things down in between conversations (e.g. ChatGPT does so), so the "One Thread" approach can be seen as the extreme special case of using memory always and for everything. The other day I've come across someone saying that their conversation with Grok (which was free to them at the time) has now grown way too long for them to switch to ChatGPT. i.e. it functions like a moat hah. LLMs are rapidly growing in the allowed maximum context length *in principle*, and it's clear that this might allow the LLM to have a lot more context and knowledge of you, but there are some caveats. Few of the major ones as an example: - Speed. A giant context window will cost more compute and will be slower. - Ability. Just because you can feed in all those tokens doesn't mean that they can also be manipulated effectively by the LLM's attention and its in-context-learning mechanism for problem solving (the simplest demonstration is the "needle in the haystack" eval). - Signal to noise. Too many tokens fighting for attention may *decrease* performance due to being too "distracting", diffusing attention too broadly and decreasing a signal to noise ratio in the features. - Data; i.e. train - test data mismatch. Most of the training data in the finetuning conversation is likely ~short. Indeed, a large fraction of it in academic datasets is often single-turn (one single question -> answer). One giant conversation forces the LLM into a new data distribution it hasn't seen that much of during training. This is in large part because... - Data labeling. Keep in mind that LLMs still primarily and quite fundamentally rely on human supervision. A human labeler (or an engineer) can understand a short conversation and write optimal responses or rank them, or inspect whether an LLM judge is getting things right. But things grind to a halt with giant conversations. Who is supposed to write or inspect an alleged "optimal response" for a conversation of a few hundred thousand tokens? Certainly, it's not clear if an LLM should have a "New Conversation" button at all in the long run. It feels a bit like an internal implementation detail that is surfaced to the user for developer convenience and for the time being. And that the right solution is a very well-implemented memory feature, along the lines of active, agentic context management. Something I haven't really seen at all so far. Anyway curious to poll if people have tried One Thread and what the word is.

English
1
0
0
113
Xiaofei Du 杜笑菲 🦄 retweetledi
Instill AI
Instill AI@instill_tech·
So much video, so little time! Instill AI - a new personal #AI assistant can load 100s of videos and allow you to search, summarize, catalog, and extract key info, quotes, or clips. It's in beta now, and we'd love your feedback - use it for free! instill-ai.com/use-cases/ai-v…
English
0
3
3
203
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
Fed up with ChatGPT’s one-off analysis or NotebookLM’s tiny scope? Big docs, structure, deep insights—brutal hurdle. Built @instill_tech's Early Access beta to turn PDFs/Word/PPT into clear tables for easy analysis. Who’s stuck too? DM me.
English
3
1
3
168
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
Does it save time? Kinda. Still spent hours for polish 😬, but for quick decks (60%), it’s a win. Worth it? Yes—Killed my “blank page” dread. Watermark off + cheap = no-brainer. I’m a subscriber now! 👍 #Gamma #AITools
English
1
0
1
43
Xiaofei Du 杜笑菲 🦄 retweetledi
Instill AI
Instill AI@instill_tech·
Sr. AI Engineer, George Strong explains how #Test_Time_Compute has disrupted AI: - Enabling #LRMs from @deepseek_ai to answer questions with greater depth - Revolutionized generative image synthesis - Making Agentic AI possible eu1.hubs.ly/H0hg-JL0
Instill AI tweet media
English
0
2
4
178
Xiaofei Du 杜笑菲 🦄 retweetledi
Instill AI
Instill AI@instill_tech·
What’s the worst OCR challenge you’ve seen? 😬 Poor exposure? Hand-written? Instill AI can understand the worst of them! See how Instill AI organizes and answers questions about the contents. Use it for free eu1.hubs.ly/H0h7SbS0
Instill AI tweet media
English
0
3
5
162
Xiaofei Du 杜笑菲 🦄
Xiaofei Du 杜笑菲 🦄@xiaofei_instill·
Chatty AI like ChatGPT or Perplexity is slick for quick answers, but when it comes to wrangling documents for real knowledge work? It’s a mess—lost context, no organization, just a pile of digital scrap paper. open.substack.com/pub/xiaofeidu/…
Xiaofei Du 杜笑菲 🦄 tweet media
English
0
2
3
146