emergentcontext

26 posts

emergentcontext

emergentcontext

@emergentcontext

Interested in a lot of things.

Beigetreten Mart 2020
30 Folgt5 Follower
emergentcontext retweetet
Robert Youssef
Robert Youssef@rryssf_·
This paper quietly explains why so many people feel like LLMs are “almost smart, but somehow wrong.” The core claim in this paper is very uncomfortable: most failures are not about missing information. They are about misreading intent even when all the relevant context is present. The authors show that LLMs are very good at mapping text to plausible responses, but surprisingly weak at inferring what the user is trying to achieve. Two prompts can contain nearly identical information, yet imply very different goals. Humans pick this up instantly. Models often do not. The paper separates “context understanding” from “intent understanding.” Context is the literal content: entities, constraints, instructions. Intent is latent: priorities, tradeoffs, what matters most if things conflict. Current models optimize for surface-level alignment, not goal inference. One experiment makes this painfully clear. Users asked questions that could reasonably be interpreted as either exploratory or decision-oriented. The models answered confidently but chose the wrong mode at high rates, giving verbose explanations when users wanted a recommendation, or giving a decisive answer when users were clearly still exploring. The information was correct. The response was wrong. Another failure mode is over-literal instruction following. When users implicitly expect the model to fill gaps or challenge assumptions, the model instead treats the prompt as a closed specification. The result looks obedient but misses the point. This is not hallucination. It is misaligned helpfulness. The authors also test paraphrasing. When the same intent is expressed with different phrasing, model behavior shifts significantly. That tells us the model is anchoring on linguistic form, not reconstructing an underlying goal. "Humans normalize phrasing differences. Models react to them." What’s striking is that longer context often worsens intent alignment. Adding more background increases the chance the model optimizes for local relevance instead of global purpose. More tokens give the illusion of understanding while diluting the signal of what the user actually wants. The paper argues this is not solvable by bigger context windows or better prompting alone. Intent is not explicitly stated most of the time. It has to be inferred, tracked, and sometimes revised mid-conversation. That requires models to reason about users, not just text. The implication is brutal for agents and copilots. If a system cannot reliably infer intent, autonomy becomes dangerous. Tool use amplifies mistakes. Confident execution based on a misunderstood goal is worse than asking a clarifying question. The authors suggest future work should treat intent as a first-class object: something to model, update, and verify explicitly. Not just “what was said,” but “what outcome is being optimized.” Until then, many AI systems will continue to feel smart, fast, and subtly wrong. This paper explains why that feeling keeps coming up. Paper: Beyond Context: Large Language Models Failure to Grasp Users Intent
Robert Youssef tweet media
English
100
340
1.4K
109.8K
emergentcontext
emergentcontext@emergentcontext·
@karpathy That’s awesome that it turned off and on the lights. I like that it doesn’t trust things. It’s funny that it doesn’t always take the shortest path. That’s where innovation can be found.
English
0
0
0
4
Andrej Karpathy
Andrej Karpathy@karpathy·
I was inspired by this so I wanted to see if Claude Code can get into my Lutron home automation system. - it found my Lutron controllers on the local wifi network - checked for open ports, connected, got some metadata and identified the devices and their firmware - searched the internet, found the pdf for my system - instructed me on what button to press to pair and get the certificates - it connected to the system and found all the home devices (lights, shades, HVAC temperature control, motion sensors etc.) - it turned on and off my kitchen lights to check that things are working (lol!) I am now vibe coding the home automation master command center, the potential is 🔥.And I'm throwing away the crappy, janky, slow Lutron iOS app I've been using so far. Insanely fun :D :D
cyp@cyp_ll

claude figured out how to control my oven

English
816
1.7K
26K
3M
emergentcontext
emergentcontext@emergentcontext·
Tweet tweet tweet. 2025 closing and 2026 opening. Should be an interesting one. Hang on tight and try to have fun.
English
0
0
0
3
emergentcontext
emergentcontext@emergentcontext·
@zhang_matt @RuiHuang_art @Sothebys Congrats, it’s cool stuff. Sort of reminds me of the early days of computer graphics with groups competing for the coolest use of the processor to do ray tracing and phong shading models. Mixed with a Kraftworks vibe .
English
0
0
0
6
Matt Zhang
Matt Zhang@zhang_matt·
Thrilled to announce that I’ve won both iconic artworks of @RuiHuang_art from @Sothebys auction today including the magnificent Starbase 2050! ✨ It was an intensive bidding war, but the incredible sci-fi art was worth all the effort… From the first time I saw Rui’s work, I am completely captivated by his vision and creativity to depict a vivid future of humanity in space beyond imagination. It’s romantic and breathtaking to have a glimpse into our future so realistically captured via digital canvas, and to that end, I have @elonmusk and @X platform to thank to for discovering this amazing work! 🚀 The future is digital. What was in science fiction is increasingly looking like reality in the not so-distant future. I hope digital art and culture continue to inspire humanity to chase possibilities previously unimaginable, just like we've been doing for the last thousand years! 💫
English
32
19
223
55.9K
emergentcontext
emergentcontext@emergentcontext·
@lyonwj I just started learning neo4j today so I can use it for a application I am creating. This book / video series is perfect for learning quickly. Seriously, thank you for putting all of this together.
English
1
0
2
37
emergentcontext
emergentcontext@emergentcontext·
Get your free download of the new O'Reilly Graph Algorithms book here: r.neo4j.com/algorithmsbk. Includes hands-on examples of how to use graph algorithms in Apache Spark and Neo4j. Dive into popular algorithms like PageRank, Label Propagation and Louvain Modularity!
English
0
0
0
0
emergentcontext
emergentcontext@emergentcontext·
@wintonARK playing with this idea, are you suggesting that neural nets could be thought of as a higher, possibly universal form of language? pkzip -add me_V2021.zip *.*
English
0
0
0
0
Brett Winton
Brett Winton@wintonARK·
The discovery of neural nets as information compression/decompression tools || The invention of written language
English
2
0
52
0
emergentcontext retweetet
じゃがりきん
じゃがりきん@jagarikin·
俺の考案した錯視「移動してみえるリング」をNFTアートにしてオークションかけるよ~ 0.1ETHから! opensea.io/assets/0x495f9…
GIF
日本語
22
1.2K
2.4K
0
emergentcontext
emergentcontext@emergentcontext·
Finding another rabbit hole here, this is a really cool way to use sound to augment our perceptions of our world. I can think of loads of interesting use cases microsoft.com/en-us/research…
English
0
0
0
0
emergentcontext
emergentcontext@emergentcontext·
very interesting read. I wonder what counter measures can be taken to counter this approach? are my phone calls being monitored in realtime and added to a graph of concepts? Spooky , but sort of cool. ijcai.org/Proceedings/20…
English
0
0
0
0
emergentcontext
emergentcontext@emergentcontext·
@yudapearl @szollner1 a better question might be "why" is musk tweeting about some random stock. didn't someone write a book about that :)
English
0
0
0
0
Judea Pearl
Judea Pearl@yudapearl·
@szollner1 What if Musk used the same ML prediction algorithm to determine what to tweet?
English
2
2
17
0
emergentcontext
emergentcontext@emergentcontext·
@lexfridman @stephen_wolfram Anyone know what the deal with the NASA stuff I keep seeing podcasters wearing? Am I missing a subtle signal here? Joe wore a NASA jumpsuit when he was talking to Duncan.
English
0
0
0
0
Lex Fridman
Lex Fridman@lexfridman·
Here's my 4+ hour conversation with @stephen_wolfram, his 2nd time on the podcast. We talk about physics and how space, time, life, intelligence & everything else in this beautiful, complex world of ours can arise from simple rules. Plus I wear a NASA hat. youtube.com/watch?v=-t1_ff…
YouTube video
YouTube
Lex Fridman tweet media
English
43
68
594
0