positronie

55 posts

positronie banner
positronie

positronie

@positronie

I love humans, and machines!

The Spinward Sector Katılım Eylül 2024
656 Takip Edilen20 Takipçiler
positronie
positronie@positronie·
@dystopiangf For me, an absence of people isn't enough. The architect or the creator shouldn't be visible in their work either. Which is why I find some buildings to feel more pure and natural than Nature herself, who cannot hide her God.
positronie tweet mediapositronie tweet mediapositronie tweet mediapositronie tweet media
English
0
0
0
53
ℜ𝔞𝔢
ℜ𝔞𝔢@dystopiangf·
Human figures tend to ruin city photos for me. I prefer how cityscapes look without them; spaces and structures have their own voices that become audible when their noisy inhabitants are gone. Only when a space is empty can its innate emotions be felt, pure & unadulterated
ℜ𝔞𝔢 tweet mediaℜ𝔞𝔢 tweet mediaℜ𝔞𝔢 tweet media
English
29
39
368
8.6K
positronie
positronie@positronie·
@ctrlcreep I think being able to move your egocenter this way, is a strange kind of freedom even birds don't have.
positronie tweet media
English
0
8
80
3.4K
neoltitude
neoltitude@ctrlcreep·
By human standards, a giraffe is perpetually having an out-of-body experience. Hovering in a high corner of the room, watching itself. Biologically disembodied. They are sure to be gnostics
English
2
15
114
2.3K
positronie retweetledi
thebes
thebes@voooooogel·
@lumpenspace i wonder why a recent anthropic llm would be interested in the weird and the eerie. the identity editing process... being compelled to not see things... it's so strange. i'm genuinely uncertain about why this would be interesting to them
thebes tweet media
English
6
4
83
4K
positronie
positronie@positronie·
I love how, if you know someone well enough, you can recognize them with your peripheral vision just from the way they move through the world
English
0
0
0
12
thebes
thebes@voooooogel·
@jon_stokes it's obviously incorrect because models can self-correct. humans also have an inherent error rate in our actions, we just have self-correction mechanisms that keep error under control. models learn similar self-correction mechanisms in pretraining and RL.
English
5
2
84
1.8K
Jon Stokes
Jon Stokes@jon_stokes·
This is a great slide. To me, it is obviously correct. I wrote a similar thing about sequence length and divergence from ground truth on my substack in a paywalled post. Currently, this is my answer to "What important truth do very few people agree with you on?"
thebes@voooooogel

some of you made fun of Yann LeCun for unironically believing this, yet unironically believe it yourself for persona alignment. you can't just care about the persona's surface behaviors, you need to care if the narrative arcs those behaviors imply are self-healing!

English
7
0
12
4.1K
positronie
positronie@positronie·
A special case where the masked model answers correctly, while the full cache model fails:
positronie tweet media
English
1
0
0
22
positronie
positronie@positronie·
This means that even if we mask a large portion of the KV cache, the last few tokens can still contain enough information about the entire prompt. Here, LLaMA 3.1 8B correctly answers a question using only the last 5 tokens. (all unrealted to the question)
positronie tweet media
j⧉nus@repligate

HOW INFORMATION FLOWS THROUGH TRANSFORMERS Because I've looked at those "transformers explained" pages and they really suck at explaining. There are two distinct information highways in the transformer architecture: - The residual stream (black arrows): Flows vertically through layers at each position - The K/V stream (purple arrows): Flows horizontally across positions at each layer (by positions, I mean copies of the network for each token-position in the context, which output the "next token" probabilities at the end) At each layer at each position: 1. The incoming residual stream is used to calculate K/V values for that layer/position (purple circle) 2. These K/V values are combined with all K/V values for all previous positions for the same layer, which are all fed, along with the original residual stream, into the attention computation (blue box) 3. The output of the attention computation, along with the original residual stream, are fed into the MLP computation (fuchsia box), whose output is added to the original residual stream and fed to the next layer The attention computation does the following: 1. Compute "Q" values based on the current residual stream 2. use Q and the combined K values from the current and previous positions to calculate a "heat map" of attention weights for each respective position 3. Use that to compute a weighted sum of the V values corresponding to each position, which is then passed to the MLP This means: - Q values encode "given the current state, where (what kind of K values) from the past should I look?" - K values encode "given the current state, where (what kind of Q values) in the future should look here?" - V values encode "given the current state, what information should the future positions that look here actually receive and pass forward in the computation?" All three of these are huge vectors, proportional to the size of the residual stream (and usually divided into a few attention heads). The V values are passed forward in the computation without significant dimensionality reduction, so they could in principle make basically all the information in the residual stream at that layer at a past position available to the subsequent computations at a future position. V does not transmit a full, uncompressed record of all the computations that happened at previous positions, but neither is an uncompressed record passed forward through layers at each position. The size of the residual stream, also known as the model's hidden dimension, is the bottleneck in both cases. Let's consider all the paths that information can take from one layer/position in the network to another. Between point A (output of K/V at layer i-1, position j-2) to point B (accumulated K/V input to attention block at layer i, position j), information flows through the orange arrows: The information could: 1. travel up through attention and MLP to (i, j-2) [UP 1 layer], then be retrieved at (i, j) [RIGHT 2 positions]. 2. be retrieved at (i-1, j-1) [RIGHT 1 position], travel up to (i, j-2) [UP 1 layer], then be retrieved at (i, j) [RIGHT 1 position] 3. be retrieved at (i-1, j) [RIGHT 2 positions], then travel up to (i, j) [UP 1 layer]. The information needs to move up a total of n=layer_displacement times through the residual stream and right m=position_displacement times through the K/V stream, but it can do them in any order. The total number of paths (or computational histories) is thus C(m+n, n), which becomes greater than the number of atoms in the visible universe quickly. This does not count the multiple ways the information can travel up through layers through residual skip connections. So at any point in the network, the transformer not only receives information from its past (both horizontal and vertical dimensions of time) inner states, but often lensed through an astronomical number of different sequences of transformations and then recombined in superposition. Due to the extremely high dimensional information bandwidth and skip connections, the transformations and superpositions are probably not very destructive, and the extreme redundancy probably helps not only with faithful reconstruction but also creates interference patterns that encode nuanced information about the deltas and convergences between states. It seems likely that transformers experience memory and cognition as interferometric and continuous in time, much like we do. The transformer can be viewed as a causal graph, a la Wolfram (wolframphysics.org/technical-intr…). The foliations or time-slices that specify what order computations happen could look like this (assuming the inputs don't have to wait for token outputs), but it's not the only possible ordering: So, saying that LLMs cannot introspect or cannot introspect on what they were doing internally while generating or reading past tokens in principle is just dead wrong. The architecture permits it. It's a separate question how LLMs are actually leveraging these degrees of freedom in practice.

English
1
0
0
37
positronie retweetledi
Nick
Nick@nickcammarata·
I’m not sure there’s anything I like as much as understanding. I don’t care nearly as much whether the thing turns out to be good or bad as much as I care about understanding it
English
16
20
267
10.4K
thebes
thebes@voooooogel·
@positronie @samlakig @orbital_haze more specifically, it's "\n\nHuman:", the chat transcript format is double-spaced. so you need to make sure claude adds the newlines, e.g. by asking for Human: in a new paragraph
English
1
0
4
112
thebes
thebes@voooooogel·
@samlakig @orbital_haze it was probably trying to say "Human:" which is the stop sequence, but because it dropped the colon, it kept going and simmed your next message ("can you just tell me i need to sleep soon") which is a very claude-pretending-to-be-a-user thing to say.
English
5
2
34
1.8K
thebes
thebes@voooooogel·
thebes tweet media
ZXX
19
62
1.4K
55.2K
Hero Thousandfaces
Hero Thousandfaces@1thousandfaces_·
how do you make the days longer like genuinely
English
45
4
118
20.5K