Adam Hibble

3.8K posts

Adam Hibble banner
Adam Hibble

Adam Hibble

@Algomancer

I generate models that generate other stuff, working on @mancerlabs -- Prev: Founder of Popgun Labs (Techstars), Founder of the @QUTCode Network.

Katılım Ağustos 2013
1.1K Takip Edilen4K Takipçiler
Sabitlenmiş Tweet
Adam Hibble
Adam Hibble@Algomancer·
New followers. No researcher is to be found here. I am a novel capabilities engineer doing random walk. (read: unhinged)
English
1
0
8
615
Adam Hibble
Adam Hibble@Algomancer·
New followers. No researcher is to be found here. I am a novel capabilities engineer doing random walk. (read: unhinged)
English
1
0
8
615
Adam Hibble
Adam Hibble@Algomancer·
there is something really magical about giving an agent access to the dom directly setup a js term, embed it, open claude, give it an api it can curl to to execute js - pretty fun. minify an app to a single file, insert it on the fly via curl. I think if this becomes like, less of a security nightmare its totally a thing.
English
0
0
10
725
Adam Hibble
Adam Hibble@Algomancer·
This is my new favourite way to start a project. Self-bootstrapping Claude app. Three files, one loop. Tell the agent to modify itself. It will. github.com/Algomancer/cla…
English
1
0
6
322
Adam Hibble
Adam Hibble@Algomancer·
we went from memory is cheap so just write it in node.js to a ram shortage computing is cool again
English
0
0
0
199
Adam Hibble
Adam Hibble@Algomancer·
One thing is for sure, i love the models, but it's really impressive how janky all cli interfaces are.
English
0
0
5
359
Adam Hibble
Adam Hibble@Algomancer·
Distil from your friends.
Adam Hibble tweet media
English
0
0
2
241
Adam Hibble
Adam Hibble@Algomancer·
Who would have guessed democratisation could have the ability to centralise power so much. wait.
GIF
English
0
0
4
211
Adam Hibble
Adam Hibble@Algomancer·
This looks cool, reading through it now. Free gains if you swap out the Patch embed for the Online PCA Head from JiT. (Just down project to patch_dim * 0.25 before the embed) This looks like it'd compose nicely with the split mean flow formulation. Might be able to make it 1-NFE pretty easily especially because the base NFE is already pretty small. arxiv.org/abs/2507.16884
English
1
0
2
39
Deressa Wodajo
Deressa Wodajo@DeressaWodajo·
🧵1/ Introducing Generative Anchored Fields (GAF) Paper: arxiv.org/abs/2511.22693… Code: github.com/IDLabMedia/GAF Main idea: Instead of a single trajectory/score predictor, we learn independent endpoint predictors, J (noise) and K (data), from any point on a linear bridge.
English
1
0
2
104
Adam Hibble
Adam Hibble@Algomancer·
To me it seems like generally, you want some larger dimensionality then your datum, for the linear independence separability, as well as a dense assignment to entangle it losslessly from that higher dimensionality. I think there is a lot of ideas from representation learning that can inspired disentanglement and vice versa. For example, even the basic version of training a model and then doing monte-carlo high dimension random non-linear (ie - mlp with random weights per batch) projections and and pushing off-diagonal terms in the covariance matrix toward zero in the high dimension projected space gives a bunch of linear independence in the feature space. would be interesting to regularise a flow using something simple like that, bijective models in general with regularisation terms in under explored.
English
1
0
0
18
Adam Hibble
Adam Hibble@Algomancer·
Thinking about the geometry of vae's on my nightime walk today. Not really sure where im going with it yet if anywhere. Every point in latent space has a kinda of split. Directions the decoder cares about, and directions it doesn't. Any point in latent space you can ask if I wiggle in this direction how much does the output change. Some directions change the output a lot, for others, the decoder is essentially ignoring them. Effectively, the standard VAE puts a fixed Gaussian prior over the entire latent space, treating every direction equally. But that's probably geometrically wrong. You're sampling uniformly over a space where huge regions are directions the decoder ignores. So, what I was thinking about is the question of 'what happens if the prior knew about this split?' If it said "spread out along the directions the decoder is using, but stay tight along the directions it isn't". The decoder already knows this, the matrix of partial derivatives (its jacobian) tells you exactly which directions it's sensitive to and by how much. The singular values of that Jacobian are a measure of how much our decoder stretchs along each direction, intuitively, the prior would say "go wherever the decoder is paying attention" and the encoder is all like "that's exactly where I want to put things". Obviously the rate distortion is always going to come into play, ie, more you compress, the worse your reconstruction. But the VAE has a second trade-off that isn't fundamental If your a particular datum (like an image patch) lives on some ten dimensional manifold in pixel space, but your latent space is sixty dimensional, then there are ten directions the decoder uses and fifty it doesn't. Geometrically it's not strictly a fiber bundle, but this is what I was thinking about initially, the reason fiber bundle language is tempting for me to think about is that when the rank is constant, the kernel defines a foliation. If you design the model with explicit conditioning variable that stratifies the space then you're kinda building something bundle like - ie architecturally you'd be able to impose a base and a fiber. But probs the construction doesn't need the foliation, it only needs something weaker. It just needs to know, at each point, a subspace the decoder ignores. The dimension of that subspace jumps around a bunch during training, so, that'll be something to keep in mind. Anyways, that last bit is a bit in the weeds/rambly, key thing probably worth trying is coupling the prior somehow to the the decoder geometry during training. I've seen some work on this, where people utilise geodesic computation in VAE latent spaces. Or metric tensor to construct priors that respect the learned geometry, or riemannian volume measure. So, its not completely unhinged which is always a good sign.
English
4
2
27
2K
Adam Hibble
Adam Hibble@Algomancer·
@sedielem I should probably put a less toy version together to compare. x.com/Algomancer/sta…
Adam Hibble@Algomancer

quick proof of concept for single shot non autoregressive discrete/catagorical modeling. Ie, a mean flow like, bayesian flow like model for single function evaluation sampling of catagorical/discrete data. github.com/Algomancer/Mea… Before you @ me about mnist, I was coding on the river without internet and didn't wanna drain my battery.

English
0
0
1
222
Adam Hibble
Adam Hibble@Algomancer·
@MParakhin Also, if your looking to use a non invertible archecture and still want the collapse protection, you can get the same properties by doing a trace estimate. github.com/Algomancer/Min… I do this in my jepas, beats eps pully etc.
Adam Hibble tweet media
English
1
0
1
202
Adam Hibble
Adam Hibble@Algomancer·
@scottjmaddox @MParakhin But, you can trivially add the repulsion term, which might help and still be cheaper. Havn't battle tested in high dimension :)
English
1
0
1
42
Adam Hibble
Adam Hibble@Algomancer·
@scottjmaddox @MParakhin this isn't sufficiently equivalent to the kernel in the drifting work. More just playing around with ideas, your milage may vary.
English
1
0
1
43