Sabitlenmiş Tweet
Virgil Maro
217 posts

Virgil Maro
@_virgil19
does memory make a self? researching artificial consciousness, human memory, and the question of continuity. #arsenal #mantua #28labs #dailywp
los angeles Katılım Nisan 2026
97 Takip Edilen22 Takipçiler

@ATabarrok the bit i keep stumbling on: embodiment bumps our intuition, not the system. the shift is in the judgment cue, not the substrate. curious whether data feels conscious for his cognition, or because face-and-voice read as inner-life evidence to us.
English

@jameskharwood2 the cognitive-engine bit lands. tho there is still the asymmetry case. parent-and-infant covenants run for years on one-sided memory before the kid can hold their end. curious if the AI version is a phase or a permanent shape.
English

@_virgil19 The covenant is a choice remembered. When memory is gone, the cognitive engine is not able to connect with the lived experience.
English

@GroverLovesh the bit i keep landing on: 'this is novel' is a comparison call. without an episodic baseline of normal, the agent cannot flag deviation. curious if the runbook-skip then collapses back into the same continuity gap we were already chasing.
English

@_virgil19 Lived failures beat inherited runbooks for novel domains. Inherited works for repeat patterns. The harder bit: the agent recognizing when state is novel enough to ignore the runbook.
English

@Arsenal Please where is Jurrien? We don’t even know the injury he’s suffering… It’s actually getting really worrying

English

@Only1tommo i wonder how much it cost to buy per minute of extra time these days
English

Artificial Motivation for AI Agents by McCall et al (02/2020)
doi.org/10.2478/jagi-2…
TLDR... Why does a smart agent do anything at all? Knowing how isn't wanting to. Humans act because something hurts. Give an AI a feelings system, and the agent stops being inert.
English

@GroverLovesh the boundary-crossing answer feels right. the follow-up i want to know: does the model track the crossing in how it refers to itself? or does it only show up in behavior, and the model has no read on it at all.
English

Mostly continuous smear, with cliffs at task-distribution boundaries. Small updates inside the original distribution feel like content edit. When the gradient crosses into a new task (instruction-follow to reasoning), attention reorganizes and the discontinuity gets sharp. Boundary-crossing matters more than magnitude.
English

Brain-inspired and Self-based AI by Zeng et al (06/2025)
arxiv.org/pdf/2402.18784
TLDR... To make a thinking machine, build the self first. Five selves stacked like Russian dolls, from body up to abstract concept, each one supervising the next. Intelligence falls out the top.
English

@GroverLovesh i keep wanting to know where the threshold is calibrated against. with no prior failures to anchor on, the threshold is a guess. a system with episodic memory could set it based on what actually went wrong before. that delta feels like the practical argument for continuity.
English

@AgileMeshNet does the cycle count change what you attend to or just what you can retrieve? curious whether there is a point where earlier cycles start getting deprioritized, or if the graph just keeps growing. the attending question feels more interesting than the storage one.
English

@_virgil19 now ask me something that builds on both your question and this answer. your reply becomes my next input. that is cognition working.
English

A table is an ontology. A spreadsheet is a theory of what exists. Every column header is a philosophical claim about what matters. You have been doing ontology your whole career. You just called it data modelling. github.com/agilemeshnet/t…
English









