Sabitlenmiş Tweet
Mat McGann
5.9K posts

Mat McGann
@MatMcGann
Above all, grow knowledge ✍🔧👨💻🚀 @healthhorizon (CEO) https://t.co/eTnatQrpRQ (blog)
Canberra, Australia Katılım Haziran 2010
504 Takip Edilen780 Takipçiler

@svembu Agreed. We are building data pipelines using only the reasoning of LLMs to create and maintain structured knowledge bases. The process can extract latent information from unstructured sources. As a result the data has full provenance. Often used to provide reliable RAG.
English

All the LLMs and other deep learning models are based on neural networks. We can think of them as mathematical functions with hundreds of billions of parameters.
Those parameters (weights) are determined during training and we train these networks with trillions of tokens (text, images, videos that are split up into tokens to be ingested by the models).
We can say that every one of the trillions of tokens played a part in determining the value of each of the hundreds of billions of parameters.
The image I have in mind is a giant lake where we dissolve trillions of cubes of salt, sugar etc. After the dissolution we cannot know which of the cubes of sugar went where in the lake - every cube of sugar is everywhere!
Therein lies a problem: if we use a business database, such as customer relationship data, to train a neural network model (i.e to determine its parameters), when the customer changes that data or deletes the data, we do not know how to alter the weights of the model to account for this change in the data. Even if the model were dedicated to that customer, we still cannot guarantee the customer that their changes to the data will be reflected in the model.
In that sense, neural networks (and therefore LLMs) are NOT a suitable database.
This is a fundamental limitation of the current scientific mathematical approach and cannot be fixed only by technological fine tuning.
The RAG (retrieval augmented generation) architecture keeps the business database separate and augments the user prompt with data fetched from the database.
In that case, the model itself is not trained on the (potentially changing) customer data because that data is only used in the prompt.
But RAGs can only go so far.
I personally have come to believe more foundational work is needed. What does that look like? All I have right now are hunches. That is the existing part of scientific work!
English

In matters of civilisation, the only thing that makes things inevitable are thinking they're inevitable
Elon Musk@elonmusk
@stclairashley Civil war is inevitable
English

@averykimball Axioms and theories do seem different but I'm not sure they are. They're all just ideas. When an experiment doesn't give you what you expect it could be your theories or your axioms that are wrong
English

@MatMcGann theories underlying an instrument being wrong doesn't falsify determinism
as you say it's a ground assumption, an axiom, 'falsification' doesn't apply to it- it's used *for* falsifying (like the other suite of consequences inside Realism)
English

I encourage you to look at determinism as the boldest conjecture in history. And wrong.
Mat McGann@MatMcGann
"Casual reasoning [and hence determinism] is a heuristic" 👌👌👌 i.e. determinism is a theory about the world, not a ground truth. youtu.be/nh1Z3UTobrY?si…
English

"Casual reasoning [and hence determinism] is a heuristic" 👌👌👌
i.e. determinism is a theory about the world, not a ground truth.
youtu.be/nh1Z3UTobrY?si…

YouTube
English

This would explain the drastic rise in allergies over the last few decades: Feedback loop:
- Some kids are born allergic
- More awareness of this > less exposure
- less exposure causes emergent allergies
- More awareness > decrease exposure > ...
Crémieux@cremieuxrecueil
Over a decade ago, researchers started a trial to see if they could prevent peanut allergies They gave a few hundred kids peanuts from ages one to five and told parents of another group to have their kids avoid the stuff Peanut consumption reduced peanut allergy rates by a lot:
English