Marcelo Magnasco

505 posts

Marcelo Magnasco

Marcelo Magnasco

@MagnascoLab

biophysicist, neuroscientist, dolphin communication research.

Se unió Ağustos 2015
408 Siguiendo191 Seguidores
Marcelo Magnasco retuiteado
Bill Kristol
Bill Kristol@BillKristol·
If I can see it coming. If Trump loses, the election commentariat is going to blame Trump's mistakes in the last week. You know who won't get credit they deserve? Kamala Harris, Nancy Pelosi, Liz Cheney, and tens of millions of women who will be responsible for defeating Trump.
English
2.2K
6.9K
43.8K
1.1M
Marcelo Magnasco retuiteado
Liam Nissan™
Liam Nissan™@theliamnissan·
Yup
Liam Nissan™ tweet media
1.1K
5.1K
27.1K
591.7K
Marcelo Magnasco retuiteado
Carla ‘Bluechecked’ Marinucci
I was arrested & interrogated in Argentina during the 70's "Dirty War'' in which its president sought to rid the country of "the enemy within" with the help of his military. 30,000 Argentines disappeared. Americans should be terrified at such talk from a presidential candidate.
Kaitlan Collins@kaitlancollins

“Yes, I think we should take those words seriously,” former Trump Defense Secretary Mark Esper says after Trump suggested using the U.S. military against the “enemy from within.”

English
370
5.1K
12K
522.8K
Marcelo Magnasco retuiteado
No Context Brits
No Context Brits@NoContextBrits·
No Context Brits tweet media
ZXX
1K
24.3K
325K
5.7M
Marcelo Magnasco retuiteado
tern
tern@1goodtern·
Imagine doing a study on willing participants to see if catching covid would damage their brains and then conducting it and finding out that catching covid has damaged their brains. thelancet.com/journals/eclin…
English
69
595
2K
61.3K
Marcelo Magnasco retuiteado
Sam Wang is at samwang.bsky.social
It’s a beautiful day in New York City. You might think that people here have no political power because everything is in the bag for Democrats. But within 100 miles are 10 swing congressional districts where you can canvass or donate. This and more at votemaximizer.org
Sam Wang is at samwang.bsky.social tweet media
English
2
20
55
16.8K
davinci
davinci@leothecurious·
everytime i think about it i end up converging at this conclusion at an intuitive level but i'm not totally sure yet. do concepts that are utterly beyond the comprehension of the natural human mind exist? or is it just all about sample-efficiency above some minimum intelligence threshold?
Tsarathustra@tsarnick

Scott Aaronson says the idea that while we have concepts that are totally inconceivable to a sea snail there should likewise be concepts that are equally inconceivable to us, may not be true as there may be a ceiling on computational universality

English
3
0
1
183
davinci
davinci@leothecurious·
deep RNNs were rarely practical to train due to vanishing gradients and even LSTMs often had gradient problems as well. the brain is highly recurrent but BPTT is most definitely not what it uses. local parameter updates however seem much more practical and also far more efficient. why wouldn't we want an RNN with local update rules instead of a global memory-intensive end2end backprop?
English
1
1
3
598
Leo Dirac
Leo Dirac@leopd·
An excellent reminder about the power of autoregressive transformers to model things other than language. But @karpathy glosses over an important subtlety in expanding LLMs to other problem domains like vision or chemistry. In order for the “autoregressive” part to work, you need to define a single linear sequence for your tokens. For vision we just use taster order because it’s convenient and works well enough (or does it? We recently got scooped on this one.) But for chemistry or other domains picking the sequence order is not obvious, and certainly has some impact on the model’s inductive bias. While in theory a transformer can imitate any neural network architecture, I’m having real doubts that about their generalization for problem domains that aren’t intrinsically linear in their data representation. You can train transformers in a non-autoregressive way, but you lose a lot of advantages to do so.
Andrej Karpathy@karpathy

It's a bit sad and confusing that LLMs ("Large Language Models") have little to do with language; It's just historical. They are highly general purpose technology for statistical modeling of token streams. A better name would be Autoregressive Transformers or something. They don't care if the tokens happen to represent little text chunks. It could just as well be little image patches, audio chunks, action choices, molecules, or whatever. If you can reduce your problem to that of modeling token streams (for any arbitrary vocabulary of some set of discrete tokens), you can "throw an LLM at it". Actually, as the LLM stack becomes more and more mature, we may see a convergence of a large number of problems into this modeling paradigm. That is, the problem is fixed at that of "next token prediction" with an LLM, it's just the usage/meaning of the tokens that changes per domain. If that is the case, it's also possible that deep learning frameworks (e.g. PyTorch and friends) are way too general for what most problems want to look like over time. What's up with thousands of ops and layers that you can reconfigure arbitrarily if 80% of problems just want to use an LLM? I don't think this is true but I think it's half true.

English
19
30
596
189.4K
Marcelo Magnasco
Marcelo Magnasco@MagnascoLab·
@DireDoge1029 @Parskatt @wandedob I’d be far less concerned with the model or manufacturer than the clientele profile for that particular site. The mri at the er at cornell sees different patients from the one at sloan kettering across the street.
English
0
0
1
39
umu
umu@DireDoge1029·
@Parskatt @MagnascoLab @wandedob And then assuming you do have a dataset with multiple models, it can be argued that you're just unnecessarily driving down model performance by splitting machine model away from your training set.
English
1
0
1
203
Anders Eklund
Anders Eklund@wandedob·
I am often invited to review papers on deep learning for medical images. Unfortunately many papers do the same mistake; they split data into training/validation/test on the slice/image/patch level instead of on the patient level. This will lead to inflated test scores, as images from the same patient then can appear in both training and test sets. Since these images can be very similar, the network will perform extremely well on the test images. If you use 2D networks on 3D data, the 2D slices have to be split on the patient level, not on slice level. If you use 2D networks on a dataset where each patient has several 2D images, the images have to be split on the patient level, not on image level. If you use deep learning for digital pathology, the patches have to be split on the patient level, not on patch level. nature.com/articles/s4159… link.springer.com/chapter/10.100… nature.com/articles/s4159…
English
23
108
588
90.6K
Marcelo Magnasco
Marcelo Magnasco@MagnascoLab·
@wandedob @Parskatt Different imagers… you can have a machine somewhere imaging routine cases vs the one next to the ER.
English
1
0
4
493
Marcelo Magnasco retuiteado
Bill Kristol
Bill Kristol@BillKristol·
"If one asked me to what do I think one must principally attribute the singular prosperity and growing force of this people, I would answer that it is to the superiority of its women." -- Alexis de Tocqueville, Democracy in America II, 3.12
English
37
529
2.2K
133.2K
Marcelo Magnasco retuiteado
United Farm Workers
United Farm Workers@UFWupdates·
We organize with Haitian-origin farm workers in NY. They’re as American as the apple pie their work makes possible, but right wing racists are spreading dehumanizing lies. Haitians are not eating pets. They’re feeding America. The anti-Haitian bigotry we’re seeing is repulsive.
English
1.1K
6.9K
28.9K
1M
Marcelo Magnasco retuiteado
Timothy Imholt
Timothy Imholt@TimothyImholt·
I do believe someone didn't understand what this system was designed to do.
Timothy Imholt tweet media
English
184
1.2K
15.1K
797K