Michael Tiemann (né Schober)

678 posts

Michael Tiemann (né Schober)

Michael Tiemann (né Schober)

@mschoberml

Research scientist @ Bosch Center for Artificial Intelligence (BCAI). Interested in all things dynamical systems and numerical solvers. Views are my own. He/him

Tübingen, Germany 가입일 Ağustos 2014
708 팔로잉437 팔로워
Ben Aaronovitch
Ben Aaronovitch@Ben_Aaronovitch·
What's a good German substitution for the response 'Did I fuck?' to a question?
English
24
3
42
12.3K
Michael Tiemann (né Schober) 리트윗함
Emtiyaz Khan
Emtiyaz Khan@EmtiyazKhan·
We don't expect Bayesian methods to do so well at large scale, but we can now get decent improvements with variational learning to GPT-2. I wrote a blog about this (first one in a long time). Check it out! team-approx-bayes.github.io/blog/ivon/ Paper: arxiv.org/abs/2402.17641 A thread below.
Emtiyaz Khan tweet media
English
12
54
269
33.9K
Michael Tiemann (né Schober) 리트윗함
Saurabh Srivastava
Saurabh Srivastava@_saurabh·
More than 50% of the reported reasoning abilities of LLMs might not be true reasoning. How do we evaluate models trained on the entire internet? I.e., what novel questions can we ask of something that has seen all written knowledge? Below: new eval, results, code, and paper. Functional benchmarks are a new way to do reasoning evals. Take a popular benchmark, e.g., MATH, and manually rewrite its reasoning into code, MATH(). Run the code to get a snapshot that asks for the same reasoning but not the same question. A reasoning gap exists if a model’s performance is different on snapshots. Big question: Are current SOTA models closer to gap 0 (proper reasoning) or gap 100 (lots of memorization)? What we find: Gaps in the range of 58% to 80% in a bunch of SOTA models. Motivates us to build Gap 0 models. We’re releasing the paper, code, and 3 snapshots of functional MATH() today. arxiv draft: arxiv.org/abs/2402.19450 github repo: github.com/ConsequentAI/f… 1/🧵
Saurabh Srivastava tweet media
English
44
221
1.2K
486.5K
Michael Tiemann (né Schober) 리트윗함
François Chollet
François Chollet@fchollet·
My view of the capabilities of LLMs is probably far below that of the median tech industry person. And yet, the more time passes the more I realize my 2023 views were actually overestimating their future potential and current usefulness. Parallel to self-driving: circa 2016-2017 my view on the timeline for full-scale self-driving deployment was much more pessimistic than most people in the industry -- I was envisioning ~2023, when everyone else targeted 2020 or earlier. And yet, as time passed I started realizing that I was being grossly overoptimistic.
English
63
147
1.6K
346.3K
Michael Tiemann (né Schober) 리트윗함
Lancelot Da Costa
Lancelot Da Costa@lancelotdacosta·
Gaussian processes are the standard for probability distributions over trajectories or paths. But over what paths? Here we fully characterize the sample path regularity of GPs in relation to the covariance kernel arxiv.org/abs/2312.14886
English
3
9
51
10.4K
Michael Tiemann (né Schober) 리트윗함
François Chollet
François Chollet@fchollet·
The "aha" moment when I realized that curve-fitting was the wrong paradigm for achieving generalizable modeling of problems spaces that involve symbolic reasoning was in early 2016. I was trying every possible way to get a LSTM/GRU based model to classify first-order logic statements, and each new attempt was showing a bit more clearly than the last that my models were completely unable to learn to perform actual first-order logic -- despite the fact that this ability was definitely part of the representable function space. Instead, the models would inevitably latch onto statistical keyword associations to make their predictions. It has been fascinating to see this observation echo again and again over the past 8 years.
English
43
197
1.6K
419.5K
Michael Tiemann (né Schober) 리트윗함
François Chollet
François Chollet@fchollet·
Video generation models and Neural Radiance Fields have been improving regularly since 2016, and now they're in the spotlight. As a result there's a been a lot of debate about whether such systems embed a *model of physics*. Let's take a look...
English
20
176
1.2K
318.7K
Michael Tiemann (né Schober) 리트윗함
0xDesigner
0xDesigner@0xDesigner·
in what fucking world can i text prompt a hollywood-level, blockbuster movie-like scene but i can't prompt a simple ui mockup? ai was supposed to take MY job not christopher nolan's why am i still working
English
154
331
5.2K
662.1K
Michael Tiemann (né Schober) 리트윗함
Jascha Sohl-Dickstein
Jascha Sohl-Dickstein@jaschasd·
Have you ever done a dense grid search over neural network hyperparameters? Like a *really dense* grid search? It looks like this (!!). Blueish colors correspond to hyperparameters for which training converges, redish colors to hyperparameters for which training diverges.
English
298
2.2K
11.3K
1.8M
Michael Tiemann (né Schober) 리트윗함
François Chollet
François Chollet@fchollet·
People seem to be falling for two rather thoughtless extremes: 1. "LLMs are AGI, they work like the human brain, they can reason, etc." 2. "LLMs are dumb and useless." Reality is that LLMs are not AGI -- they're a big curve fit to a very large dataset. They work via memorization and interpolation. But that interpolative curve can be tremendously useful, if you want to automate a known task that's a match for its training data distribution. Memorization works, as long as you don't need to adapt to novelty. You don't *need* intelligence to achieve usefulness across a set of known, fixed scenarios. In fact, that's the entire story of the field of AI so far: achieve increasing levels of usefulness and automation, while bypassing the problem of creating intelligence.
English
72
341
2K
238.6K
Michael Tiemann (né Schober) 리트윗함
The Cultural Tutor
The Cultural Tutor@culturaltutor·
A little tour through the impossible and mind-bending worlds of M.C. Escher...
The Cultural Tutor tweet media
English
76
1.1K
10.1K
930.8K
Michael Tiemann (né Schober) 리트윗함
John Burn-Murdoch
John Burn-Murdoch@jburnmurdoch·
NEW: an ideological divide is emerging between young men and women in many countries around the world. I think this one of the most important social trends unfolding today, and provides the answer to several puzzles.
John Burn-Murdoch tweet media
English
1.5K
12.6K
49.5K
27.5M
Michael Tiemann (né Schober) 리트윗함
Jie Huang
Jie Huang@jefffhj·
I authored a critique paper titled "Large Language Models Cannot Self-Correct Reasoning Yet" (arxiv.org/abs/2310.01798) 20 days ago. I’ve observed two distinct groups misinterpreting the content in two different ways: For LLM Critics: "LLMs Cannot Self-Correct Reasoning" != "LLMs Cannot Reason" Consider an individual capable of reasoning but who provides an incorrect solution to a problem and fails to correct their own error. This incapacity for self-correction does not negate their reasoning ability. I did, however, express doubts about whether LLMs can genuinely reason in my survey paper last year (arxiv.org/abs/2212.10403). For LLM Enthusiasts: Leveraging external feedback for improvement does not equate to LLMs having the capacity to "self"-improve. High-quality external feedback is often unavailable, and even when it is, it may not be characterized as "self"-critique but rather as "critique with external feedback". My Two Cents: 1) Avoid overclaiming your results; 2) Do not exaggerate your "critique"; otherwise, you become no different from those who overstate their results.
Yann LeCun@ylecun

Anyone who thinks Auto-Regressive LLMs are getting close to human-level AI, or merely need to be scaled up to get there, *must* read this. AR-LLMs have very limited reasoning and planning abilities. This will not be fixed by making them bigger and training them on more data.

English
5
30
154
71.3K
Michael Tiemann (né Schober) 리트윗함
Mark Tenenholtz
Mark Tenenholtz@marktenenholtz·
Business analysts: please god save us from Excel. we'll do anything. Microsoft:
Mark Tenenholtz tweet media
English
23
44
487
112.6K
Michael Tiemann (né Schober) 리트윗함
Karpi
Karpi@karpi·
I've asked an AI to generate a trailer for a HEIDI movie and now I can never sleep again
English
2.1K
12.9K
50.7K
19.3M
Michael Tiemann (né Schober) 리트윗함
Dr. Casey Fiesler is no longer on here
If AI ethicists need a doctorate in CS to be qualified to critique AI, then AI researchers need to have a doctorate in ethics/humanities/philosophy/HCI/etc. to be qualified to build AI.
English
8
182
851
68.3K
Michael Tiemann (né Schober) 리트윗함
ELLIS
ELLIS@ELLISforEurope·
A statement by the ELLIS Board: In this text, the members of the ELLIS Board share their view on the global conversation about the societal risks of #AI. ➡️ellis.eu/news/our-view-…
ELLIS tweet media
English
0
34
53
14.8K