Carsten T. Lüth

93 posts

Carsten T. Lüth

Carsten T. Lüth

@CarTLueth

PhD Student at the Interactive Machine Learning Research Group, working on efficient label use and low data settings for Deep Learning.

Katılım Ocak 2022
186 Takip Edilen76 Takipçiler
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
When: November 5, 18:00 (Berlin Time) Where: Zoom 🙏 3/3
English
0
0
0
14
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
💥 Event alert: Test-Time Training Agents to Solve Challenging Problems We’re thrilled to welcome Jonas Huebotter, PhD student at ETH Zurich, to our joint heidelberg.ai / NCT Data Science Seminar series on November 5th at 6 pm (CET). 👇Details Below 1/3
Carsten T. Lüth tweet media
English
1
0
1
51
Carsten T. Lüth retweetledi
Heidelberg AI
Heidelberg AI@HeidelbergAi·
🚀 Deploying General AI in the Private World What’s holding AI back in the real world? Join Prof. Seong Joon Oh (@ Tübingen) at the heidelberg.ai / NCT Seminar for insights & new research. 📅 Aug 14, 17:00 (CET) 💻 Online — link in comments 👇 #AI #MachineLearning
Heidelberg AI tweet media
English
1
2
3
158
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
If you ever wondered how Augmented Reality is used in games like Pokémon Go, do not miss this event! Can’t make it live? We’ll upload the recording to our YouTube channel afterward. 📺 YouTube: @HeidelbergAi" target="_blank" rel="nofollow noopener">youtube.com/@HeidelbergAi 3/3
English
0
0
0
72
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
We’re thrilled to welcome Eric Brachmann (@eric_brachmann), Senior Staff Scientist at Niantic Spatial, Inc. and a leading expert in visual relocalisation & pose estimation, to our heidelberg.ai / NCT Data Science Seminar on July 16th at 4 PM. 👇Details Below 1/3
Carsten T. Lüth tweet media
English
2
2
3
137
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
I am 100% in favor of this. We need to establish a practice where we accumulate our knowledge and actively question it actively, with the individuals involved gaining well-accredited credentials.
Rylan Schaeffer@RylanSchaeffer

New position paper! Machine Learning Conferences Should Establish a “Refutations and Critiques” Track Joint w/ @sanmikoyejo @JoshuaK92829 @yegordb @bremen79 @koustuvsinha @in4dmatics @JesseDodge @suchenzang @BrandoHablando @MGerstgrasser @is_h_a @ObbadElyas 1/6

English
0
0
4
196
Carsten T. Lüth retweetledi
Heidelberg AI
Heidelberg AI@HeidelbergAi·
We’re thrilled to welcome Eric Brachmann (@eric_brachmann), Senior Staff Scientist at Niantic Spatial, Inc. and a leading expert in visual relocalisation & pose estimation, to our heidelberg.ai / NCT Data Science Seminar on July 16th at 4 PM. 👇Details Below 1/3
Heidelberg AI tweet media
English
1
2
2
132
Carsten T. Lüth retweetledi
Alex Dimakis
Alex Dimakis@AlexGDimakis·
There are still posts about 'new papers showing AI models cannot reason'. There are unfortunately problems into how these evaluations were done and also many of those limitations are known, peer-reviewed and published. Here is a simplified version of what's going on as far as I understand: Say you interview me, and you ask me to write all the integers from 1 to 2^n, for n=13. I will tell you, here is a Python program that writes them. (Even if you force a human to write them down, they will almost certainly make a typo somewhere and that is what's happening in LMs due to sampling). If you then score me zero despite my correct Python solution, this is not evidence I am not reasoning, it just means the evaluation is flawed. The fact that transformers cannot generalize in unseen domains is known e.g. n-digit multiplication, maze solving and many other puzzles. It was shown also how to make transformers generalize with recursive self-improvement. Further science on different reasoning puzzles is important but should not be taken out of its narrow scientific context, and should be peer-reviewed not reduced to viral soundbites. On a higher level, when AI researchers say a model is reasoning, it means it scores well on reasoning benchmarks like AIME. We should not map how we think humans reason to how machines reason. What is practically relevant is that if you post-train a reasoning model to do a task, it will do it well, and generalize in unseen questions from that domain. (like getting a Gold medal in a Math olympiad, doing your taxes, giving legal or medical pointers etc.). We do not know of any domain where this does not happen when enough data and RL is thrown on the domain. That means a ton of tasks can be automated. Outside well-defined tasks, the question 'but is it truly reasoning?' is not a meaningful question to me.
English
7
18
145
14.2K
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
@ziv_ravid I would really like to hear your thoughts on this. For me LLMs including Reasoning Models serve as a database which query and access unstructered knowledge, I never expected even the reasoning models to“really“ reason.
English
0
0
0
131
Ravid Shwartz Ziv
Ravid Shwartz Ziv@ziv_ravid·
My opinion on Apple's "Illusion of Thinking" paper and the pushback against it: It proved exactly what I thought before!
English
4
3
54
6.1K
Carsten T. Lüth retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
@willccbb Theoretical physicists are the intellectual embryonic stem cell, I’ve now seen them become ~everything.
English
72
111
2.2K
144.7K
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
@giffmana @y_m_asano @giffmana, we would also be very happy to host you for a talk at heidelberg.ai to hear about your latest research. If you would like to revisit Heidelberg again and meet with the research group from which Mint Medical was created as a spin-off, we could do it in person.
English
1
0
1
56
Carsten T. Lüth
Carsten T. Lüth@CarTLueth·
This event was jointly organized by: heidelberg.ai & NCT Data Science Seminar series in collaboration with the ELLIS Unit Heidelberg
English
1
0
0
104