Celian Ringwald

883 posts

Celian Ringwald banner
Celian Ringwald

Celian Ringwald

@ringwald_c

Datartisan working with knowledge, graphs and texts - Phd Student at INRIA, 3IA, CNRS, I3S, UCA

@[email protected] Katılım Mart 2015
959 Takip Edilen334 Takipçiler
Celian Ringwald retweetledi
Gallicagram
Gallicagram@gallicagram·
🔴Gallicagram v2, nous voici !🔴 Nouvelle interface en react vachement plus stable et rapide, nouveaux corpus, recherche contextuelle, comparaisons inter-corpus, filtre rubrique, bilingue, infinite scrolling... on vous explique tout ! gallicagram.com
Français
1
11
17
2.1K
Celian Ringwald
Celian Ringwald@ringwald_c·
@zehavoc I had the experience with deepl recently that proposed me a full sentence when i wanted to translate a simple word from french to english…
English
0
0
1
22
Djamé..
Djamé..@zehavoc·
Weird... I was used to see google search hallucinates stuff a bit on this "research synthesis" but not on the links themselves. Look at the third link. I was like "Quake 3 ? cool" but the paper only talks about some poker game #Enshitification
Djamé.. tweet media
English
1
0
0
87
Celian Ringwald
Celian Ringwald@ringwald_c·
My PhD thesis is almost finished, so I wrote an article on Medium to explain it in simple terms. "Once upon a time, there was a Kastor 🦫 (semantic beaver) in the forest of Wikipedia knowledge..." @3iacotedazur/semantic-relation-extraction-with-frugal-language-models-3314db2559a6" target="_blank" rel="nofollow noopener">medium.com/@3iacotedazur/… #SLM #KG #RelationExtraction #SHACL #NLP #SemanticWeb
English
0
0
1
112
Celian Ringwald
Celian Ringwald@ringwald_c·
-Echos from the room of text, triples & constraints- Really excited to visit the KCL Knowledge Graphs Lab over the next two weeks! A stay that sounds like a hyperbolic time chamber, to discuss and join forces on LLM/Constraints RQs, before writing the last pages of my PhD
Celian Ringwald tweet media
English
0
0
1
119
Alexander Doria
Alexander Doria@Dorialexander·
Oulipo-styled benchmark: write a story without the letter e. So far Gemini winning vs. OpenAI. And generally better than I would expect given tokenization.
Alexander Doria tweet media
English
4
0
6
1.4K
Celian Ringwald retweetledi
Wimmics Team
Wimmics Team@wimmics·
LLM & Linked Data by Fabien Gandon, Inria @W3C AC 2025 member meeting. He highlighted how Linked Data standards enable knowledge extraction, sharing, and machine learning across domains like robotics, culture, medicine, and chemistry. youtu.be/CVFhPYTVBlI?si…
YouTube video
YouTube
English
0
3
7
460
Celian Ringwald
Celian Ringwald@ringwald_c·
I am looking forward #ESWC2025 in Portorož next week ! The occasion to present our last work: Kastor - 🦫 A framework that includes Human-in-the-loop and combines #KG and #SLMs to produce #RDF and shape-based relation extractors. > Author version: hal.science/hal-05078493v1
Celian Ringwald tweet media
English
0
2
11
371
Celian Ringwald
Celian Ringwald@ringwald_c·
@debayan woo ! i didn’t realized that it was really possible and acceptable to do it: youtube.com/watch?v=1kwbp8… in this talk the sankana team are proposing to create ia generated paper track, maybe a better idea than bothering reviewers in regular tracks
YouTube video
YouTube
English
0
0
0
30
debayan
debayan@debayan·
@ringwald_c I see! I have a feeling I am reviewing some AI written papers!
English
1
0
0
35
debayan
debayan@debayan·
Is AI capable of writing code that is SoTA on a dataset, and also the paper supporting the code?
English
1
0
0
121
Celian Ringwald
Celian Ringwald@ringwald_c·
@debayan really good questions, i was thinking about small models but with llm this is even harder to evaluate 😶‍🌫️ high perf of these models make us forget sometimes this potential memorisation/generalization illusion
English
0
0
0
37
debayan
debayan@debayan·
@ringwald_c Yes, but should we consider the date overlap as a factor during review? Should authors really try to use an LLM that dates older than dataset release to be completely scientifically sound from an evaluation perspective?
English
1
0
1
44
debayan
debayan@debayan·
When reviewing a paper that uses an LLM backend for a popular QA dataset, how do I exclude the possibility of test-set leakage, as the LLM may have memorised the dataset during pre-training?
English
1
0
1
120
Celian Ringwald
Celian Ringwald@ringwald_c·
Just received my #LookingForAPostDoc tshirt :) See you soon for talking about shapes, rdf based relation extraction and Kastor (beavers) in a wood of knowledge.
Celian Ringwald tweet media
English
0
1
8
245
Celian Ringwald retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
When working with LLMs I am used to starting "New Conversation" for each request. But there is also the polar opposite approach of keeping one giant conversation going forever. The standard approach can still choose to use a Memory tool to write things down in between conversations (e.g. ChatGPT does so), so the "One Thread" approach can be seen as the extreme special case of using memory always and for everything. The other day I've come across someone saying that their conversation with Grok (which was free to them at the time) has now grown way too long for them to switch to ChatGPT. i.e. it functions like a moat hah. LLMs are rapidly growing in the allowed maximum context length *in principle*, and it's clear that this might allow the LLM to have a lot more context and knowledge of you, but there are some caveats. Few of the major ones as an example: - Speed. A giant context window will cost more compute and will be slower. - Ability. Just because you can feed in all those tokens doesn't mean that they can also be manipulated effectively by the LLM's attention and its in-context-learning mechanism for problem solving (the simplest demonstration is the "needle in the haystack" eval). - Signal to noise. Too many tokens fighting for attention may *decrease* performance due to being too "distracting", diffusing attention too broadly and decreasing a signal to noise ratio in the features. - Data; i.e. train - test data mismatch. Most of the training data in the finetuning conversation is likely ~short. Indeed, a large fraction of it in academic datasets is often single-turn (one single question -> answer). One giant conversation forces the LLM into a new data distribution it hasn't seen that much of during training. This is in large part because... - Data labeling. Keep in mind that LLMs still primarily and quite fundamentally rely on human supervision. A human labeler (or an engineer) can understand a short conversation and write optimal responses or rank them, or inspect whether an LLM judge is getting things right. But things grind to a halt with giant conversations. Who is supposed to write or inspect an alleged "optimal response" for a conversation of a few hundred thousand tokens? Certainly, it's not clear if an LLM should have a "New Conversation" button at all in the long run. It feels a bit like an internal implementation detail that is surfaced to the user for developer convenience and for the time being. And that the right solution is a very well-implemented memory feature, along the lines of active, agentic context management. Something I haven't really seen at all so far. Anyway curious to poll if people have tried One Thread and what the word is.
English
668
551
6.6K
829.3K