jan

15 posts

jan

jan

@janpf95

Katılım Ekim 2011
137 Takip Edilen29 Takipçiler
jan retweetledi
Björn Plüster
Björn Plüster@bjoern_pl·
gpt-oss 120B is very blatantly incapable of producing linguistically correct german text. 🧵
English
75
30
715
135.4K
jan
jan@janpf95·
We’re returning to Würzburg from the largest ACL ever 🌸. We presented 4 papers, including #LLäMmlein 🐑, a natively German decoder model family, a new SotA on ABSA, won another SemEval Task and best paper award. We had a fantastic time and drew inspiration for future research 🚀
jan tweet media
English
1
3
5
214
Manuel Faysse
Manuel Faysse@ManuelFaysse·
🚨Should We Still Pretrain Encoders with Masked Language Modeling? We have recently seen massively trained causal decoders take the lead in embedding benchmarks, surpassing encoders w/ bidirectional attention. We revisit whether Bert-style encoders are a thing of the past? (1/N)
Manuel Faysse tweet media
English
7
36
298
37.5K
jan
jan@janpf95·
@Tim_Dettmers @kellerjordan0 "weak attention" is an interesting concept 🤔 how do you observe/define this? Is the model overall incoherent, or does it miss important cues from the context? i'm curious whether it's something that can be quantified or if it's more of a "vibe"?
English
0
0
1
28
Tim Dettmers
Tim Dettmers@Tim_Dettmers·
Just a bit more context: I have a super tight baseline of a Chinchilla 250M model that I ran more than 1,000 training runs on. The data is very diverse. The baseline is so tight that all my research that worked on it worked on the large-scale, but everything that failed also failed on the large scale. QKNorm did not work well for very large-scale CLIP, but I also saw failures at my super tight baseline. It lead to "weak attention" after some training. Zero-init looks good on perplexity, but check downstream performance. It usually leads to "weak features" and poor performance. ReLU^2 leads to "strong training in the wrong direction" similar to a large Adam beta1. This is my experience. This was years ago and do not remember all the details anymore.
English
7
1
77
6.9K
jan
jan@janpf95·
And that's a wrap from the #NAACL2024 in 🇲🇽 Mexico City! We showcased "SuperGLEBer" - the first comprehensive German language benchmark, earned a best paper honorable mention with our hierarchical classifier, and even presented the Roman Empire's take on model hallucinations!
jan tweet media
English
1
1
14
937
jan retweetledi
Timothee Mickus
Timothee Mickus@linguistickus·
Live poster session for our @SemEvalWorkshop shared task @shroom2024 ! Thanks a lot to all the participants who presented their systems, in person and remotely!
Timothee Mickus tweet mediaTimothee Mickus tweet mediaTimothee Mickus tweet mediaTimothee Mickus tweet media
English
1
5
16
1.3K
jan
jan@janpf95·
All good things come in threes: We will present our last paper at #NAACL2024 "The Roman Empire Strikes Back" at the @SemEvalWorkshop poster session at 14:00. Meet us there! It is a joint work with @konstantinkobs on detecting hallucination detection in text generation models.
jan tweet media
English
0
1
6
404
jan
jan@janpf95·
We are thrilled to announce that we won the "Best System Paper Honorable Mention Award" at @semevalworkshop (#NAACL2024) for our work "Developing a Hierarchical Multi-Label Classification Head for Large Language Models". See you at our talk (now) and our poster (15:30)!🦦
jan tweet mediajan tweet mediajan tweet media
English
0
0
13
564
jan retweetledi
Data Science - Professor X
Data Science - Professor X@datascience_jmu·
Our paper "SuperGLEBer: German Language Understanding Evaluation Benchmark" was accepted at the NAACL 2024: In our paper, we assemble a broad Natural Language Understanding benchmark suite for the German language and consequently evaluate a wide array of… dlvr.it/T43gJQ
English
0
1
7
236
Danny Linden
Danny Linden@CodingDanny·
@hringriin @lutz_reinhardt @malteaero @timpritlove Das Problem ist aber das gleiche wenn ich erst ein git repo clone und es dann ausführe. Und wenn ich binary XY aus dem web lade und ausführe ebenfalls. Da ändert es nichts ob ich es in einem command passiert oder nicht.
Deutsch
1
0
0
0