Alan Akbik

75 posts

Alan Akbik

Alan Akbik

@alan_akbik

Professor of Machine Learning at Humboldt-Universität zu Berlin

Berlin Katılım Kasım 2014
398 Takip Edilen481 Takipçiler
Alan Akbik
Alan Akbik@alan_akbik·
🎓This semester, you can join us online in my "Deep Learning and NLP" course at HU Berlin! In the course, we go all the way from zero to "NLP hero": from classical NLP to transformers and LLMs, plus in-depth into to PyTorch🔥 How to enrol: hu-berlin.de/en/studies/cou…
Alan Akbik tweet media
English
0
0
3
138
Alan Akbik
Alan Akbik@alan_akbik·
Announcing 🐤 Büble-LM 🐤, our new state-of-the-art 2B language model (LM) for German! Trained by @pieterdelobelle with a novel trans-tokenization approach, it outperforms other German LMs like Sauerkraut and LLäMmlein on most benchmarks. Try it out! huggingface.co/flair/bueble-l…
English
2
10
33
1.2K
Alan Akbik
Alan Akbik@alan_akbik·
Problem: There are thousands of language models on 🤗 @HuggingFace - but which one is the best for *your* NLP task? Solution: ⚖️ TransformerRanker ⚖️! Our newest library directly connects to🤗 HF and quickly/efficiently ranks LMs for your task! github.com/flairNLP/trans…
English
0
0
4
298
Alan Akbik
Alan Akbik@alan_akbik·
2 papers accepted to #EMNLP2024 main conference! See you in Miami in November! ☀️🏖️🍹
Alan Akbik tweet media
English
2
0
20
905
Alan Akbik
Alan Akbik@alan_akbik·
@krasul Most newspapers have topics (such as financial, sports, etc.). Those are automatically parsed and assigned to the respective crawled article!
English
0
0
1
25
Kashif Rasul
Kashif Rasul@krasul·
@alan_akbik are the articles also tagged by type? e.g. financial, sports etc?
English
1
0
0
66
Alan Akbik
Alan Akbik@alan_akbik·
Crawl 1 million news articles in 7 hours*! Announcing the release of 🗞️ Fundus v0.4 🗞️! github.com/flairNLP/fundus *crawling speed depends on your internet connection. I launched it 1.5 days ago and already gathered over 5 million news articles across 41 languages 🔥🔥🔥!
English
1
2
15
687
Alan Akbik
Alan Akbik@alan_akbik·
More details: - The position is funded through the HEIBRiDS programme: heibrids.berlin - PhD project is on NLP for scientific literature (material science) and co-supervised by Dr. Thomas Unold from the Helmholtz Center Berlin - Apply to project 60029!
English
0
0
1
435
Alan Akbik
Alan Akbik@alan_akbik·
Want do a PhD in NLP? We have a new PhD position available, fully funded for 4 years! - ⏰ Application deadline: August 23rd - 🗓️ Start of PhD project: January 2025 - 🔥 Requires strong PyTorch/Python skills, and knowledge of NLP/ML - ✍️ Apply here: heibrids.berlin/admission/open…
English
1
1
10
1.4K
Alan Akbik
Alan Akbik@alan_akbik·
Pieter joins as a visiting researcher from KU Leuven until the end of the year. Hope you enjoy your stay at HU and in Berlin!
English
0
0
1
203
Alan Akbik
Alan Akbik@alan_akbik·
Excited to have Dr. Pieter Delobelle join our lab! Pieter is an LLM-expert, well known for creating Dutch foundation LLMs 🇳🇱🇧🇪. @pieterdelobelle: Welcome to Berlin! We look forward to working together on Dutch/German/English LLMs! 💪 pieter.ai
English
1
0
9
1.6K
Alan Akbik
Alan Akbik@alan_akbik·
@rasbt @rasbt any idea why they are using NLL over the token probabilities in their teacher-student setup? Could they not also use MSE over the final layer hidden state?
English
1
0
0
119
Sebastian Raschka
Sebastian Raschka@rasbt·
3) Knowledge distillation (as in MiniLLM): The general idea is to transfer knowledge from a larger model (the teacher) to a smaller model (the student). Here, they trained a 27B (teacher) model from scratch and then trained the smaller 2B and 9B (student) models on the outputs of the larger teacher model. The 27B model doesn't use knowledge distillation but was trained from scratch to serve as a "teacher" for the smaller models.
English
2
2
22
2.4K
Sebastian Raschka
Sebastian Raschka@rasbt·
What's noteworthy in the newly released Gemma 2 LLMs? The main theme is that they explore techniques w/o necessarily increasing the dataset sizes but rather focus on developing relatively small & efficient LLMs. The are 3 main design choices to create the 2B & 9B models:
Sebastian Raschka tweet media
English
7
80
417
35.6K
Alan Akbik
Alan Akbik@alan_akbik·
Announcing the 🧠LM Pub Quiz🧠- the ultimate test of your LLMs factual knowledge! Probe any LM (masked/causal) on Huggingface with our new library. We mitigate biases (answer distribution, domain, etc.) to give the most exact reading possible. Try it! lm-pub-quiz.github.io
Alan Akbik tweet media
English
0
0
9
374
Alan Akbik
Alan Akbik@alan_akbik·
Our paper "🗞️Fundus🗞️: A Simple-to-Use News Scraper Optimized for High Quality Extractions" accepted to ACL 2024 Demos! Fundus allows you to easily build a high quality corpus of news data for your NLP project. Try it out :) github.com/flairNLP/fundus #NLProc #ACL2024NLP
Alan Akbik tweet media
English
0
0
6
294
Alan Akbik
Alan Akbik@alan_akbik·
@GuillaumeLample Awesome! Have you tried benchmarking on PECC, our problem extraction + coding benchmark? It's extremely challenging for all models we tried, and it would be cool to see how well Codestral-22B fares! hallerpatrick.github.io/pecc/
English
1
0
4
230
Guillaume Lample @ NeurIPS 2024
Guillaume Lample @ NeurIPS 2024@GuillaumeLample·
Today we are releasing Codestral-22B, our first code model! Codestral is trained on more than 80 programming languages and outperforms the performance of previous code models, including the largest ones. It is available on our API platform, through instruct and fill-in-the-middle endpoints, and can be easily integrated into VScode plugins. You can also use it for free on Le Chat: chat.mistral.ai
Guillaume Lample @ NeurIPS 2024 tweet mediaGuillaume Lample @ NeurIPS 2024 tweet media
English
43
157
1.2K
178.9K
Alan Akbik
Alan Akbik@alan_akbik·
Announcing 📊PECC📊, our extremely challenging LLM benchmark for coding and math problems! Even very strong LLMs get less than 50% of coding questions correct. And less than 10% of math! - Presented 9:00 tomorrow (Friday) at #lreccoling2024 - Paper: arxiv.org/abs/2404.18766
Alan Akbik tweet media
English
3
6
19
874