Eval4NLP

60 posts

Eval4NLP

Eval4NLP

@eval4nlp

Workshop on Evaluation and Comparison of NLP Systems, co-located with #AACL2025.

Katılım Ekim 2019
36 Takip Edilen304 Takipçiler
Eval4NLP
Eval4NLP@eval4nlp·
📢📢 This year Eval4NLP is co-located with #aacl2025. Our CFP is now out: eval4nlp.github.io/2025/cfp.html Paper submission deadline: September 29, 2025. Direct submission via OpenReview. ARR commitment deadline: October 27, 2025 Notification of acceptance: November 3, 2025
English
0
3
5
396
Eval4NLP retweetledi
NLLG
NLLG@NLLG_lab·
📢📢📢The NLLG lab has three new open fully funded PhD positions: 1⃣ - Next Generation LLMs 2⃣ - NLP for Science 3⃣ - Multimodal evaluation metrics Deadlines: End of October ℹ️More information: nl2g.github.io/positions
English
0
3
9
1K
Eval4NLP retweetledi
NLLG
NLLG@NLLG_lab·
📢📢👉👉@chrleiter presenting his work on how explainability can improve evaluation metrics for MT and summarization tomorrow at #eacl2024 twitter.com/ChrLeiter/stat…
Christoph Leiter@ChrLeiter

Excited to present our paper "BMX: Boosting Natural Language Generation Metrics with Explainability" at #EACL2024! Join us in Virtual Poster Session B on 20.03.2024 at 2 p.m. as we unveil how explanations can enhance NLG evaluation metrics.

English
1
1
3
362
Eval4NLP
Eval4NLP@eval4nlp·
Eval4NLP23 has concluded. We thank everyone + congratulate our shared task winners on inducing high-quality metrics for MT+summ. using prompting and efficient models: "HIT-MI&T Lab" (even beating GEMBA + COMET🚀) & "DSBA". Shared task overview paper: arxiv.org/pdf/2310.19792…
English
1
4
16
2K
Eval4NLP
Eval4NLP@eval4nlp·
📢📢To accommodate the recent ARR author response period, Eval4NLP @aaclmeeting extends the deadline for pre-reviewed papers until September 30th. Pre-reviewed papers must include: the paper along with its original reviews and scores. More details: eval4nlp.github.io
English
0
4
5
715
Eval4NLP
Eval4NLP@eval4nlp·
📢📢 Don't forget: Pre-reviewed papers can be submitted to Eval4NLP @aaclmeeting until September 25 via Openreview Just include the paper and your meta-reviews along with reviews and all the scores. eval4nlp.github.io
English
0
2
5
1.2K
Dan Deutsch
Dan Deutsch@_danieldeutsch·
Interested in document-level MT but have been held back by the lack of automatic metrics? If so, you won't want to miss our new paper! We study the quality of sentence-level metrics on long-form text and augment them with paragraph-level training data. arxiv.org/abs/2308.13506
Dan Deutsch tweet media
English
3
8
42
5.7K
Eval4NLP
Eval4NLP@eval4nlp·
@gg42554 @ReviewAcl The reviewing quality is bad in NLP, agreed. But it's also often because there are so many junior people (who sometimes need to step in because there aren't enough reviewers). Exposing them publicly may also be problematic for various reasons.
English
0
0
1
529
Goran Glavaš
Goran Glavaš@gg42554·
Writing the EMNLP rebuttals. I'm now convinced (also after having served for a year as EiC for @ReviewAcl) that nothing short of publicly releasing reviews *with reviewer identities* will substantially improve the (currently appaling) average review quality in #NLProc.
English
4
3
27
3.8K
Eval4NLP
Eval4NLP@eval4nlp·
Due to popular demand, the Eval4NLP workshop @ @aaclmeeting submission deadline has been moved to September 1. We look forward to your submissions! 📣📣 More infos: eval4nlp.github.io
English
0
7
4
1.8K
Markus Freitag
Markus Freitag@markuseful·
Many of you asked me the question about an automatic metric that can give us similar insights as MQM. We (mostly Patrick -- hands down one of the best student researchers I ever worked with) investigated how well LLMs can do MQM like error annotation. We present ... 🥳AutoMQM🥳
Patrick Fernandes@psanfernandes

LLMs still lag behind our best metrics for MT evaluation. But what if we prompted them for fine-grained, interpretable feedback (much like human annotators)? arxiv.org/abs/2308.07286 TLDR: We analyzed their capabilities for MT eval, and propose *AutoMQM* to improve them! 1/14

English
1
2
33
5.4K
Eval4NLP
Eval4NLP@eval4nlp·
📢📢📢 We have released the description of Eval4NLP's shared task on "Prompting LLMs as Explainable Evaluation Metrics" (for MT & summarization). Dev phase: Aug. 7 Test phase: Sep. 18 System Submission Deadline: Sep. 23 More details: eval4nlp.github.io/2023/shared-ta… 🚀🚀🚀
Eval4NLP tweet media
English
0
7
12
1.1K
Eval4NLP
Eval4NLP@eval4nlp·
This year, Eval4NLP is accepting submissions of papers with reviews from other venues (see the CFP for more details #ARR" target="_blank" rel="nofollow noopener">eval4nlp.github.io/2022/cfp.html#…). Submit your paper and reviews here openreview.net/group?id=aclwe… by September 21, AOE!
English
0
3
6
0