BigCode

272 posts

BigCode banner
BigCode

BigCode

@BigCodeProject

Open and responsible research and development of large language models for code. #BigCodeProject run by @huggingface + @ServiceNowRSRCH

Bergabung Ağustos 2022
3 Mengikuti9.2K Pengikut
Tweet Disematkan
BigCode
BigCode@BigCodeProject·
Introducing: StarCoder2 and The Stack v2 ⭐️ StarCoder2 is trained with a 16k token context and repo-level information for 4T+ tokens. All built on The Stack v2 - the largest code dataset with 900B+ tokens. All code, data and models are fully open! hf.co/bigcode/starco…
BigCode tweet media
English
13
202
665
222.4K
BigCode me-retweet
BigCode
BigCode@BigCodeProject·
BigCodeArena cannot be built without the support of the BigCode community. We are grateful for the huge credits provided by the @e2b team. We thank @hyperbolic_labs, @nvidia, and @Alibaba_Qwen for providing the model inference endpoints.
English
2
1
9
957
BigCode
BigCode@BigCodeProject·
Introducing BigCodeArena, a human-in-the-loop platform for evaluating code through execution. Unlike current open evaluation platforms that collect human preferences on text, it enables interaction with runnable code to assess functionality and quality across any language.
BigCode tweet media
English
4
29
81
43.9K
BigCode me-retweet
Terry Yue Zhuo
Terry Yue Zhuo@terryyuezhuo·
BigCodeBench @BigCodeProject evaluation framework has been fully upgraded! Just pip install -U bigcodebench With v0.2.0, it's now much easier to use compared to the previous v0.1.* versions. The new version adopts the @Gradio Client API interface from @huggingface Spaces by default, w/o the need for local environment setup, and can be replaced with a custom API if desired. Moreover, the latest version no longer requires running separate commands for each stage (like generate, sanitize, and evaluate), significantly simplifying the workflow. The new version also features Batch Inference — running the LLMs on the BigCodeBench-Full set now takes under 5mins for generation and execution! BTW, the benchmark data has been updated to v0.1.2, improving task instructions and test examples. Some of the updates in this release were inspired by EvalPlus @JiaweiLiu_ . A big thank you for the continued maintenance of EvalPlus and the strong support for BigCodeBench 🤗
Terry Yue Zhuo tweet media
English
1
5
32
6.2K
BigCode me-retweet
Josh
Josh@JoshPurtell·
Evaluating LM agents has come a long way since gpt-4 released in March of 2023. We now have SWE-Bench, (Visual) Web Arena, and other evaluations that tell us a lot about how the best models + architectures do on hard and important tasks. There's still lots to do, though 🧵
English
2
10
44
12K
BigCode me-retweet
Terry Yue Zhuo
Terry Yue Zhuo@terryyuezhuo·
People may think BigCodeBench @BigCodeProject is nothing more than a straightforward coding benchmark, but it is not. BigCodeBench is a rigorous testbed for LLM agents using code to solve complex and practical challenges. Each task demands significant reasoning capabilities for selecting appropriate library APIs and logically connecting them to craft a program. Rather than merely providing high-level instructions, each task comes with detailed requirements to evaluate the model's ability to adhere to all aspects. While language models typically perform well on short and simple tasks, they often struggle with longer and more complex problems (e.g., BigCodeBench-Hard). A model needs to perform well on BigCodeBench before it can be used for agentic software development.
English
5
9
42
6.6K
BigCode me-retweet
Qian Liu
Qian Liu@sivil_taram·
By popular demand, I have released the StarCoder2 code documentation dataset, please check it out ⬇️ hf.co/datasets/Sivil…
English
0
11
50
5K
BigCode me-retweet
BigCode me-retweet
Terry Yue Zhuo
Terry Yue Zhuo@terryyuezhuo·
Today, we are happy to announce the beta mode of real-time Code Execution for BigCodeBench @BigCodeProject, which has been integrated into our Hugging Face leaderboard. We understand that setting up a dependency-based execution environment can be cumbersome, even with the built-in Docker image and Dockerfile. To make the evaluation process more reproducible, we've built an interactive environment for you, with guidance from the @Gradio team! (Special thanks to @evilpingwin 🤗) Please note: (1) The execution process might be slightly slower than what you experience on a local machine, as we are using the basic CPU option. There are some compatibility issues with the upgraded CPU environment, and we are currently exploring solutions. (2) Four tasks in the full set require some tricky setup, which has resulted in a pass rate of 99.6%. We will work to fix these in the next iterations :)
Terry Yue Zhuo tweet media
Terry Yue Zhuo@terryyuezhuo

In the past few months, we’ve seen SOTA LLMs saturating basic coding benchmarks with short and simplified coding tasks. It's time to enter the next stage of coding challenge under comprehensive and realistic scenarios! -- Here comes BigCodeBench, benchmarking LLMs on solving practical and challenging programming tasks! So, can LLMs solve these tasks? - Not yet! 🏆 Pass@1: Humans ace 97%, GPT-4o only hits 50-60%, but DeepSeek-Coder-V2 is tighy at its heels! Check out our leaderboard, data, code, and paper: bigcode-bench.github.io 1/🧵

English
1
14
50
20.3K
BigCode me-retweet
Terry Yue Zhuo
Terry Yue Zhuo@terryyuezhuo·
Ppl are curious about the performance of DeepSeek-Coder-V2-Lite on BigCodeBench. We've added its results, along with a few other models, to the leaderboard! huggingface.co/spaces/bigcode… DeepSeek-Coder-V2-Lite-Instruct is a beast indeed, similar to Magicoder-S-DS-6.7B, but with only 2.4B activated parameters! 🤯 We also update all the code generation results here: github.com/bigcode-projec… Feel free to submit a PR if you want to see other models on BigCodeBench 🤗github.com/bigcode-projec…
Terry Yue Zhuo tweet media
BigCode@BigCodeProject

Introducing 🌸BigCodeBench: Benchmarking Large Language Models on Solving Practical and Challenging Programming Tasks! BigCodeBench goes beyond simple evals like HumanEval and MBPP and tests LLMs on more realistic and challenging coding tasks.

English
0
7
24
5.6K
BigCode me-retweet
Philipp Schmid
Philipp Schmid@_philschmid·
It is time to deprecate HumanEval! 🧑🏻‍💻 @BigCodeProject just released BigCodeBench, a new benchmark to evaluate LLMs on challenging and complex coding tasks focused on realistic, function-level tasks that require the use of diverse libraries and complex reasoning! 👀 🧩 Contains 1,140 tasks with 5.6 test cases each, covering 139 libraries in Python. 📊 Uses Pass@1 with greedy decoding and Elo rating for comprehensive evaluation. 🏆 Best model is GPT-4o 61.1%, followed by DeepSeek-Coder-V2. 🥈 Best open Model is DeepSeek-Coder-V2 with 59.7%, better than Claude 3 Opus or Gemini. 👥 Tasks are created in a three-stage process, including synthetic data generation and cross-validation by humans. 🧱 Evaluation framework and Docker images available for easy reproduction 🔜 Plans to extend to multilingualism. Blog: hf.co/blog/leaderboa… Leaderboard: huggingface.co/spaces/bigcode… Code: github.com/bigcode-projec…
Philipp Schmid tweet media
English
4
51
240
35.9K
BigCode me-retweet
Terry Yue Zhuo
Terry Yue Zhuo@terryyuezhuo·
In the past few months, we’ve seen SOTA LLMs saturating basic coding benchmarks with short and simplified coding tasks. It's time to enter the next stage of coding challenge under comprehensive and realistic scenarios! -- Here comes BigCodeBench, benchmarking LLMs on solving practical and challenging programming tasks! So, can LLMs solve these tasks? - Not yet! 🏆 Pass@1: Humans ace 97%, GPT-4o only hits 50-60%, but DeepSeek-Coder-V2 is tighy at its heels! Check out our leaderboard, data, code, and paper: bigcode-bench.github.io 1/🧵
Terry Yue Zhuo tweet media
BigCode@BigCodeProject

Introducing 🌸BigCodeBench: Benchmarking Large Language Models on Solving Practical and Challenging Programming Tasks! BigCodeBench goes beyond simple evals like HumanEval and MBPP and tests LLMs on more realistic and challenging coding tasks.

English
1
37
123
46.8K
BigCode
BigCode@BigCodeProject·
BigCodeBench contains 1,140 function-level tasks to challenge LLMs to follow instructions and compose multiple function calls as tools from 139 Python libraries. To evaluate LLMs rigorously, each programming task encompasses 5.6 test cases with an average branch coverage of 99%.
English
1
0
18
2.1K
BigCode
BigCode@BigCodeProject·
Introducing 🌸BigCodeBench: Benchmarking Large Language Models on Solving Practical and Challenging Programming Tasks! BigCodeBench goes beyond simple evals like HumanEval and MBPP and tests LLMs on more realistic and challenging coding tasks.
GIF
English
9
61
212
102.1K