This week a few more languages have got 1,000 annotations for the educational quality of data from HuggingFaceFW/fineweb-2.
Why should you care?
The quality of pre-training data can have a big impact on the performance of downstream language models trained on that data (HuggingFaceFW/blogpost-fineweb-v1).
Being able to filter by educational quality is on way of improving the quality of the data you use for training an LLM. Very importantly this approach can also reduce the amount of data needed for pertaining.
Why not use an LLM?
LLMs can be used to annotate educational quality for a subset of data. This data can then be used to train a smaller encoder only model to label the full dataset. However, this may not work well for languages outside of english. This is where fineweb-c (community) comes in.
The community is annotating the educational quality of fineweb2 data. Currently 114 languages have some annotations. These annotations will enable a number of things:
- Evaluate whether an LLM can label the educational quality for texts in that language well - Directly be used for training quality classifiers - Help discover other rules and huerisitcs for refining fineweb2 further for different languages.
🚀 Argilla v2.5.0 is out! 🎉 We’re excited to announce the latest version of Argilla, packed with features to make your data annotation workflows more powerful and seamless. Here’s what’s new:
✨ 1. Argilla Webhooks With Argilla webhooks, you can: * Trigger custom workflows * Seamlessly integrate with external tools * Build custom event-driven pipelines
🐍 2. Support for Python 3.13 and Pydantic v2 Argilla v2.5.0 now runs on: * Python 3.13 for enhanced compatibility and speed * Pydantic v2 for improved performance and type validation
🎨 3. Redesigned Home Page Argilla's home page has been redesigned to provide a better user experience, showing a new dataset card view, which provides a better overview of the datasets and annotation progress.