AI & ML interests

None defined yet.

Recent Activity

alanakbik  updated a model about 4 hours ago
Boldt/Boldt-DC-350M
alanakbik  updated a model about 5 hours ago
Boldt/Boldt-DC-1B
alanakbik  updated a model about 5 hours ago
Boldt/Boldt-1B
View all activity

Organization Card

Welcome to Boldt!

Boldt is a family of German language models developed by the Chair of Machine Learning @ Humboldt-Universität zu Berlin. This organization hosts our models, datasets, and research artifacts related to the Boldt project.

Feel free to explore, download, and experiment with our latest releases! 🚀

🌟 The Boldt Model Family

Our models are trained on our German Dense-Core subset of FineWeb-2, utilizing a multi-epoch training recipe on high-quality data.

Model Parameters Context Window Description
Boldt-DC-350M 350M 2048 Ultra-lightweight base model for constrained environments.
Boldt-DC-1B 1B 2048 Highly optimized 1B base model with top-tier German performance.
Boldt-1B 1B 4096 Extended context and vocabulary, augmented with 6B tokens of high-quality German news data.
Boldt-1B-IT-Preview 1B 4096 Instruction-tuned preview model for chat and zero-shot tasks.

📊 Comparison

Boldt-1B compares favorably on German LLM benchmarks against other similarly-sized models:

Boldt-1B Performance Comparison

It is even competitive with many larger (2B parameter) models. See our paper for the full evaluation.

📖 Research & Artifacts