Bobbie-model Official

| Stage | Dataset | Tokens | Purpose | |-------|---------|--------|---------| | 1 | RedPajama (v2) | 1.2T | Base language modeling | | 2 | SlimPajama + CodeAlpaca | 400B | Code & reasoning | | 3 | Synthetic multi-turn chat | 50B | Instruction following |

They explicitly filtered out any data containing eval benchmark examples (MMLU, GSM8K, HumanEval) using 13-gram overlap detection. This means Bobbie's benchmarks are likely not contaminated. 4. Performance Benchmarks We ran Bobbie-7B-Instruct against Llama-3-8B-Instruct and Mistral-7B-v0.3 on an RTX 4090. bobbie-model

If you’ve been following the open-source LLM space, you’ve likely memorized the specs of Llama 3, Mixtral, and Qwen. But a new contender has been quietly gaining traction in the "small model" category: . | Stage | Dataset | Tokens | Purpose