We just shipped a blog on everything latest on vision language models, including ๐ค GUI agents, agentic VLMs, omni models ๐ multimodal RAG โฏ๏ธ video LMs ๐ค๐ป smol models ..and more! https://huggingface.co/blog/vlms-2025
Seed-Coder released and it's designed for coding tasks, featuring base, instruct, and reasoning variants at an 8B parameter scale developed by ByteDance Seed team. Unlike traditional open source LLMs that rely on human crafted rules or annotated data for curating code pretraining datasets Seed-Coder introduces a model-centric data pipeline. The pipeline processes raw data from GitHub and web archives into four categories: file-level codes, repository-level codes, GitHub commits, and code-related web data.A quality filter LLM, evaluates code (for readability, modularity, clarity, and reusability) by removing the lowest 10% to create a 6 trillion token dataset supporting 89 programming languages. Models: ByteDance-Seed/seed-coder-680de32c15ead6555c75b0e4 Github: https://github.com/ByteDance-Seed/Seed-Coder/tree/master Paper: https://github.com/ByteDance-Seed/Seed-Coder/blob/master/Seed-Coder.pdf
Seed-Coder released and it's designed for coding tasks, featuring base, instruct, and reasoning variants at an 8B parameter scale developed by ByteDance Seed team. Unlike traditional open source LLMs that rely on human crafted rules or annotated data for curating code pretraining datasets Seed-Coder introduces a model-centric data pipeline. The pipeline processes raw data from GitHub and web archives into four categories: file-level codes, repository-level codes, GitHub commits, and code-related web data.A quality filter LLM, evaluates code (for readability, modularity, clarity, and reusability) by removing the lowest 10% to create a 6 trillion token dataset supporting 89 programming languages. Models: ByteDance-Seed/seed-coder-680de32c15ead6555c75b0e4 Github: https://github.com/ByteDance-Seed/Seed-Coder/tree/master Paper: https://github.com/ByteDance-Seed/Seed-Coder/blob/master/Seed-Coder.pdf