Antoine Déchappe

About meAbout me

Latest research review

  • SmolVLM: Redefining small and efficient multimodal models

    May 12, 2025

Recent writing

  • Beware of poetry package named differently from project structure

    Mar 27, 2025

  • Pre-commit options in pyproject.toml should be committed first

    Mar 12, 2025

  • Build a chess dataset from PGNs

    Feb 25, 2025

Antoine Déchappe

BlogBlogResearch reviewResearch reviewProjectsProjectsBookmarksBookmarks

Home

❯

Research review

❯

LLMs

❯

Training
    • Research review
      • LLMs
        • Training
          • Branch-Train-Merge: embarrassingly parallel training of expert Language Models
          • Smaller, weaker, yet better: training LLM reasoners via compute-optimal sampling

  • GitHub
  • LinkedIn
  • Email