Close Menu
SkytikSkytik

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    At Least 32 People Dead After a Mine Bridge Collapsed Due to Overcrowding

    November 17, 2025

    Here’s how I turned a Raspberry Pi into an in-car media server

    November 17, 2025

    Beloved SF cat’s death fuels Waymo criticism

    November 17, 2025
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    SkytikSkytik
    • Home
    • AI Tools
    • Online Tools
    • Tech News
    • Guides
    • Reviews
    • SEO & Marketing
    • Social Media Tools
    SkytikSkytik
    Home»AI Tools»Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training
    AI Tools

    Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training

    AwaisBy AwaisApril 7, 2026No Comments2 Mins Read0 Views
    Facebook Twitter Pinterest LinkedIn Telegram Tumblr Email
    Measuring Intelligence Efficiency of Local AI
    Share
    Facebook Twitter LinkedIn Pinterest Email

    [Submitted on 30 Mar 2026 (v1), last revised 4 Apr 2026 (this version, v2)]

    View a PDF of the paper titled OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training, by Haiyue Song and Masao Utiyama

    View PDF
    HTML (experimental)

    Abstract:Continual pre-training is widely used to adapt LLMs to target languages and domains, yet the mixture ratio of training data remains a sensitive hyperparameter that is expensive to tune: they must be fixed before training begins, and a suboptimal choice can waste weeks of compute. In this work, we propose OptiMer, which decouples ratio selection from training: we train one CPT model per dataset, extract each model’s distribution vector, which represents the parameter shift induced by that dataset, and search for optimal composition weights post-hoc via Bayesian optimization. Experiments on Gemma 3 27B across languages (Japanese, Chinese) and domains (Math, Code) show that OptiMer consistently outperforms data mixture and model averaging baselines with 15-35 times lower search cost. Key findings reveal that 1) the optimized weights can be interpreted as data mixture ratios, and retraining with these ratios improves data mixture CPT, and 2) the same vector pool can be re-optimized for a given objective without any retraining, producing target-tailored models on demand. Our work establishes that data mixture ratio selection, traditionally a pre-training decision, can be reformulated as a post-hoc optimization over distribution vectors, offering a more flexible paradigm for continual pre-training.

    Submission history

    From: Haiyue Song [view email]
    [v1]
    Mon, 30 Mar 2026 18:00:02 UTC (1,202 KB)
    [v2]
    Sat, 4 Apr 2026 01:16:42 UTC (1,210 KB)

    Continual data Distribution merging Mixing Optimal Pretraining Vector
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Awais
    • Website

    Related Posts

    [2604.03057] Querying Structured Data Through Natural Language Using Language Models

    April 7, 2026

    [2506.16255] Category-based Galaxy Image Generation via Diffusion Models

    April 6, 2026

    The Geometry Behind the Dot Product: Unit Vectors, Projections, and Intuition

    April 6, 2026

    Stabilizing Rubric Integration Training via Decoupled Advantage Normalization

    April 6, 2026

    ChatGPT Search Is Citing Fewer Sites, Data Shows

    April 6, 2026

    How to Run Claude Code Agents in Parallel

    April 6, 2026
    Leave A Reply Cancel Reply

    Top Posts

    At Least 32 People Dead After a Mine Bridge Collapsed Due to Overcrowding

    November 17, 20250 Views

    Here’s how I turned a Raspberry Pi into an in-car media server

    November 17, 20250 Views

    Beloved SF cat’s death fuels Waymo criticism

    November 17, 20250 Views
    Don't Miss

    Higher standards, AI influence, and a web still catching up

    April 7, 2026

    Is it possible to get an accurate view of the current state of SEO? There…

    Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training

    April 7, 2026

    What is AI lead generation? (+tools)

    April 7, 2026

    Navigate Feedly Faster with Go To

    April 7, 2026
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    The Top 6 Search Engines & The AI Search Engines To Know

    April 6, 2026

    Is AI Killing Web Traffic? How AI Overviews Impact Organic Website Traffic

    April 6, 2026
    Most Popular

    13 Trending Songs on TikTok in Nov 2025 (+ How to Use Them)

    November 18, 20257 Views

    How to watch the 2026 GRAMMY Awards online from anywhere

    February 1, 20263 Views

    Corporate Reputation Management Strategies | Sprout Social

    November 19, 20252 Views
    Our Picks

    At Least 32 People Dead After a Mine Bridge Collapsed Due to Overcrowding

    November 17, 2025

    Here’s how I turned a Raspberry Pi into an in-car media server

    November 17, 2025

    Beloved SF cat’s death fuels Waymo criticism

    November 17, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest YouTube Dribbble
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions
    • Disclaimer

    © 2025 skytik.cc. All rights reserved.

    Type above and press Enter to search. Press Esc to cancel.