Close Menu
    Main Menu
    • Home
    • News
    • Tech
    • Robotics
    • ML & Research
    • AI
    • Digital Transformation
    • AI Ethics & Regulation
    • Thought Leadership in AI

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    FORT Robotics Launches Wi-fi E-Cease Professional: Actual-Time Wi-fi Security for Advanced Industrial Environments

    January 26, 2026

    Konni Hackers Deploy AI-Generated PowerShell Backdoor Towards Blockchain Builders

    January 26, 2026

    The 5 Varieties Of Organizational Buildings For The New World Of Work

    January 26, 2026
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Home»News»Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    News

    Combining contrastive studying and masked language modeling for self-supervised speech pre-training

    Amelia Harper JonesBy Amelia Harper JonesMay 15, 2025No Comments2 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    Motivated by the success of masked language modeling (MLM) in pre-training pure language processing fashions, the builders suggest w2v-BERT that explores MLM for self-supervised speech illustration studying.

    w2v-BERT is a framework that mixes contrastive studying and MLM, the place the previous trains the mannequin to discretize enter steady speech indicators right into a finite set of discriminative speech tokens, and the latter trains the mannequin to study contextualized speech representations through fixing a masked prediction job consuming the discretized tokens.

    In distinction to present MLM-based speech pre-training frameworks similar to HuBERT, which depends on an iterative re-clustering and re-training course of, or vq-wav2vec, which concatenates two individually educated modules, w2v-BERT might be optimized in an end-to-end vogue by fixing the 2 self-supervised duties (the contrastive job and MLM) concurrently.

    The experiments present that w2v-BERT achieves aggressive outcomes in comparison with present state-of-the-art pre-trained fashions on the LibriSpeech benchmarks when utilizing the Libri-Mild~60k corpus because the unsupervised information.

    Specifically, when in comparison with revealed fashions similar to conformer-based wav2vec~2.0 and HuBERT, the represented mannequin reveals 5% to 10% relative WER discount on the test-clean and test-other subsets. When utilized to Google’s Voice Search site visitors dataset, w2v-BERT outperforms our inside conformer-based wav2vec~2.0 by greater than 30% comparatively.

    You may view the complete article right here

    There’s additionally a tutorial video on YouTube.

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Amelia Harper Jones
    • Website

    Related Posts

    Pricing Choices and Useful Scope

    January 25, 2026

    Yumchat AI Chatbot Assessment: Key Options & Pricing

    January 24, 2026

    A Missed Forecast, Frayed Nerves and a Lengthy Journey Again

    January 24, 2026
    Top Posts

    Evaluating the Finest AI Video Mills for Social Media

    April 18, 2025

    Utilizing AI To Repair The Innovation Drawback: The Three Step Resolution

    April 18, 2025

    FORT Robotics Launches Wi-fi E-Cease Professional: Actual-Time Wi-fi Security for Advanced Industrial Environments

    January 26, 2026

    Midjourney V7: Quicker, smarter, extra reasonable

    April 18, 2025
    Don't Miss

    FORT Robotics Launches Wi-fi E-Cease Professional: Actual-Time Wi-fi Security for Advanced Industrial Environments

    By Arjun PatelJanuary 26, 2026

    Designed to unlock robotic productiveness in warehousing, manufacturing, and development, the brand new Professional mannequin…

    Konni Hackers Deploy AI-Generated PowerShell Backdoor Towards Blockchain Builders

    January 26, 2026

    The 5 Varieties Of Organizational Buildings For The New World Of Work

    January 26, 2026

    5 Breakthroughs in Graph Neural Networks to Watch in 2026

    January 26, 2026
    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    UK Tech Insider
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms Of Service
    • Our Authors
    © 2026 UK Tech Insider. All rights reserved by UK Tech Insider.

    Type above and press Enter to search. Press Esc to cancel.