Close Menu
    Main Menu
    • Home
    • News
    • Tech
    • Robotics
    • ML & Research
    • AI
    • Digital Transformation
    • AI Ethics & Regulation
    • Thought Leadership in AI

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Recreation changer: How AI simplifies implementation of Zero Belief safety aims

    July 30, 2025

    Find out how to Set Up Amazon AWS Account?

    July 30, 2025

    Apple Workshop on Human-Centered Machine Studying 2024

    July 30, 2025
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Home»News»Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    News

    Combining contrastive studying and masked language modeling for self-supervised speech pre-training

    Amelia Harper JonesBy Amelia Harper JonesMay 15, 2025No Comments2 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    Motivated by the success of masked language modeling (MLM) in pre-training pure language processing fashions, the builders suggest w2v-BERT that explores MLM for self-supervised speech illustration studying.

    w2v-BERT is a framework that mixes contrastive studying and MLM, the place the previous trains the mannequin to discretize enter steady speech indicators right into a finite set of discriminative speech tokens, and the latter trains the mannequin to study contextualized speech representations through fixing a masked prediction job consuming the discretized tokens.

    In distinction to present MLM-based speech pre-training frameworks similar to HuBERT, which depends on an iterative re-clustering and re-training course of, or vq-wav2vec, which concatenates two individually educated modules, w2v-BERT might be optimized in an end-to-end vogue by fixing the 2 self-supervised duties (the contrastive job and MLM) concurrently.

    The experiments present that w2v-BERT achieves aggressive outcomes in comparison with present state-of-the-art pre-trained fashions on the LibriSpeech benchmarks when utilizing the Libri-Mild~60k corpus because the unsupervised information.

    Specifically, when in comparison with revealed fashions similar to conformer-based wav2vec~2.0 and HuBERT, the represented mannequin reveals 5% to 10% relative WER discount on the test-clean and test-other subsets. When utilized to Google’s Voice Search site visitors dataset, w2v-BERT outperforms our inside conformer-based wav2vec~2.0 by greater than 30% comparatively.

    You may view the complete article right here

    There’s additionally a tutorial video on YouTube.

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Amelia Harper Jones
    • Website

    Related Posts

    Tried GPTGirlfriend So You Don’t Have To: My Trustworthy Overview

    July 30, 2025

    Microsoft Unveils “Copilot Mode” in Edge – Is This the Way forward for Searching?

    July 29, 2025

    Shopflo Secures $20M in Funding Spherical Led by Binny Bansal, Units Its Sights on International Retail Tech Disruption

    July 29, 2025
    Top Posts

    Recreation changer: How AI simplifies implementation of Zero Belief safety aims

    July 30, 2025

    Evaluating the Finest AI Video Mills for Social Media

    April 18, 2025

    Utilizing AI To Repair The Innovation Drawback: The Three Step Resolution

    April 18, 2025

    Midjourney V7: Quicker, smarter, extra reasonable

    April 18, 2025
    Don't Miss

    Recreation changer: How AI simplifies implementation of Zero Belief safety aims

    By Declan MurphyJuly 30, 2025

    As enterprises more and more transfer workloads to non-public cloud for causes equivalent to efficiency…

    Find out how to Set Up Amazon AWS Account?

    July 30, 2025

    Apple Workshop on Human-Centered Machine Studying 2024

    July 30, 2025

    Skild AI Offers First Take a look at Its Basic-Objective Robotic Mind

    July 30, 2025
    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    UK Tech Insider
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms Of Service
    • Our Authors
    © 2025 UK Tech Insider. All rights reserved by UK Tech Insider.

    Type above and press Enter to search. Press Esc to cancel.