Close Menu
    Main Menu
    • Home
    • News
    • Tech
    • Robotics
    • ML & Research
    • AI
    • Digital Transformation
    • AI Ethics & Regulation
    • Thought Leadership in AI

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Malicious Go Packages Impersonate Google’s UUID Library to Steal Delicate Information

    December 6, 2025

    AI denial is changing into an enterprise threat: Why dismissing “slop” obscures actual functionality positive aspects

    December 6, 2025

    The 6 Disciplines of Strategic Pondering For Leaders With Michael Watkins Mega Finest-Promoting Creator of “The First 90 Days”

    December 6, 2025
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Home»News»Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    News

    Combining contrastive studying and masked language modeling for self-supervised speech pre-training

    Amelia Harper JonesBy Amelia Harper JonesMay 15, 2025No Comments2 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Combining contrastive studying and masked language modeling for self-supervised speech pre-training
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    Motivated by the success of masked language modeling (MLM) in pre-training pure language processing fashions, the builders suggest w2v-BERT that explores MLM for self-supervised speech illustration studying.

    w2v-BERT is a framework that mixes contrastive studying and MLM, the place the previous trains the mannequin to discretize enter steady speech indicators right into a finite set of discriminative speech tokens, and the latter trains the mannequin to study contextualized speech representations through fixing a masked prediction job consuming the discretized tokens.

    In distinction to present MLM-based speech pre-training frameworks similar to HuBERT, which depends on an iterative re-clustering and re-training course of, or vq-wav2vec, which concatenates two individually educated modules, w2v-BERT might be optimized in an end-to-end vogue by fixing the 2 self-supervised duties (the contrastive job and MLM) concurrently.

    The experiments present that w2v-BERT achieves aggressive outcomes in comparison with present state-of-the-art pre-trained fashions on the LibriSpeech benchmarks when utilizing the Libri-Mild~60k corpus because the unsupervised information.

    Specifically, when in comparison with revealed fashions similar to conformer-based wav2vec~2.0 and HuBERT, the represented mannequin reveals 5% to 10% relative WER discount on the test-clean and test-other subsets. When utilized to Google’s Voice Search site visitors dataset, w2v-BERT outperforms our inside conformer-based wav2vec~2.0 by greater than 30% comparatively.

    You may view the complete article right here

    There’s additionally a tutorial video on YouTube.

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Amelia Harper Jones
    • Website

    Related Posts

    How Lloyds Banking Group Is Betting Crypto + Code Will Rewrite the Mortgage Rulebook

    December 5, 2025

    Evolving AI from Chatbots to Colleagues That Make An Affect

    December 5, 2025

    Mother! They’ve Acquired Her – However It Was Simply AI”: How a “Actual-Life Horror Film” Performed Out in Kansas

    December 4, 2025
    Top Posts

    Evaluating the Finest AI Video Mills for Social Media

    April 18, 2025

    Utilizing AI To Repair The Innovation Drawback: The Three Step Resolution

    April 18, 2025

    Midjourney V7: Quicker, smarter, extra reasonable

    April 18, 2025

    Meta resumes AI coaching utilizing EU person knowledge

    April 18, 2025
    Don't Miss

    Malicious Go Packages Impersonate Google’s UUID Library to Steal Delicate Information

    By Declan MurphyDecember 6, 2025

    A hidden hazard has been lurking within the Go programming ecosystem for over 4 years.…

    AI denial is changing into an enterprise threat: Why dismissing “slop” obscures actual functionality positive aspects

    December 6, 2025

    The 6 Disciplines of Strategic Pondering For Leaders With Michael Watkins Mega Finest-Promoting Creator of “The First 90 Days”

    December 6, 2025

    Pixi: A Smarter Approach to Handle Python Environments

    December 6, 2025
    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    UK Tech Insider
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms Of Service
    • Our Authors
    © 2025 UK Tech Insider. All rights reserved by UK Tech Insider.

    Type above and press Enter to search. Press Esc to cancel.