Close Menu
    Main Menu
    • Home
    • News
    • Tech
    • Robotics
    • ML & Research
    • AI
    • Digital Transformation
    • AI Ethics & Regulation
    • Thought Leadership in AI

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    The Way forward for Agentic Coding – O’Reilly

    February 13, 2026

    GPT‑5.3-Codex vs Claude Opus 4.6

    February 13, 2026

    The Scale vs Ethics Debate Defined

    February 13, 2026
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Facebook X (Twitter) Instagram
    UK Tech InsiderUK Tech Insider
    Home»Robotics»How can robots purchase abilities by interactions with the bodily world? An interview with Jiaheng Hu
    Robotics

    How can robots purchase abilities by interactions with the bodily world? An interview with Jiaheng Hu

    Arjun PatelBy Arjun PatelFebruary 13, 2026No Comments6 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    How can robots purchase abilities by interactions with the bodily world? An interview with Jiaheng Hu
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    One of many key challenges in constructing robots for family or industrial settings is the necessity to grasp the management of high-degree-of-freedom methods reminiscent of cell manipulators. Reinforcement studying has been a promising avenue for buying robotic management insurance policies, nonetheless, scaling to complicated methods has proved difficult. Of their work SLAC: Simulation-Pretrained Latent Motion House for Entire-Physique Actual-World RL, Jiaheng Hu, Peter Stone and Roberto Martín-Martín introduce a way that renders real-world reinforcement studying possible for complicated embodiments. We caught up with Jiaheng to seek out out extra.

    What’s the subject of the analysis in your paper and why is it an attention-grabbing space for examine?

    This paper is about how robots (specifically, family robots like cell manipulators) can autonomously purchase abilities by way of interacting with the bodily world (i.e. real-world reinforcement studying). Reinforcement studying (RL) is a normal studying framework for studying from trial-and-error interplay with an surroundings, and has enormous potential in permitting robots to study duties with out people hand-engineering the answer. RL for robotics is a really thrilling discipline, as it might open potentialities for robots to self-improve in a scalable approach, in the direction of the creation of general-purpose family robots that may help folks in our on a regular basis lives.

    What had been a number of the points with earlier strategies that your paper was making an attempt to handle?

    Beforehand, many of the profitable functions of RL to robotics had been carried out by coaching completely in simulation, then deploying the coverage within the real-world immediately (i.e. zero-shot sim2real). Nevertheless, such a way has massive limitations: on one hand, it’s not very scalable, as it is advisable to create task-specific, high-fidelity simulation environments that extremely match the real-world surroundings that you just wish to deploy the robotic in, and this will typically take days or months for every job. Then again, some duties are literally very arduous to simulate, as they contain deformable objects and contact-rich interactions (for instance, pouring water, folding garments, wiping whiteboard). For these duties, the simulation is commonly fairly completely different from the actual world. That is the place real-world RL comes into play: if we will permit a robotic to study by immediately interacting with the bodily world, we don’t want a simulator anymore. Nevertheless, whereas a number of makes an attempt have been made in the direction of realizing real-world RL, it’s truly a really arduous drawback since: 1. Pattern-inefficiency: RL requires a whole lot of samples (i.e. interplay with the surroundings) to study good habits, which is commonly unimaginable to gather in massive portions within the real-world. 2. Security Points: RL requires exploration, and random exploration within the real-world is commonly very very harmful. The robotic can break itself and can by no means have the ability to get well from that.

    May you inform us in regards to the methodology (SLAC) that you just’ve launched?

    So, creating high-fidelity simulations may be very arduous, and immediately studying within the real-world can also be actually arduous. What ought to we do? The important thing thought of SLAC is that we will use a low-fidelity simulation surroundings to help subsequent real-world RL. Particularly, SLAC implements this concept in a two-step course of: in step one, SLAC learns a latent motion house in simulation by way of unsupervised reinforcement studying. Unsupervised RL is a way that permits the robotic to discover a given surroundings and study task-agnostic behaviors. In SLAC, we design a particular unsupervised RL goal that encourages these behaviors to be protected and structured.

    Within the second step, we deal with these realized behaviors as the brand new motion house of the robotic, the place the robotic does real-world RL for downstream duties reminiscent of wiping whiteboards by making selections on this new motion house. Importantly, this methodology permit us to avoid the 2 greatest drawback of real-world RL: we don’t have to fret about questions of safety for the reason that new motion house is pretrained to be at all times protected; and we will study in a sample-efficient approach as a result of our new motion house is educated to be very structured.

    The robotic finishing up the duty of wiping a whiteboard.

    How did you go about testing and evaluating your methodology, and what had been a number of the key outcomes?

    We check our strategies on an actual Tiago robotic – a excessive degrees-of-freedom, bi-manual cell manipulation, on a collection of very difficult real-world duties, together with wiping a big whiteboard, cleansing a desk, and sweeping trash right into a bag. These duties are difficult from three features: 1. They’re visuo-motor duties that require processing of high-dimensional picture data. 2. They require the whole-body movement of the robotic (i.e. controlling many degrees-of-freedom on the similar time), and three. They’re contact-rich, which makes it arduous to simulate precisely. On all of those duties, our methodology permits us to study high-performance insurance policies (>80% success charge) inside an hour of real-world interactions. By comparability, earlier strategies merely can not clear up the duty, and sometimes danger breaking the robotic. So to summarize, beforehand it was merely not attainable to resolve these duties by way of real-world RL, and our methodology has made it attainable.

    What are your plans for future work?

    I feel there may be nonetheless much more to do on the intersection of RL and robotics. My eventual objective is to create actually self-improving robots that may study completely by themselves with none human involvement. Extra just lately, I’ve been all for how we will leverage basis fashions reminiscent of vision-language fashions (VLMs) and vision-language-action fashions (VLAs) to additional automate the self-improvement loop.

    About Jiaheng

    Jiaheng Hu is a 4th-year PhD scholar at UT-Austin, co-advised by Prof. Peter Stone and Prof. Roberto Martín-Martín. His analysis curiosity is in Robotic Studying and Reinforcement Studying, with the long-term objective of creating self-improving robots that may study and adapt autonomously in unstructured environments. Jiaheng’s work has been revealed at top-tier Robotics and ML venues, together with CoRL, NeurIPS, RSS, and ICRA, and has earned a number of greatest paper nominations and awards. Throughout his PhD, he interned at Google DeepMind and Ai2, and is a recipient of the Two Sigma PhD Fellowship.

    Learn the work in full

    SLAC: Simulation-Pretrained Latent Motion House for Entire-Physique Actual-World RL, Jiaheng Hu, Peter Stone, Roberto Martín-Martín.




    AIhub
    is a non-profit devoted to connecting the AI neighborhood to the general public by offering free, high-quality data in AI.

    AIhub
    is a non-profit devoted to connecting the AI neighborhood to the general public by offering free, high-quality data in AI.




    Lucy Smith
    is Senior Managing Editor for Robohub and AIhub.

    Lucy Smith
    is Senior Managing Editor for Robohub and AIhub.

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Arjun Patel
    • Website

    Related Posts

    Methods to Choose the Proper Conduit System for Industrial Automation Purposes

    February 13, 2026

    7 RoboDK Academy Programs That Will Develop your Robotics Experience

    February 12, 2026

    Apptronik brings in one other $520M to ramp up Apollo manufacturing

    February 12, 2026
    Top Posts

    Evaluating the Finest AI Video Mills for Social Media

    April 18, 2025

    Utilizing AI To Repair The Innovation Drawback: The Three Step Resolution

    April 18, 2025

    Midjourney V7: Quicker, smarter, extra reasonable

    April 18, 2025

    Meta resumes AI coaching utilizing EU person knowledge

    April 18, 2025
    Don't Miss

    The Way forward for Agentic Coding – O’Reilly

    By Oliver ChambersFebruary 13, 2026

    AI coding assistants have shortly moved from novelty to necessity, the place as much as…

    GPT‑5.3-Codex vs Claude Opus 4.6

    February 13, 2026

    The Scale vs Ethics Debate Defined

    February 13, 2026

    The Rising Danger Of Uncovered ChatGPT API Keys

    February 13, 2026
    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    UK Tech Insider
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms Of Service
    • Our Authors
    © 2026 UK Tech Insider. All rights reserved by UK Tech Insider.

    Type above and press Enter to search. Press Esc to cancel.