Prediction of Delay-Free Scene for Quadruped Robot Teleoperation: Integrating Delayed Data With User Commands

Seunghyeon Ha, Seongyong Kim, Soo Chul Lim

Research output: Contribution to journalArticlepeer-review

Abstract

Teleoperation systems are utilized in various controllable systems, including vehicles, manipulators, and quadruped robots. However, during teleoperation, communication delays can cause users to receive delayed feedback, which reduces controllability and increases the risk faced by the remote robot. To address this issue, we propose a delay-free video generation model based on user commands that allows users to receive real-time feedback despite communication delays. Our model predicts delay-free video by integrating delayed data (video, point cloud, and robot status) from the robot with the user's real-time commands. The LiDAR point cloud data, which is part of the delayed data, is used to predict the contents of areas outside the camera frame during robot rotation. We constructed our proposed model by modifying the transformer-based video prediction model VPTR-NAR to effectively integrate these data. For our experiments, we acquired a navigation dataset from a quadruped robot, and this dataset was used to train and test our proposed model. We evaluated the model's performance by comparing it with existing video prediction models and conducting an ablation study to verify the effectiveness of its utilization of command and point cloud data.

Original languageEnglish
Pages (from-to)2846-2853
Number of pages8
JournalIEEE Robotics and Automation Letters
Volume10
Issue number3
DOIs
StatePublished - 2025

Keywords

  • Deep learning methods
  • telerobotics and teleoperation
  • visual learning

Fingerprint

Dive into the research topics of 'Prediction of Delay-Free Scene for Quadruped Robot Teleoperation: Integrating Delayed Data With User Commands'. Together they form a unique fingerprint.

Cite this