aiai-perceptionroboticsresearch

AI Trends in Robotics 2025: From LLM to Embodied AI

Overview of AI trends in Robotics 2025 — foundation models, sim-to-real transfer, and Embodied AI transforming the industry.

Nguyen Anh Tuan10 tháng 7, 20253 phút đọc
AI Trends in Robotics 2025: From LLM to Embodied AI

AI and Robotics are Converging

The AI trends in Robotics 2025 mark a turning point when AI is no longer just a data processing tool but becomes the true brain of robots. The development of Large Language Models (LLM) and Vision-Language Models (VLM) has opened the possibility for robots to understand natural language and reason about the physical world.

AI brain visualization with neural networks and robotics

1. Foundation Models for Robotics

Google DeepMind's RT-2 and Open X-Embodiment proved that a single model can control many different robot types. Instead of training a separate model for each robot, a foundation model learns "common sense" about physics and can transfer to new robots with minimal data.

2. Improved Sim-to-Real Transfer

NVIDIA Isaac Sim and MuJoCo have elevated simulation to the point where robots trained entirely in virtual environments can perform well in the real world. Domain randomization techniques and digital twins in manufacturing help narrow the sim-to-real gap.

At VnRobo, we use Isaac Sim to train navigation policies for AMRs before deploying to real robots, reducing testing time by 90%.

Humanoid robot representing Embodied AI trends

3. LLM as Task Planner

Instead of hardcoding action sequences, robots use LLMs to analyze natural language requests and create execution plans. Example: "Pick up the blue block from shelf 3 and place it on the conveyor" gets analyzed by LLM into action primitives.

Notable frameworks:

  • SayCan (Google): Combines LLM reasoning with robot affordances
  • Code as Policies: LLM generates Python code to control robot
  • VoxPoser: Uses VLM to create 3D value maps for manipulation

4. Dexterous Manipulation

Robot hands are making major advances thanks to tactile sensing and RL. OpenAI's Shadow Hand (Rubik's cube solving) inspired numerous research efforts. In 2025, affordable dexterous hands from Leap Hand and Chinese startups are bringing manipulation closer to real applications.

5. Edge AI for Robots

Specialized AI chips (NVIDIA Jetson Orin, Hailo-8, Qualcomm RB5) enable AI deployment on embedded devices without cloud dependency. This is critical for:

  • Low latency (under 10ms for control loop)
  • Offline operation in factories without stable internet
  • Data security — data stays within factory

AI chips and edge computing devices for autonomous robots

Impact on Vietnam Market

Vietnam is transitioning from labor-intensive manufacturing to automation. These AI trends create major opportunities:

  • Lower barriers: Foundation models accelerate robot deployment, don't need deep AI specialists
  • Cost reduction: Edge AI chips getting cheaper, AMRs under 500M VND already appearing
  • Workforce: Demand for Robotics + AI engineers surging, opportunity for Vietnamese students

To go deeper into AI applications in manufacturing, see computer vision for quality inspection — one of the most common AI robotics applications in Vietnamese factories.

Conclusion

AI is transforming robots from repetitive machines into adaptive intelligent systems. Whether you're a traditional automation engineer or AI specialist, now is the best time to enter robotics. VnRobo commits to sharing knowledge and tools so Vietnam's engineering community doesn't miss this opportunity.

NT

Nguyễn Anh Tuấn

Robotics & AI Engineer. Building VnRobo — sharing knowledge about robot learning, VLA models, and automation.

Bài viết liên quan

NEWTutorial
NVIDIA Newton 1.0: GPU Physics 475x Nhanh Hơn MJX
simulationnvidiaphysics-enginegpusim-to-realisaac-labrobotics

NVIDIA Newton 1.0: GPU Physics 475x Nhanh Hơn MJX

Hướng dẫn thực hành NVIDIA Newton 1.0 — physics engine mã nguồn mở nhanh nhất cho sim-to-real robotics, tăng tốc 475x so với MJX trên GPU.

17/4/202611 phút đọc
NEWTutorial
Hướng dẫn GigaBrain-0: VLA + World Model + RL
vlaworld-modelreinforcement-learninggigabrainroboticsmanipulation

Hướng dẫn GigaBrain-0: VLA + World Model + RL

Hướng dẫn chi tiết huấn luyện VLA bằng World Model và Reinforcement Learning với framework RAMP từ GigaBrain — open-source, 3.5B params.

12/4/202611 phút đọc
NEWDeep Dive
Gemma 4 cho Robotics: AI mã nguồn mở chạy trên Edge
ai-perceptionedge-computinggemmagoogleopen-source

Gemma 4 cho Robotics: AI mã nguồn mở chạy trên Edge

Phân tích Gemma 4 của Google — mô hình AI mã nguồn mở hỗ trợ multimodal, agentic, chạy trên Jetson và Raspberry Pi cho robotics.

12/4/202612 phút đọc