At Foxlink, we are not just building robots—we are creating a comprehensive ecosystem for intelligent autonomy. With NVIDIA’s full-stack robotics platform, we have seamlessly integrated autonomous patrolling, dexterous manipulation, and smart manufacturing to bring AI to life across real-world industrial applications.
Our development framework is built on upon NVIDIA Isaac Sim, NVIDIA Jetson AGX, NVIDIA Omniverse, TensorRT, and NVIDIA DGX, enabling rapid iteration from simulation to training and final deployment.
Our primary platforms focus on two domains:
A dual-arm manipulation system with vision-language understanding powered by NVIDIA Isaac GR00T Dexterity workflow.
Foxlink Jeff: AI-Powered Quadruped with Perception and Mobility
Foxlink Jeff is an all-terrain AI robotic dog designed for patrolling, industrial inspection, and smart infrastructure integration. Built on the NVIDIA platform, Jeff executes real-world perception, decision-making, and autonomous actions. Jeff is not just mobile—it sees, thinks, adapts, and acts as a next-generation AI patrol agent.
Sim2Real Training Workflow: From Isaac Lab to Field Deployment
We designed a complete reinforcement learning training loop using NVIDIA Isaac Lab, an open source, modular robot learning framework, enabling behavioral policies to be learned in simulation and transferred seamlessly to the NVIDIA Jetson AGX platform for real-time inferencing and control.
Simulation ➜ Policy Learning ➜ DGX Training ➜ AGX Deployment ➜ Real-World Execution
Foxlink has also built a three-tier integration framework between simulation and real-world operation:
Via the Metropolis interface, users can:
The result: a fully remotely operable autonomous patrol system.
We use NVIDIA GR00T as our intelligent dual-arm robotic system, integrating computer vision and language understanding.
We use NVIDIA GR00T N1 robot foundation model as our intelligent dual-arm robotic system, integrating computer vision and language understanding.
Core components include:
Data workflow integrates VisionSync teleoperation with fine-tuning pipelines, enabling a flexible and adaptive architecture. This system continuously improves through learning-by-demonstration and policy reinforcement.
Example task:
Command: “Pick up the batteries and place them into the battery box.”
The post-trained GR00T N model interprets the instruction, parses visual inputs, plans the path, and executes the grasping action.
Our data pipeline enables zero-shot learning, data-efficient training, and robust generalization across tasks.
Real Deployment in Foxlink Factory
Deployed use cases include:
What’s Next
We are actively integrating the following advancements:
Final Thoughts Redefining Real AI Robotics
At Foxlink, our mission is not just to get robots moving—but to empower them to understand the world, learn from change, and adapt intelligently.
By leveraging the full NVIDIA AI robotics stack, we are redefining how machines see, think, and act in real-world environments.
Join us on our journey to build the next generation of truly autonomous AI robots.