AI EducademyAIEducademy
🌳

AI 学习路径

🌱
AI 种子

从零开始

🌿
AI 萌芽

打好基础

🌳
AI 枝干

付诸实践

🏕️
AI 树冠

深入探索

🌲
AI 森林

精通AI

🔨

工程技能路径

✏️
AI 草图

从零开始

🪨
AI 雕刻

打好基础

⚒️
AI 匠心

付诸实践

💎
AI 打磨

深入探索

🏆
AI 杰作

精通AI

查看所有学习计划→

实验室

已加载 7 个实验
🧠神经网络游乐场🤖AI 还是人类?💬提示实验室🎨图像生成器😊情感分析器💡聊天机器人构建器⚖️伦理模拟器
进入实验室→
📝

博客

关于AI、教育和技术的最新文章

阅读博客→
nav.faq
🎯
使命

让AI教育触达每一个人、每一个角落

💜
价值观

开源、多语言、社区驱动

⭐
Open Source

在 GitHub 上公开构建

认识创始人→在 GitHub 上查看
立即开始
AI EducademyAIEducademy

MIT 许可证。开源项目

学习

  • 学习计划
  • 课程
  • 实验室

社区

  • GitHub
  • 参与贡献
  • 行为准则
  • 关于
  • 常见问题

支持

  • 请我喝杯咖啡 ☕
AI & 工程学习计划›🌳 AI 枝干›课程›AI 与机器人
🤖
AI 枝干 • 中级⏱️ 15 分钟阅读

AI 与机器人

AI and Robotics - From Factory Arms to Humanoid Helpers

Robots have worked in factories for decades - welding car frames, painting surfaces, and assembling electronics with tireless precision. But those classical robots followed rigid, pre-programmed instructions. They could not adapt if a bolt was slightly crooked or a box was in the wrong place. AI-powered robotics changes everything: robots that perceive, reason, and learn from their environment.

Classical Robotics vs AI-Powered Robotics

Traditional industrial robots operate through hard-coded instructions: move to position X, rotate 45 degrees, close gripper. They are fast, precise, and reliable - but only in controlled environments where nothing unexpected happens.

AI-powered robots, by contrast, use perception, planning, and learning to handle uncertainty. They can recognise objects they have never seen, plan paths through cluttered spaces, and improve their performance over time through experience.

| Aspect | Classical Robotics | AI-Powered Robotics | |---|---|---| | Programming | Hard-coded trajectories | Learned behaviours | | Environment | Structured, predictable | Unstructured, dynamic | | Adaptability | None | Continuous learning | | Sensing | Minimal (position encoders) | Rich (cameras, LiDAR, touch) |

Diagram showing the perception-planning-action loop in an AI-powered robot
AI robots continuously perceive their environment, plan actions, and execute - adapting in real time.

The Perception-Planning-Action Loop

Every AI robot operates through a continuous cycle:

  1. Perception - Cameras, LiDAR, depth sensors, and force sensors build a model of the surrounding environment. Computer vision identifies objects, estimates poses, and detects obstacles.
  2. Planning - Given the perceived world and a goal, the robot plans a sequence of actions. This might involve path planning (avoiding obstacles), grasp planning (how to pick up an object), or task planning (what steps to complete a job).
  3. Action - Motors and actuators execute the plan. Feedback from sensors continuously refines execution.

This loop runs many times per second. The faster and more accurate each stage, the more capable the robot becomes in dynamic environments. Modern robots often run perception at 30–60 frames per second and replanning at similar rates, enabling fluid, reactive behaviour.

Computer Vision in Robots

Vision is arguably the most critical sense for modern robots. Key capabilities include:

  • Object detection and recognition - Identifying items on a conveyor belt or shelf.
  • Pose estimation - Determining an object's exact position and orientation for precise grasping.
  • Semantic segmentation - Understanding the scene at pixel level (floor, obstacle, target object).
  • Depth perception - Stereo cameras or structured-light sensors provide 3D understanding.
\ud83e\udd2f
Amazon's warehouse robots process thousands of different product types daily, using computer vision to identify items they have never been explicitly trained on.

Boston Dynamics and Legged Locomotion

Boston Dynamics has pushed the boundaries of what robots can physically do. Their robots - Spot (quadruped), Atlas (humanoid), and Stretch (warehouse) - demonstrate remarkable agility.

Atlas can run, jump, perform backflips, and navigate uneven terrain. Spot patrols construction sites, inspects industrial facilities, and navigates stairs. The key challenge with legged locomotion is balance control: the robot must constantly adjust its posture in response to terrain, obstacles, and its own momentum.

Recent breakthroughs use reinforcement learning to train locomotion policies in simulation, then transfer them to physical robots - a technique called sim-to-real transfer. The results are striking: robots trained this way recover from pushes, navigate rubble, and traverse ice - all without explicit programming for each scenario.

\ud83e\udde0小测验

What is the main advantage of AI-powered robots over classical industrial robots?

Warehouse Robots - Amazon and Beyond

Amazon's fulfilment centres deploy over 750,000 robots working alongside human employees. These robots:

  • Transport shelving units across warehouse floors (Kiva/Proteus).
  • Sort packages using computer vision and robotic arms (Sparrow).
  • Navigate dynamic environments shared with human workers.

The economics are compelling: robots reduce order processing time, operate around the clock, and improve accuracy. But they also raise important questions about workforce displacement and the changing nature of warehouse work.

Surgical Robots - Precision Beyond Human Hands

The da Vinci Surgical System by Intuitive Surgical has performed over 12 million procedures worldwide. It does not operate autonomously - a surgeon controls it - but AI enhances the experience:

  • Tremor filtering eliminates natural hand tremors.
  • Motion scaling translates large hand movements into tiny, precise instrument movements.
  • 3D visualisation provides magnified, high-definition views of the surgical site.

Fully autonomous surgical robots remain a research frontier, but AI-assisted systems already enable procedures that would be impossible with human hands alone.

\ud83e\udd14
Think about it:Surgical robots can perform operations with superhuman precision, yet fully autonomous surgery remains years away. What technical and ethical hurdles must be cleared before a robot operates without a human surgeon in the loop?

Humanoid Robots - The Next Frontier

The race to build general-purpose humanoid robots is accelerating:

  • Tesla Optimus - Designed for repetitive manual labour in factories and potentially homes. Tesla leverages its self-driving AI expertise for Optimus's perception and planning.
  • Figure 01 and Figure 02 - Figure AI's humanoids combine large language models with physical manipulation, enabling robots that understand spoken instructions and execute complex tasks.
  • Agility Robotics Digit - A bipedal robot designed for logistics, already deployed in Amazon warehouses for testing.

The humanoid form factor is attractive because human environments - buildings, vehicles, tools - are designed for human bodies. A humanoid robot can theoretically use existing infrastructure without modification.

\ud83e\udde0小测验

Why is the humanoid form factor considered advantageous for general-purpose robots?

The Real World Is Messy

The biggest challenge in robotics is not any single algorithm - it is the gap between simulation and reality. The real world is unpredictable:

  • Lighting changes throughout the day.
  • Objects have unexpected shapes, weights, and textures.
  • Surfaces are slippery, uneven, or cluttered.
  • Humans behave unpredictably.

This is why robots that work flawlessly in a lab often struggle in a real kitchen or construction site. Robustness remains the field's central engineering challenge.

Sim-to-Real and Embodied AI

Sim-to-real transfer trains robots in simulated environments - which are cheap, fast, and safe - and then deploys them in the physical world. Techniques like domain randomisation (varying simulation parameters) help policies generalise.

Embodied AI goes further: it argues that true intelligence requires a physical body interacting with a physical world. Language models understand the word "heavy," but a robot that has struggled to lift a concrete block has a fundamentally different, richer understanding.

\ud83e\udd2f
NVIDIA's Isaac Sim platform can simulate millions of robotic interactions per hour, dramatically accelerating training that would take months in the real world.
\ud83e\udde0小测验

What is domain randomisation in sim-to-real transfer?

\ud83e\udd14
Think about it:As robots become more capable and autonomous, how should society handle the economic impact on jobs that are currently performed by humans?

📚 Further Reading

  • Boston Dynamics - Inside the Lab - Videos and technical overviews of Spot, Atlas, and Stretch
  • NVIDIA Isaac Sim - Simulation platform for training and testing AI-powered robots at scale
第 9 课,共 14 课已完成 0%
←多模态AI
自动驾驶汽车→