2024-12-29 00:10:14 +08:00
< h1 align = "center" > 具身智能入门指南 Embodied-AI-Guide< / h1 >
< p align = "center" > < / p >
2024-12-31 12:16:23 +08:00
> Embodied AI( 具身智能) 入门的路径以及高质量信息的总结, 期望是按照路线走完后, 新手可以快速建立关于这个领域的认知, 希望能帮助到各位入门具身智能的朋友, 欢迎点Star、分享与提PR🌟~<br>【 <a href="https://github.com/tianxingchen/Embodied-AI-Guide">Embodied-AI-Guide</a>, Latest Update: Dec 29, 2024 】<img src="https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2Ftianxingchen%2FEmbodied-AI-Guide&count_bg=%232B8DD9&title_bg=%237834C6&icon=github.svg&icon_color=%23E7E7E7&title=Page+Viewers&edge_flat=false"/> <img alt="GitHub repo stars" src="https://img.shields.io/github/stars/TianxingChen/Embodied-AI-Guide">
2024-12-29 12:46:44 +08:00
2024-11-30 01:29:35 +08:00
## Contents - 目录
2024-12-31 22:09:15 +08:00
2024-06-09 19:01:52 +08:00
< nav >
< ul >
2024-12-31 22:09:15 +08:00
< li > < a href = "#start" > 1. Start Up - 从这里开始< / a > < / li >
< li > < a href = "#info" > 2. Useful Info - 有利于搭建认知的资料< / a > < / li >
< li > < a href = "#algorithm" > 3. Algorithm - 算法< / a >
2024-12-30 00:46:19 +08:00
< ul >
2024-12-31 22:09:15 +08:00
< li > < a href = "#common-tools" > 3.1 Common Tools - 常用工具< / a > < / li >
< li > < a href = "#foundation-models" > 3.2 Foundation Models - 基础模型< / a > < / li >
< li > < a href = "#robot-learning" > 3.3 Robot Learning - 机器人学习< / a >
< ul >
< li > < a href = "#rl" > 3.3.1 Reinforcement Learning - 强化学习< / a > < / li >
< li > < a href = "#il" > 3.3.2 Imitation Learning - 模仿学习< / a > < / li >
< / ul >
< / li >
< li > < a href = "#llm_robot" > 3.4 LLM for Robotics - 大模型在机器人学中的应用< / a > < / li >
< li > < a href = "#cv" > 3.5 Computer Vision - 计算机视觉< / a >
< ul >
< li > < a href = "#3dv" > 3.5.1 3D Vision - 三维视觉< / a > < / li >
< / ul >
< / li >
< li > < a href = "#embodied-ai-4-x" > 3.6 Embodied AI for X - 具身智能+X< / a >
< ul >
< li > Embodied AI for Healthcare - 具身智能+医疗< / li >
< / ul >
< / li >
2024-12-30 00:46:19 +08:00
< / ul >
< / li >
2024-12-31 22:09:15 +08:00
< li > < a href = "#hardware" > 4. Hardware - 硬件< / a >
2024-12-31 12:27:43 +08:00
< ul >
2024-12-31 22:09:15 +08:00
< li > < a href = "#control" > 4.1 Control - 控制学< / a > < / li >
2025-01-01 01:03:25 +08:00
< li > < a href = "#companies" > 4.2 Sensors - 传感器< / a > < / li >
< li > < a href = "#companies" > 4.3 Companies - 公司< / a > < / li >
2024-12-31 12:27:43 +08:00
< / ul >
< / li >
2024-12-31 22:09:15 +08:00
< li > < a href = "#software" > 5. Software - 软件< / a >
2024-12-29 00:10:14 +08:00
< ul >
2024-12-31 22:09:15 +08:00
< li > < a href = "#benchmarks" > 5.1 Benchmarks & Simulators - 基准 & 仿真器< / a > < / li >
2024-12-29 00:10:14 +08:00
< / ul >
< / li >
2024-12-31 22:09:15 +08:00
< li > < a href = "#paper_list" > 6. Paper Lists - 论文列表< / a > < / li >
< li > < a href = "#communities" > 7. Communities - 社区< / a > < / li >
< li > < a href = "#acknowledgement" > 9. Acknowledgement - 致谢< / a > < / li >
2024-06-09 19:01:52 +08:00
< / ul >
< / nav >
2024-12-31 22:09:15 +08:00
2024-06-09 19:01:52 +08:00
< section id = "start" > < / section >
2024-12-31 22:09:15 +08:00
## 1. Start Up - 从这里开始
2024-06-09 19:01:52 +08:00
2024-12-31 12:02:23 +08:00
> 具身智能是指一种基于物理身体进行感知和行动的智能系统,其通过智能体与环境的交互获取信息、理解问题、做出决策并实现行动,从而产生智能行为和适应性。
### How - 如何食用这份指南
我们希望的是帮助新人快速建立领域认知,所以设计理念是:**简要**介绍目前具身智能涉及到的主要技术,让大家知道不同的技术能够解决什么问题,未来想要深入发展的时候能够有头绪。
### About us - 关于我们
2024-12-29 00:10:14 +08:00
我们是一个由具身初学者组成的团队,希望能够通过我们自己的学习经验,为后来者提供一些帮助,加快具身智能的普及。欢迎更多朋友加入我们的项目,也很欢迎交友、学术合作,有任何问题,可以联系邮箱`chentianxing2002@gmail.com`。
2024-06-08 14:24:31 +08:00
2024-12-31 12:02:23 +08:00
< p > < b > 🦉Contributors< / b > : < a href = "https://tianxingchen.github.io" > 陈天行 (25' 港大PhD)< / a > , < a href = "https://yudezou.github.io/" > 邹誉德 (25' 上交-浦江实验室联培PhD)< / a > , < a href = "" > 陈思翔 (25' 北大PhD)< / a > , < a href = "https://github.com/27yw" > 叶雯 (25' 中科院自所PhD)< / a > , < a href = "https://github.com/zanxinchen" > 陈攒鑫 (深大本科生)< / a > , < a href = "https://github.com/ShijiaPeng03" > 彭时佳 (深大本科生)< / a > , < a href = "https://gkw0010.github.io/" > 王冠锟 (港中文-华为联培PhD)< / a > , < a href = "https://ngchikit.github.io" > 吴志杰 (港中文PhD)< / a > , < a href = "https://github.com/csyufei" > 朱宇飞 (25' 上科大Ms)< / a > .< / p >
2024-12-29 00:10:14 +08:00
< a href = "https://github.com/TianxingChen/Embodied-AI-Guide/graphs/contributors" >
< img src = "https://contrib.rocks/image?repo=TianxingChen/Embodied-AI-Guide" / >
< / a >
2024-12-27 02:33:01 +08:00
< section id = "info" > < / section >
2024-12-31 22:09:15 +08:00
## 2. Useful Info - 有利于搭建认知的资料
2024-12-29 00:10:14 +08:00
2024-12-27 02:40:35 +08:00
* 具身智能基础技术路线-YunlongDong [2]: [PDF ](./files/具身智能基础技术路线-YunlongDong.pdf ), [bilibili ](https://www.bilibili.com/video/BV1d5ukedEsi/?buvid=XXCD799C01878A6CFDECF3FB4427E2F070877&from_spmid=default-value&is_story_h5=false&mid=iWFclAyh36UYMh2G6ZcsDw%3D%3D&p=1&plat_id=114&share_from=ugc&share_medium=android&share_plat=android&share_session_id=9c0dccf5-ec0b-4369-8b89-ff1d848467ee&share_source=WEIXIN&share_tag=s_i&spmid=united.player-video-detail.0.0×tamp=1716466406&unique_k=Q0CaIUj&up_id=249218043 )
2024-12-27 02:33:01 +08:00
2024-12-31 23:29:16 +08:00
* 社交媒体:
* 可以关注的公众号: **石麻日记 (超高质量!!!)** , 机器之心, 新智元, 量子位, 具身知识库, 具身智能之心, 自动驾驶之心, 3D视觉工坊, 将门创投, RLCN强化学习研究, CVHub
* AI领域值得关注的博主列表 [3]: [zhihu ](https://zhuanlan.zhihu.com/p/682110383 )
2024-12-27 02:33:01 +08:00
2024-12-27 02:40:35 +08:00
* Robotics实验室总结 [4]: [zhihu_1 ](https://zhuanlan.zhihu.com/p/682671294?utm_psn=1782122763157188608 ), [zhihu_2 ](https://zhuanlan.zhihu.com/p/682692024?utm_psn=1782122945184796672 )
2024-12-27 02:33:01 +08:00
2024-12-31 23:22:01 +08:00
* 具身智能会投稿的较高质量会议与期刊: RSS, TRO, Science Robotics, IROS, ICRA, ICCV, ECCV, ICRA, AAAI, ICML, CVPR, NIPS, ICLR, IJRR, ACL等。
2024-12-31 22:09:15 +08:00
2025-01-01 01:03:25 +08:00
* 斯坦福机器人学导论:[website](https://www.bilibili.com/video/BV17T421k78T/?spm_id_from=333.337.search-card.all.click)
* 共建全网最全具身智能知识库 [6]: [website ](https://yv6uc1awtjc.feishu.cn/wiki/WPTzw9ON0ivIVrkLjVocNZh8nLf )
2024-12-31 22:09:15 +08:00
< section id = "algorithm" > < / section >
## 3. Algorithm - 算法
< section id = "common-tools" > < / section >
### 3.1 Common Tools - 常用工具
> 这个部分是关于具身中常用技巧的分享
* 点云降采样: [zhihu ](https://zhuanlan.zhihu.com/p/558683732?utm_campaign=shareopn&utm_medium=social&utm_psn=1772067996070236160&utm_source=wechat_session ), 包括随机降采样、均匀降采样、最远点降采样、法线空间降采样等, 需要了解清楚每一种降采样的优劣, 这个技巧的选择对于3D应用来说是至关重要的。
2024-12-31 12:02:23 +08:00
< section id = "foundation-models" > < / section >
2024-12-31 22:09:15 +08:00
### 3.2 Foundation Models - 基础模型
2024-12-31 12:02:23 +08:00
2024-12-31 12:27:43 +08:00
> 以下是部分具身智能中常用的基础模型, 计算机视觉中发展的非常好的工具可以直接赋能具身智能的下游应用。
2024-12-31 12:10:36 +08:00
* CLIP: [website ](https://github.com/openai/CLIP ), 来自OpenAI的研究, 最基本的应用是可以计算图像与语言描述的相似度, 中间层的视觉特征对各种下游应用非常有帮助。
2024-12-31 12:16:23 +08:00
* DINO: [DINO repo ](https://github.com/facebookresearch/dino ), [DINO-v2 repo ](https://github.com/facebookresearch/dinov2 ), 来自Meta的研究, 可以提供图像的高层视觉特征, 对corresponding之类的信息提取非常有帮助, 比如不同个体之间的鼻子都有类似的几何特征, 这个时候不同图像中关于不同鼻子的视觉特征值可能是近似的。
2024-12-31 12:02:23 +08:00
* SAM: [website ](https://segment-anything.com/ ), 来自Meta的研究, 可以基于提示点或者框, 对图像的物体进行分割。
* SAM2: [website ](https://ai.meta.com/sam2/ ), 来自Meta的研究, SAM的升级版, 可以在视频层面持续对物体进行分割追踪。
2024-12-31 12:16:23 +08:00
* Grounding-DINO: [repo ](https://github.com/IDEA-Research/GroundingDINO ), [在线尝试 ](https://deepdataspace.com/playground/grounding_dino ), **这个DINO与上面Meta的DINO没有关系** , 是一个由IDEA研究院( 做了很多不错开源项目的机构) 开发集成的图像目标检测的框架, 很多时候需要对目标物体进行检测的时候可以考虑使用。
* Grounded-SAM: [repo ](https://github.com/IDEA-Research/Grounded-SAM-2 ), 比Grounding-DINO多了一个分割功能, 也就是支持检测后分割, 也有很多下游应用, 具体可以翻一下README。
2024-12-31 12:02:23 +08:00
2024-12-31 12:16:23 +08:00
* FoundationPose: [website ](https://github.com/NVlabs/FoundationPose ), 来自Nvidia的研究, 物体姿态追踪模型。
2024-12-31 12:02:23 +08:00
2024-12-31 12:16:23 +08:00
* Stable Diffusion: [repo ](https://github.com/CompVis/stable-diffusion ), [website ](https://ommer-lab.com/research/latent-diffusion-models/ ), 22年的文生图模型, 现在虽然不是SOTA了, 但是依然可以作为不错的应用, 例如中间层特征支持下游应用、生成Goal Image (目标状态) 等等。
2024-12-31 12:02:23 +08:00
2024-12-31 22:13:04 +08:00
* Depth Anything (v1 & v2): [repo ](https://github.com/LiheYoung/Depth-Anything ), [repo ](https://github.com/DepthAnything/Depth-Anything-V2 ), 港大和字节的研究工作,单目深度估计模型。
2024-12-31 13:33:46 +08:00
2024-12-31 22:13:04 +08:00
* Point Transformer (v3): [repo ](https://github.com/Pointcept/PointTransformerV3 ), 点云特征提取的工作。
* RDT-1B: [website ](https://rdt-robotics.github.io/rdt-robotics/ ), 清华朱军老师团队的工作, 机器人双臂操作的基础模型, 具有强大的few-shot能力。
2024-12-31 12:02:23 +08:00
2024-12-30 00:46:19 +08:00
< section id = "robot-learning" > < / section >
2024-12-31 22:09:15 +08:00
### 3.3 Robot Learning - 机器人学习
2024-12-27 02:33:01 +08:00
2024-12-31 12:02:23 +08:00
机器人学习 Robot Learning 的发展: [zhihu ](https://zhuanlan.zhihu.com/p/26988866 )
2024-06-09 19:01:52 +08:00
< section id = "rl" > < / section >
2024-12-31 22:09:15 +08:00
#### 3.3.1 Reinforcement Learning - 强化学习
2024-12-31 12:27:43 +08:00
* 推荐直接跟着李宏毅老师一套走: bilibili上课+刷蘑菇书巩固+gymnasium动手实践, 重点了解一下PPO。
* 台湾大学李宏毅公开课: [bilibili ](https://www.bilibili.com/video/BV1XP4y1d7Bk/?spm_id_from=333.337.search-card.all.click&vd_source=ab9cf5374617c2867aaea34af29b53c9 )< br >
* EasyRL - 蘑菇书: [website ](https://datawhalechina.github.io/easy-rl/#/ ), 基本是配套李宏毅老师的课程< br >
* 实践[gymnasium](https://gymnasium.farama.org/),可以尝试一下把玩一下登月着陆等经典强化学习场景,思考+动手, 观察阶段agent的表现并分析, 有助于深入理解强化学习
<!-- * UCB CS285 深度强化学习: [website ](https://rail.eecs.berkeley.edu/deeprlcourse/ ) | [youtube ](https://www.youtube.com/playlist?list=PL_iWQOsE6TfVYGEGiAOMaOzzv41Jfm_Ps )<br> -->
<!-- * 强化学习的数学原理 - 西湖大学赵世钰: [bilibili ](https://space.bilibili.com/2044042934/channel/collectiondetail?sid=748665 )<br> -->
2024-06-08 18:14:11 +08:00
2024-06-09 19:01:52 +08:00
< section id = "il" > < / section >
2024-12-31 22:09:15 +08:00
#### 3.3.2 Imitation Learning - 模仿学习
2024-12-31 12:27:43 +08:00
2024-12-27 02:43:37 +08:00
* 模仿学习简洁教程 - 南京大学LAMDA: [PDF ](https://www.lamda.nju.edu.cn/xut/Imitation_Learning.pdf )< br >
* Supervised Policy Learning for Real Robots, RSS 2024 Workshop 教程:真实机器人的监督策略学习, [bilibili ](https://www.bilibili.com/video/BV1Fx4y1s7if/?buvid=XY415384A771A6C681C9BEB3817566ED57724&is_story_h5=false&mid=ORgXkVzTHaOKTsml0RX5Gw%3D%3D&plat_id=240&share_from=ugc&share_medium=android&share_plat=android&share_source=WEIXIN&share_tag=s_i&spmid=dt.space-dt.0.0×tamp=1721464513&unique_k=Cqj5d9J&up_id=2185804&vd_source=ab9cf5374617c2867aaea34af29b53c9 )
2024-06-08 18:15:49 +08:00
2024-12-31 22:09:15 +08:00
<!-- * 实践[RoboTwin]() -->
2024-12-31 12:27:43 +08:00
2024-06-09 19:01:52 +08:00
< section id = "llm_robot" > < / section >
2024-12-31 22:09:15 +08:00
### 3.4 LLM for Robotics - 大模型在机器人学中的应用
2024-12-27 02:43:37 +08:00
* Robotics+LLM系列通过大语言模型控制机器人 [2]: [zhihu ](https://zhuanlan.zhihu.com/p/668053911 )< br >
* PDDL-wiki: [website ](https://planning.wiki/ )< br >
* An Introduction to PDDL: [PDF ](https://www.cs.toronto.edu/~sheila/2542/s14/A1/introtopddl2.pdf )< br >
* Embodied Agent wiki: [website ](https://en.wikipedia.org/wiki/Embodied_agent )< br >
* Lilian Weng 个人博客 - AI Agent 系统综述 [5]: 中文: [website ](https://mp.weixin.qq.com/s/Jb8HBbaKYXXxTSQOBsP5Wg ) 英文: [website ](https://lilianweng.github.io/posts/2023-06-23-agent/ )< br >
2024-09-01 15:39:03 +08:00
2024-12-28 22:25:37 +08:00
< section id = "medical" > < / section >
## MLLM for Medical - 多模态大语言模型在医学中的应用
* SkinGPT-4 for dermatological diagnosis: [website ](https://www.nature.com/articles/s41467-024-50043-3 )< br >
2024-12-28 22:30:27 +08:00
* PneumoLLM for pneumoconiosis diagnosis: [website ](https://www.sciencedirect.com/science/article/abs/pii/S1361841524001737 )< br >
* BiomedGPT: [website ](https://github.com/taokz/BiomedGPT )< br >
* LLAVA-Med: [website ](https://github.com/microsoft/LLaVA-Med?tab=readme-ov-file )< br >
2024-12-28 22:25:37 +08:00
2024-06-09 19:01:52 +08:00
2024-12-31 12:27:43 +08:00
< section id = "cv" > < / section >
2024-12-31 22:09:15 +08:00
### 3.5 Computer Vision - 计算机视觉
2024-12-31 12:27:43 +08:00
2024-12-31 22:19:54 +08:00
CS231n (斯坦福计算机视觉课程): [website ](https://cs231n.stanford.edu/schedule.html ), 该课程对深度学习在计算机视觉的应用有较为全面的介绍。因为已经在具体实现某个论文的算法了,所以这个阶段可以不用做作业,只需要看课程视频和课程讲义即可。< br >
2024-08-19 20:56:57 +08:00
2024-12-31 12:27:43 +08:00
< section id = "3dv" > < / section >
2024-12-31 22:09:15 +08:00
#### 3.5.1 3D Vision - 三维视觉
2024-12-31 13:50:56 +08:00
第一阶段: 学习最基础的3DV知识, 追求广度, 了解一些基础的概念和算法< br >
2024-12-31 13:24:44 +08:00
* 三维视觉导论 - Andreas Geiger: [website ](https://uni-tuebingen.de/fakultaeten/mathematisch-naturwissenschaftliche-fakultaet/fachbereiche/informatik/lehrstuehle/autonomous-vision/lectures/computer-vision/ ) (重点是完成课程里面的作业) < br >
2024-12-31 13:21:56 +08:00
* GAMES203 - 三维重建和理解: [bilibili ](https://www.bilibili.com/video/BV1pw411d7aS/?share_source=copy_web&vd_source=0b7603f37af6d369a97df34525b149be )< br >
2024-12-31 13:50:56 +08:00
第二阶段:细分方向,追求深度,上手一些项目< br >
2024-12-31 22:19:12 +08:00
* 如果对传统图形学感兴趣,可以看下面两门(闫令琪老师开的课,讲得特别好):< br >
* GAMES101 - 现代计算机图形学入门: [website ](https://games-cn.org/intro-graphics/ )< br >
* GAMES202 - 高质量实时渲染: [website ](https://sites.cs.ucsb.edu/~lingqi/teaching/games202.html )< br >
* 如果对motion synthesis/computer animation感兴趣, 可以看:
* GAMES105 - 计算机角色动画基础: [website ](https://games-105.github.io/ )< br >
* 如何对三维重建感兴趣,可以看下面两门:
* Nerf原理代码讲解: [bilibili ](https://www.bilibili.com/video/BV1CC411V7oq/?spm_id_from=333.337.search-card.all.click&vd_source=930ef08bfb2ff0db87ec20bf72a99855 )
* 3DGS原理代码讲解: [bilibili ](https://www.bilibili.com/video/BV1zi421v7Dr?spm_id_from=333.788.recommend_more_video.-1&vd_source=930ef08bfb2ff0db87ec20bf72a99855 )
* 三维预训练最新综述:
* Advances in 3D pre-training and downstream tasks: a survey. [PDF ](https://link.springer.com/content/pdf/10.1007/s44336-024-00007-4.pdf )< br >
* 3DGS在具身上的综述:
* 3D Gaussian Splatting in Robotics: A Survey. [PDF ](https://arxiv.org/pdf/2410.12262v2 )< br >
2024-12-31 12:27:43 +08:00
2024-12-31 22:09:15 +08:00
< section id = "embodied-ai-4-x" > < / section >
### 3.6 Embodied AI for X - 具身智能+X
#### 3.6.1 Embodied AI for Healthcare - 具身智能+医疗
Coming Soon...
< section id = "hardware" > < / section >
## 4. Hardware - 硬件
2024-09-02 15:29:29 +08:00
< section id = "control" > < / section >
2024-12-31 22:09:15 +08:00
### 4.1 Control - 控制学
2024-12-31 12:10:36 +08:00
2024-12-27 02:43:37 +08:00
> 关于控制部分的学习,最好从实践出发!
2024-12-31 12:10:36 +08:00
2024-12-27 02:43:37 +08:00
* PID控制: [CSDN](https://blog.csdn.net/name_longming/article/details/115093338)
* 彻底搞懂阻抗控制、导纳控制、力位混合控制: [CSDN ](https://blog.csdn.net/a735148617/article/details/108564836 )< br >
* 具身智能ROS1基础: [website ](http://www.autolabor.com.cn/book/ROSTutorials/ )< br >
* 具身智能ROS2基础: [website ](https://zhangzhiwei-zzw.github.io/ROS2%E5%AD%A6%E4%B9%A0/ROS2/ )< br >
* 机器人系统教材: [website ](https://motion.cs.illinois.edu/RoboticSystems/ )< br >
* 动手实践Lerobot SO-100: [website](https://github.com/huggingface/lerobot/blob/main/examples/10_use_so100.md)< br >
* ROS多传感器时间戳同步: [website](https://blog.csdn.net/qq_43495930/article/details/125649446)
2025-01-01 01:03:25 +08:00
< section id = "sensors" > < / section >
### 4.2 Sensors - 传感器
Coming Soon !
< section id = "companies" > < / section >
### 4.3 Companies - 公司
| 公司 | 主营产品 | Others |
|-------|------|------|
| [松灵AgileX ](https://www.agilex.ai/ ) | [pipper机械臂 ](https://www.agilex.ai/chassis/16 )< br > 移动底盘 | 面向教育科研
| [宇树Unitree ](https://www.unitree.com/cn ) | [Go2机器狗 ](https://www.unitree.com/cn/go2 )< br > [通用人形H1](https://www.unitree.com/cn/h1)< br > [通用人形G1](https://www.unitree.com/cn/g1)< br > | 许多产出使用宇树的机器人作为硬件基础
| [方舟无限ARX ](https://www.arx-x.com/?product/ ) | [X5机械臂 ](https://www.arx-x.com/?product/21.html )< br > [X7双臂平台](https://www.arx-x.com/?product/23.html)< br > [R5机械臂](https://www.arx-x.com/?product/22.html) | 适合复现很多经典的工作, eg. [aloha ](https://mobile-aloha.github.io/cn.html )< br > [RoboTwin松灵底盘+方舟臂](https://github.com/TianxingChen/RoboTwi)
| [波士顿动力 ](https://bostondynamics.com/ ) | [spot机器狗 ](https://bostondynamics.com/products/spot/ )< br > [Atlas通用人形](https://bostondynamics.com/atlas/) | 具身智能本体制造商,从液压驱动转向电机驱动 |
| [灵心巧手]| | |
| [灵巧智能DexRobot ](https://www.dex-robot.com/ )| [Dexhand 021灵巧手 ](https://www.dex-robot.com/productionDexhand ) | 19自由度量产灵巧手 |
| [银河通用 ](https://www.galbot.com/about ) | | 已完成多轮融资 |
| [星海图Galaxea ](http://galaxea.tech/ ) | [A1机械臂 ](http://galaxea.tech/Introducing_Galaxea_Robot/product_info/A1/#discover-more ) | |
| [World Labs ](https://www.worldlabs.ai/ ) | | 专注于空间智能, 致力于打造大型世界模型( LWM) , 以感知、生成并与 3D 世界进行交互。 [相关介绍 ](https://mp.weixin.qq.com/mp/wappoc_appmsgcaptcha?poc_token=HEH5X2ejkAoWy1ZXj8DlZO_Y2Q7PsYX-3ID-rfr5&target_url=https%3A%2F%2Fmp.weixin.qq.com%2Fs%2Fi58_yTFtt904haKezJgr1Q ) |
| [星动纪元 ](https://www.robotera.com ) | [Star1人形 ](https://www.robotera.com/goods/1.html )< br > [XHAND1灵巧手 ](https://www.robotera.com/goods/2.html ) | 由清华叉院陈建宇教授创建 |
| [加速进化 ](https://boosterobotics.com/zh/ ) | [Booster T1人形 ](https://boosterobotics.com/zh/store/ )| |
2024-12-31 22:09:15 +08:00
< section id = "software" > < / section >
## 5. Software - 软件
2024-12-11 02:00:42 +08:00
2024-06-09 19:01:52 +08:00
< section id = "benchmarks" > < / section >
2024-12-31 22:09:15 +08:00
### 5.1 Benchmarks & Simulators - 基准 & 仿真器
2024-11-29 21:42:08 +08:00
具身智能常用benchmark总结 [1]: [zhihu ](https://zhuanlan.zhihu.com/p/695342864 )< br >
2024-11-21 22:42:34 +08:00
常见仿真器wiki: [wiki ](https://simulately.wiki/ )
2024-11-29 21:42:08 +08:00
| 仿真器 | 基准 |
|-------|------|
| [IsaacSim ](https://developer.nvidia.com/isaac/sim ) | [BEHAVIOR-1K(可跨平台) ](https://behavior.stanford.edu/behavior-1k )+[omniGibson(工具链)](https://behavior.stanford.edu/omnigibson/)< br > [ARNOID](https://arnold-benchmark.github.io/) |
| [MuJoCo ](https://mujoco.org/ ) | [robosuite ](https://robosuite.ai/docs/overview.html )+[robomimic(工具链)](https://robomimic.github.io/)< br > [LIBERO](https://libero-project.github.io/main.html)< br > [MetaWorld](https://meta-world.github.io/)< br > [Gymnasium-Robotics(Fetch; Shadow Dexterous Hand; Maze; Adroit Hand; Franka Kitchen; MaMuJoCo)](https://robotics.farama.org/)< br > [RoboCasa](Docs.qq.com/sheet/DYmppSU55cFNpaVJo?tab=BB08J2)< br > [RoboHive](https://github.com/vikashplus/robohive) |
| [Sapien ](https://sapien.ucsd.edu/ ) | [ManiSkill ](https://maniskill.readthedocs.io/en/latest/index.html )< br > [RoboTwin](https://github.com/TianxingChen/RoboTwin) |
| [CoppeliaSim ](https://www.coppeliarobotics.com/ ) | [RLBench ](https://github.com/stepjam/RLBench )< br > [PerAct2](https://bimanual.github.io/)< br > [COLOSSEUM](https://robot-colosseum.github.io/) |
| [PyBullet ](https://pybullet.org/wordpress/ ) | [Calvin ](https://github.com/mees/calvin?tab=readme-ov-file )< br > [Ravens](https://github.com/google-research/ravens)< br > [VimaBench](https://github.com/vimalabs/VimaBench) |
2024-12-29 12:59:41 +08:00
| [Genesis ](https://genesis-embodied-ai.github.io/ ) ||
2024-06-09 19:01:52 +08:00
2024-12-31 12:16:23 +08:00
< section id = "paper_list" > < / section >
2024-12-31 22:09:15 +08:00
## 6. Paper Lists - 论文列表
2024-12-31 12:16:23 +08:00
* Awesome Humanoid Robot Learning - Yanjie Ze: [repo ](https://github.com/YanjieZe/awesome-humanoid-robot-learning )
* Paper Reading List - DeepTimber Community: [repo ](https://github.com/DeepTimber-Robot-Lab/Paper-Reading-List )
* Paper List - Yanjie Ze: [repo ](https://github.com/YanjieZe/Paper-List )
* Paper List For EmbodiedAI - Tianxing Chen: [repo ](https://github.com/TianxingChen/Paper-List-For-EmbodiedAI )
* SOTA Paper Rating - Weiyang Jin: [website ](https://waynejin0918.github.io/SOTA-paper-rating.io/ )
* Awesome-LLM-Robotics: A repo contains a curative list of papers using Large Language/Multi-Modal Models for Robotics/RL: [website ](https://github.com/GT-RIPL/Awesome-LLM-Robotics )
2024-12-29 00:10:14 +08:00
2024-06-09 19:01:52 +08:00
< section id = "communities" > < / section >
2024-12-31 22:09:15 +08:00
## 7. Communities - 社区
2024-12-27 02:32:28 +08:00
> 以下部分资料引用自[7]
2024-12-27 02:43:37 +08:00
* DeepTimber Robotics Innovations Community, 深木科研交流社区: [website ](https://gamma.app/public/DeepTimber-Robotics-Innovations-Community-A-Community-for-Multi-m-og0uv8mswl1a3q7?mode=doc )
2024-06-08 21:52:32 +08:00
2024-12-27 02:43:37 +08:00
* 宇树具身智能社群: [website ](https://www.unifolm.com/#/ )
2024-12-27 02:30:47 +08:00
2024-12-27 02:43:37 +08:00
* Simulately: Handy information and resources for physics simulators for robot learning research: [website ](https://simulately.wiki/ )
2024-12-27 02:30:47 +08:00
2024-12-27 02:43:37 +08:00
* DeepTimber-地瓜机器人社区: [website ](https://cn.developer.d-robotics.cc/forumList?id=156&title=Deeptimber )
2024-12-27 02:30:47 +08:00
2024-12-27 02:43:37 +08:00
* HuggingFace LeRobot (Europe, check the Discord): [website ](https://github.com/huggingface/lerobot )
2024-12-27 02:30:47 +08:00
2024-12-27 02:43:37 +08:00
* K-scale labs (US, check the Discord): [website ](https://kscale.dev/ )
2024-06-09 19:01:52 +08:00
2024-12-27 02:30:47 +08:00
< section id = "acknowledgement" > < / section >
< a name = "acknowledgement" > < / a >
2024-12-31 22:09:15 +08:00
## 9. Acknowledgement - 致谢
2024-12-27 11:50:12 +08:00
本文转载/引用了一些博主的文章,我们对他们的知识分享表示感谢,引用列表如下:
2024-12-31 14:01:12 +08:00
[1] 知乎 [穆尧 ](https://www.zhihu.com/people/mu-yao-12-34 ), [2] 知乎 [东林钟声 ](https://www.zhihu.com/people/dong-lin-zhong-sheng-76 ), Github [Yunlong Dong ](https://github.com/yunlongdong ), [3] 知乎 [强化学徒 ](https://www.zhihu.com/people/heda-he-28 ), [4] 知乎 [Biang哥 ](https://www.zhihu.com/people/qi-da-guang ), [5] OpenAI [Lilian Weng ](https://lilianweng.github.io/ ), [6] B站 [木木具身 ](https://space.bilibili.com/350563565 ), [7] Github [Zhuoheng Li ](https://github.com/StarCycle/EmbodiedAI-Reading-List-For-Lists?tab=readme-ov-file ), [8] 知乎 [Flood Sung ](https://www.zhihu.com/people/flood-sung ), [9] Github [Sida Peng ](https://github.com/pengsida/learning_research )
2024-07-08 18:11:40 +08:00
2024-11-30 01:29:35 +08:00
## 🏷️ License - 许可证
2024-07-08 18:11:40 +08:00
This repository is released under the MIT license. See LICENSE for additional details.
2024-12-29 00:10:14 +08:00