Datasets:
Create README.md (#1)
Browse files- Create README.md (4ee2105a7f167e68532c06e74b1d305822e5cc0b)
Co-authored-by: Manthan Patel <[email protected]>
README.md
ADDED
@@ -0,0 +1,92 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: en
|
3 |
+
license: cc-by-4.0
|
4 |
+
tags:
|
5 |
+
- robotics
|
6 |
+
- perception
|
7 |
+
- navigation
|
8 |
+
- SLAM
|
9 |
+
- semantic-segmentation
|
10 |
+
- occupancy-mapping
|
11 |
+
- synthetic-data
|
12 |
+
- multimodal
|
13 |
+
- ground-robot
|
14 |
+
- dataset
|
15 |
+
size_categories:
|
16 |
+
- 10M<n<100M
|
17 |
+
---
|
18 |
+
|
19 |
+
# TartanGround: A Large-Scale Dataset for Ground Robot Perception and Navigation
|
20 |
+
|
21 |
+

|
22 |
+
|
23 |
+
## Dataset Description
|
24 |
+
|
25 |
+
**TartanGround** is a large-scale, multi-modal dataset designed to advance the perception and autonomy of ground robots operating in diverse environments. Collected across 63 photorealistic simulation environments, it provides comprehensive data streams for various robotic tasks.
|
26 |
+
|
27 |
+
## Key Features
|
28 |
+
|
29 |
+
- **Environments**: 63 diverse simulation environments categorized into:
|
30 |
+
- Indoor
|
31 |
+
- Nature
|
32 |
+
- Rural
|
33 |
+
- Urban
|
34 |
+
- Industrial/Infrastructure
|
35 |
+
- Historical/Thematic
|
36 |
+
|
37 |
+
- **Trajectories**: 878 trajectories captured across the environments.
|
38 |
+
|
39 |
+
- **Samples**: Over 1.44 million samples.
|
40 |
+
|
41 |
+
- **Robot Platforms**:
|
42 |
+
- Omnidirectional (`P0000`, `P0001`, ...)
|
43 |
+
- Differential Drive (`P1000`, `P1001`, ...)
|
44 |
+
- Quadrupedal (`P2000`, `P2001`, ...)
|
45 |
+
|
46 |
+
- **Sensor Modalities**:
|
47 |
+
- RGB Stereo Camera Pairs (front, back, left, right, top, bottom)
|
48 |
+
- Depth Maps
|
49 |
+
- Semantic Segmentation
|
50 |
+
- Optical Flow
|
51 |
+
- Stereo Disparity
|
52 |
+
- LiDAR Point Clouds
|
53 |
+
- IMU Data
|
54 |
+
- Ground Truth Poses (6-DOF)
|
55 |
+
- Semantic Occupancy Maps (3D voxel grids)
|
56 |
+
- Proprioceptive Data (for quadruped trajectories)
|
57 |
+
|
58 |
+
## Applications
|
59 |
+
|
60 |
+
TartanGround supports a wide range of robotic perception and navigation tasks, including:
|
61 |
+
|
62 |
+
- Semantic Occupancy Prediction
|
63 |
+
- Open-Vocabulary Occupancy Prediction
|
64 |
+
- Visual SLAM
|
65 |
+
- Neural Scene Representation
|
66 |
+
- Bird's-eye-view Prediction
|
67 |
+
- Navigation and more
|
68 |
+
|
69 |
+
## License
|
70 |
+
|
71 |
+
The dataset is licensed under the [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
|
72 |
+
|
73 |
+
## Citation
|
74 |
+
|
75 |
+
If you use TartanGround in your research, please cite the following paper:
|
76 |
+
|
77 |
+
```
|
78 |
+
@article{patel2025tartanground,
|
79 |
+
title={TartanGround: A Large-Scale Dataset for Ground Robot Perception and Navigation},
|
80 |
+
author={Patel, Manthan and Yang, Fan and Qiu, Yuheng and Cadena, Cesar and Scherer, Sebastian and Hutter, Marco and Wang, Wenshan},
|
81 |
+
journal={arXiv preprint arXiv:2505.10696},
|
82 |
+
year={2025}
|
83 |
+
}
|
84 |
+
```
|
85 |
+
|
86 |
+
|
87 |
+
## Links
|
88 |
+
|
89 |
+
- [Dataset Website](https://tartanair.org/tartanground)
|
90 |
+
- [Paper on arXiv](https://arxiv.org/abs/2505.10696)
|
91 |
+
- [GitHub Repository](https://github.com/castacks/tartanairpy)
|
92 |
+
- [Metadata Sheet](https://docs.google.com/spreadsheets/d/1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0t1u2v3w4x5y6z7a8b9c0d1e2f3g4h5i6j7k8l9m0n1o2p3q4r5s6t7u8v9w0x1y2z3a4b5c6d7e8f9g0h1i2j3k4l5m6n7o8p9q0r1s2t3u4v5w6x7y8z9a0b1c2d3e4f5g6h7i8j9k0l1m2n3o4p5q6r7s8t9u0v1w2x3y4z5a6b7c8d9e0f1g2h3i4j5k6l7m8n9o0p1q2r3s4t5u6v7w8x9y0z1a2b3c4d5e6f7g8h9i0j1k2l3m4n5o6p7q8r9s0)
|