MTJester commited on
Commit
4130947
·
verified ·
1 Parent(s): 400731a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +182 -3
README.md CHANGED
@@ -1,3 +1,182 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ tags:
4
+ - 6d-pose-estimation
5
+ size_categories:
6
+ - 10K<n<100K
7
+ dataset_info:
8
+ features:
9
+ - name: scene_id
10
+ dtype: string
11
+ - name: image_id
12
+ dtype: string
13
+ - name: camera_type
14
+ dtype: string
15
+ - name: image
16
+ dtype: image
17
+ - name: depth
18
+ dtype: image
19
+ - name: width
20
+ dtype: int64
21
+ - name: height
22
+ dtype: int64
23
+ - name: split
24
+ dtype: string
25
+ - name: labels
26
+ dtype: string
27
+ - name: camera_params
28
+ dtype: string
29
+ - name: source_image_id
30
+ dtype: string
31
+ configs:
32
+ - config_name: default
33
+ data_files:
34
+ - split: test
35
+ path: data/*
36
+ pretty_name: CHIP
37
+ ---
38
+ # CHIP: A multi-sensor dataset for 6D pose estimation of chairs in industrial settings
39
+ <div style="display: flex; justify-content: center; gap: 1rem; flex-wrap: wrap; margin: 2rem 0;">
40
+
41
+ <a href="https://tev-fbk.github.io/CHIP/"
42
+ target="_blank"
43
+ style="
44
+ display: inline-block;
45
+ background-image: linear-gradient(to right, #3b82f6, #06b6d4);
46
+ color: white;
47
+ font-weight: bold;
48
+ font-size: 1rem;
49
+ padding: 0.85rem 1.75rem;
50
+ border-radius: 0.75rem;
51
+ text-decoration: none;
52
+ transition: opacity 0.3s;
53
+ "
54
+ onmouseover="this.style.opacity='0.85'"
55
+ onmouseout="this.style.opacity='1'">
56
+ 🏠 Homepage
57
+ </a>
58
+
59
+ <a href="https://arxiv.org/abs/2506.09699"
60
+ target="_blank"
61
+ style="
62
+ display: inline-block;
63
+ background-image: linear-gradient(to right, #34d399, #10b981);
64
+ color: white;
65
+ font-weight: bold;
66
+ font-size: 1rem;
67
+ padding: 0.85rem 1.75rem;
68
+ border-radius: 0.75rem;
69
+ text-decoration: none;
70
+ transition: opacity 0.3s;
71
+ "
72
+ onmouseover="this.style.opacity='0.85'"
73
+ onmouseout="this.style.opacity='1'">
74
+ 📄 Paper
75
+ </a>
76
+
77
+ </div>
78
+
79
+ ![CHIP Dataset Teaser](resources/teaser.png)
80
+
81
+ ## Introduction
82
+
83
+ Accurate 6D pose estimation of complex objects in 3D environments is essential for effective robotic manipulation. Yet, existing benchmarks fall short in evaluating 6D pose estimation methods under realistic industrial conditions, as most datasets focus on household objects in domestic settings, while the few available industrial datasets are limited to artificial setups with objects placed on tables. To bridge this gap, we introduce CHIP, the first dataset designed for 6D pose estimation of chairs manipulated by a robotic arm in a real-world industrial environment. CHIP includes seven distinct chairs captured using three different RGBD sensing technologies and presents unique challenges, such as distractor objects with fine-grained differences and severe occlusions caused by the robotic arm and human operators. CHIP comprises 77,811 RGBD images annotated with ground-truth 6D poses automatically derived from the robot's kinematics, averaging 11,115 annotations per chair. We benchmark CHIP using three zero-shot 6D pose estimation methods, assessing performance across different sensor types, localization priors, and occlusion levels. Results show substantial room for improvement, highlighting the unique challenges posed by the dataset.
84
+
85
+ ## Dataset Summary
86
+ - **Number of images:** 77,811 RGBD images
87
+ - **Number of object classes:** 7 distinct chair models
88
+ - **Sensors used:** Intel RealSense D435, Intel RealSense L515, Stereo Labs ZED
89
+ - **Annotations:** Ground-truth 6D poses derived from robot kinematics (~11,115 annotations per chair)
90
+ - **Occlusion levels:** No occlusions, moderate occlusions
91
+
92
+ ### Object Classes
93
+ CHIP includes seven high-quality chair models from Andreu World, covering a variety of structures and materials.
94
+
95
+ #### Frameonly designs:
96
+ - **000001**: Smile si0325 — [Andreu World link](https://andreuworld.com/en/products/smile-si0325)
97
+ - **000003**: Carlotta si0991 — [Andreu World link](https://andreuworld.com/en/products/carlotta-si0991)
98
+ - **000006**: Carola so0903 — [Andreu World link](https://andreuworld.com/en/products/carola-so0903)
99
+ - **000007**: Rizo so2043 — [Andreu World link](https://andreuworld.com/en/products/rizo-so2043)
100
+
101
+ #### Solid-wood designs:
102
+ - **000002**: Happy si0374 — [Andreu World link](https://andreuworld.com/en/products/happy-si0374)
103
+ - **000004**: Duos si2750 — [Andreu World link](https://andreuworld.com/en/products/duos-si2750)
104
+ - **000005**: Rdl si7291 — [Andreu World link](https://andreuworld.com/en/products/rdl-si7291)
105
+
106
+ ![Example Chair Models](resources/chairs.jpg)
107
+
108
+ ### Data Fields
109
+ ```
110
+ - scene_id: Unique identifier for each scene in the dataset (BOP format).
111
+ - image_id: Unique identifier for each image within a scene and camera type (BOP format).
112
+ - camera_type: Type of camera used to capture the image (e.g., 'zed', 'rs_l515', 'rs_d435').
113
+ - image: RGB image captured by the specified camera.
114
+ - depth: Depth image corresponding to the RGB image, captured by the specified camera.
115
+ - width: Width of the image in pixels.
116
+ - height: Height of the image in pixels.
117
+ - split: Dataset split to which the image belongs (e.g., 'test_no_occlusions', 'test_moderate_occlusions').
118
+ - source_image_id: Original image identifier from the CHIP dataset.
119
+ - labels: JSON string containing object annotations, including 6D poses and visibility information.
120
+ - camera_params: JSON string containing intrinsic and extrinsic camera parameters for the specified camera.
121
+ ```
122
+
123
+ ### Uses
124
+
125
+ The CHIP dataset can be used for a wide range of research tasks, including:
126
+
127
+ - Benchmarking 6D pose estimation in realistic industrial environments
128
+ - Evaluating robustness to challenging conditions, such as occlusions, clutter, and visually similar distractor objects
129
+ - Studying cross-sensor generalization across LiDAR-based, passive-stereo, and active-stereo RGB-D sensors
130
+ - 6D pose tracking during robotic manipulation, including motion sequences with dynamic occlusions
131
+
132
+ ## Citation
133
+ If you find CHIP useful for your work please cite:
134
+ ```
135
+ @inproceedings{nardon2025chip,
136
+ title={CHIP: A multi-sensor dataset for 6D pose estimation of chairs in industrial settings},
137
+ author={Nardon, Mattia and Mujika Agirre, Mikel and González Tomé, Ander and Sedano Algarabel, Daniel and Rueda Collell, Josep and Caro, Ana Paola and Caraffa, Andrea and Poiesi, Fabio and Chippendale, Paul Ian and Boscaini, Davide},
138
+ booktitle={British Machine Vision Conference (BMVC)},
139
+ year={2025}}
140
+ ```
141
+
142
+ ## Acknowledgement
143
+ <style>
144
+ .list_view{
145
+ display:flex;
146
+ align-items:center;
147
+ }
148
+ .list_view p{
149
+ padding:10px;
150
+ }
151
+ </style>
152
+
153
+ <div class="list_view">
154
+ <a href="https://aiprism.eu/" target="_blank">
155
+ <img src="resources/logos/Ai-Prism_Logo_Square.png" alt="Ai-Prism logo" style="max-width:200px">
156
+ </a>
157
+ <p>
158
+ This work was supported by the European Union's Horizon Europe research and innovation programme under grant agreement No. 101058589 (AI-PRISM).
159
+ </p>
160
+ </div>
161
+
162
+ ### Partners
163
+ <div style="background-color: #f8fafc; border: 1px solid #e2e8f0; border-radius: 1rem; padding: 2rem; max-width: 672px; margin: 2rem auto; text-align: center; font-family: sans-serif; color: #334155;">
164
+ <div style="display: flex; justify-content: center; align-items: center; gap: 30px; flex-wrap: wrap; margin-top: 1.5rem;">
165
+ <a href="https://www.fbk.eu/" target="_blank" style="opacity: 0.9; transition: opacity 0.3s;" onmouseover="this.style.opacity='1'" onmouseout="this.style.opacity='0.9'">
166
+ <img src="resources/logos/logo_fbk.png" alt="FBK logo" style="max-width:120px;">
167
+ </a>
168
+ <a href="https://www.andreuworld.com/en/" target="_blank" style="opacity: 0.9; transition: opacity 0.3s;" onmouseover="this.style.opacity='1'" onmouseout="this.style.opacity='0.9'">
169
+ <img src="resources/logos/Logo_Andreu_World.png" alt="Andreu World logo" style="max-width:120px;">
170
+ </a>
171
+ <a href="https://www.ikerlan.es/en" target="_blank" style="opacity: 0.9; transition: opacity 0.3s;" onmouseover="this.style.opacity='1'" onmouseout="this.style.opacity='0.9'">
172
+ <img src="resources/logos/Ikerlan_BRTA_V.png" alt="Ikerlan logo" style="max-width:120px;">
173
+ </a>
174
+ </div>
175
+ </div>
176
+
177
+ ### Contact
178
+ For any questions regarding the dataset, please contact Mattia Nardon ([email protected]).
179
+
180
+ Homepage link: https://tev-fbk.github.io/CHIP/
181
+
182
+ Paper link: https://arxiv.org/abs/2506.09699