xinjie.wang commited on
Commit
5c8b822
·
1 Parent(s): 732c3d7
embodied_gen/models/layout.py CHANGED
@@ -43,7 +43,7 @@ __all__ = [
43
  ]
44
 
45
 
46
- DISTRACTOR_NUM = 3 # Maximum number of distractor objects allowed
47
  LAYOUT_DISASSEMBLE_PROMPT = f"""
48
  You are an intelligent 3D scene planner. Given a natural language
49
  description of a robotic task, output a structured description of
 
43
  ]
44
 
45
 
46
+ DISTRACTOR_NUM = 2 # Maximum number of distractor objects allowed
47
  LAYOUT_DISASSEMBLE_PROMPT = f"""
48
  You are an intelligent 3D scene planner. Given a natural language
49
  description of a robotic task, output a structured description of
embodied_gen/scripts/imageto3d.py CHANGED
@@ -219,9 +219,10 @@ def entrypoint(**kwargs):
219
  num_images=12,
220
  )
221
 
222
- geo_flag, geo_result = GEO_CHECKER(
223
- [color_path], text=asset_node
224
- )
 
225
  logger.warning(
226
  f"{GEO_CHECKER.__class__.__name__}: {geo_result} for {seg_path}"
227
  )
 
219
  num_images=12,
220
  )
221
 
222
+ color_img = Image.open(color_path)
223
+ half_height = int(color_img.height * 2 / 3)
224
+ crop_img = color_img.crop((0, 0, color_img.width, half_height))
225
+ geo_flag, geo_result = GEO_CHECKER([crop_img], text=asset_node)
226
  logger.warning(
227
  f"{GEO_CHECKER.__class__.__name__}: {geo_result} for {seg_path}"
228
  )
embodied_gen/scripts/render_gs.py CHANGED
@@ -104,7 +104,7 @@ def load_gs_model(
104
  # Normalize vertices to [-1, 1], center to (0, 0, 0).
105
  _, scale, center = normalize_vertices_array(gs_model._means)
106
  scale, center = float(scale), center.tolist()
107
- transpose = [*[-v for v in center], *pre_quat]
108
  instance_pose = torch.tensor(transpose).to(gs_model.device)
109
  gs_model = gs_model.get_gaussians(instance_pose=instance_pose)
110
  gs_model.rescale(scale)
 
104
  # Normalize vertices to [-1, 1], center to (0, 0, 0).
105
  _, scale, center = normalize_vertices_array(gs_model._means)
106
  scale, center = float(scale), center.tolist()
107
+ transpose = [*[v for v in center], *pre_quat]
108
  instance_pose = torch.tensor(transpose).to(gs_model.device)
109
  gs_model = gs_model.get_gaussians(instance_pose=instance_pose)
110
  gs_model.rescale(scale)
embodied_gen/validators/quality_checkers.py CHANGED
@@ -249,7 +249,7 @@ class SemanticConsistChecker(BaseChecker):
249
  fewer than four legs or if the legs are unevenly distributed, are not allowed. Do not assume
250
  hidden legs unless they are clearly visible.)
251
  - Geometric completeness is required: the object must not have missing, truncated, or cropped parts.
252
- - The image must contain exactly one object. Multiple distinct objects are not allowed.
253
  A single composite object (e.g., a chair with legs) is acceptable.
254
  - The object should be shown from a slightly angled (three-quarter) perspective,
255
  not a flat, front-facing view showing only one surface.
 
249
  fewer than four legs or if the legs are unevenly distributed, are not allowed. Do not assume
250
  hidden legs unless they are clearly visible.)
251
  - Geometric completeness is required: the object must not have missing, truncated, or cropped parts.
252
+ - The image must contain exactly one object. Multiple distinct objects (e.g. multiple pens) are not allowed.
253
  A single composite object (e.g., a chair with legs) is acceptable.
254
  - The object should be shown from a slightly angled (three-quarter) perspective,
255
  not a flat, front-facing view showing only one surface.