Spaces:
Running
on
Zero
Running
on
Zero
xinjie.wang
commited on
Commit
·
5c8b822
1
Parent(s):
732c3d7
update
Browse files
embodied_gen/models/layout.py
CHANGED
|
@@ -43,7 +43,7 @@ __all__ = [
|
|
| 43 |
]
|
| 44 |
|
| 45 |
|
| 46 |
-
DISTRACTOR_NUM =
|
| 47 |
LAYOUT_DISASSEMBLE_PROMPT = f"""
|
| 48 |
You are an intelligent 3D scene planner. Given a natural language
|
| 49 |
description of a robotic task, output a structured description of
|
|
|
|
| 43 |
]
|
| 44 |
|
| 45 |
|
| 46 |
+
DISTRACTOR_NUM = 2 # Maximum number of distractor objects allowed
|
| 47 |
LAYOUT_DISASSEMBLE_PROMPT = f"""
|
| 48 |
You are an intelligent 3D scene planner. Given a natural language
|
| 49 |
description of a robotic task, output a structured description of
|
embodied_gen/scripts/imageto3d.py
CHANGED
|
@@ -219,9 +219,10 @@ def entrypoint(**kwargs):
|
|
| 219 |
num_images=12,
|
| 220 |
)
|
| 221 |
|
| 222 |
-
|
| 223 |
-
|
| 224 |
-
)
|
|
|
|
| 225 |
logger.warning(
|
| 226 |
f"{GEO_CHECKER.__class__.__name__}: {geo_result} for {seg_path}"
|
| 227 |
)
|
|
|
|
| 219 |
num_images=12,
|
| 220 |
)
|
| 221 |
|
| 222 |
+
color_img = Image.open(color_path)
|
| 223 |
+
half_height = int(color_img.height * 2 / 3)
|
| 224 |
+
crop_img = color_img.crop((0, 0, color_img.width, half_height))
|
| 225 |
+
geo_flag, geo_result = GEO_CHECKER([crop_img], text=asset_node)
|
| 226 |
logger.warning(
|
| 227 |
f"{GEO_CHECKER.__class__.__name__}: {geo_result} for {seg_path}"
|
| 228 |
)
|
embodied_gen/scripts/render_gs.py
CHANGED
|
@@ -104,7 +104,7 @@ def load_gs_model(
|
|
| 104 |
# Normalize vertices to [-1, 1], center to (0, 0, 0).
|
| 105 |
_, scale, center = normalize_vertices_array(gs_model._means)
|
| 106 |
scale, center = float(scale), center.tolist()
|
| 107 |
-
transpose = [*[
|
| 108 |
instance_pose = torch.tensor(transpose).to(gs_model.device)
|
| 109 |
gs_model = gs_model.get_gaussians(instance_pose=instance_pose)
|
| 110 |
gs_model.rescale(scale)
|
|
|
|
| 104 |
# Normalize vertices to [-1, 1], center to (0, 0, 0).
|
| 105 |
_, scale, center = normalize_vertices_array(gs_model._means)
|
| 106 |
scale, center = float(scale), center.tolist()
|
| 107 |
+
transpose = [*[v for v in center], *pre_quat]
|
| 108 |
instance_pose = torch.tensor(transpose).to(gs_model.device)
|
| 109 |
gs_model = gs_model.get_gaussians(instance_pose=instance_pose)
|
| 110 |
gs_model.rescale(scale)
|
embodied_gen/validators/quality_checkers.py
CHANGED
|
@@ -249,7 +249,7 @@ class SemanticConsistChecker(BaseChecker):
|
|
| 249 |
fewer than four legs or if the legs are unevenly distributed, are not allowed. Do not assume
|
| 250 |
hidden legs unless they are clearly visible.)
|
| 251 |
- Geometric completeness is required: the object must not have missing, truncated, or cropped parts.
|
| 252 |
-
- The image must contain exactly one object. Multiple distinct objects are not allowed.
|
| 253 |
A single composite object (e.g., a chair with legs) is acceptable.
|
| 254 |
- The object should be shown from a slightly angled (three-quarter) perspective,
|
| 255 |
not a flat, front-facing view showing only one surface.
|
|
|
|
| 249 |
fewer than four legs or if the legs are unevenly distributed, are not allowed. Do not assume
|
| 250 |
hidden legs unless they are clearly visible.)
|
| 251 |
- Geometric completeness is required: the object must not have missing, truncated, or cropped parts.
|
| 252 |
+
- The image must contain exactly one object. Multiple distinct objects (e.g. multiple pens) are not allowed.
|
| 253 |
A single composite object (e.g., a chair with legs) is acceptable.
|
| 254 |
- The object should be shown from a slightly angled (three-quarter) perspective,
|
| 255 |
not a flat, front-facing view showing only one surface.
|