openpi / scripts /lerobot_preprocess_cotracker.py
zhicao's picture
Upload folder using huggingface_hub
e4eb88a verified
#!/usr/bin/env python3
"""
Direct RLDS-to-LeRobot preprocessing with CoTracker and simulator mesh seeds.
Pipeline (single stage):
1) Load LIBERO RLDS-as-LeRobot dataset (images/states/actions/tasks).
2) Derive scene-specific 7 mesh vertex seeds (agentview + eye-in-hand) from simulator.
3) Build query sets:
- agentview: 7 mesh seeds
- wrist: 25 uniform grid + 7 mesh seeds
4) Track queries with CoTracker over the episode frames (agent/wrist separately).
5) Emit LeRobot dataset with:
- images, wrist_images, state, actions, task
- per-frame tracks/vis (agentview_tracks/vis, wrist_tracks/vis)
- mesh vertices (agentview_mesh_vertices_2d, wrist_mesh_vertices_2d)
- grid query points (agentview_grid_mesh_32, wrist_grid_mesh_32)
- track_targets_raw: horizon window (H, 7+32, 2) combining agent mesh 7 + wrist grid25+mesh7
This bypasses the intermediate ATM conversion and writes directly to a LeRobot repo.
Notes:
- Requires a working LIBERO simulator + OffScreenRenderEnv to get mesh vertex seeds.
- CoTracker2 checkpoint is loaded locally (default: /mnt/kevin/vlm_models/cotracker/scaled_offline.pth).
- Designed for 256×256 frame resolution to match existing LeRobot LIBERO data.
"""
from __future__ import annotations
import openpi.shared.local_cache_bootstrap # noqa: F401
import argparse
import json
import os
import importlib
import re
import sys
from dataclasses import dataclass
from pathlib import Path
from typing import Iterable
import numpy as np
import pyarrow.parquet as pq
import torch
from einops import rearrange
from importlib.machinery import SourceFileLoader
from lerobot.common.datasets.lerobot_dataset import LeRobotDataset
# Scene-specific vertex IDs (7 per camera)
DEFAULT_VERTEX_IDS = {
"agentview": [3001, 3004, 3014, 3067, 3040, 3081, 2898],
"eye_in_hand": [2955, 3055, 3040, 2981, 2966, 3052, 3022],
}
KITCHEN_VERTEX_IDS = {
"agentview": [3249, 3252, 3200, 3315, 3272, 3329, 3134],
"eye_in_hand": [2955, 3055, 3040, 2981, 2966, 3052, 3022],
}
LIVING_ROOM_VERTEX_IDS = {
"agentview": [3135, 3140, 3141, 3202, 3205, 3215, 3034],
"eye_in_hand": [2955, 3055, 3040, 2981, 2966, 3052, 3022],
}
STUDY_VERTEX_IDS = {
"agentview": [3352, 3355, 3365, 3418, 3391, 3432, 3249],
"eye_in_hand": [2955, 3055, 3040, 2981, 2966, 3052, 3022],
}
SCENE_VERTEX_IDS = {
"default": DEFAULT_VERTEX_IDS,
"kitchen": KITCHEN_VERTEX_IDS,
"living_room": LIVING_ROOM_VERTEX_IDS,
"living": LIVING_ROOM_VERTEX_IDS,
"study": STUDY_VERTEX_IDS,
}
def _scene_from_task(task: str) -> str:
t = task.lower()
if "kitchen" in t:
return "kitchen"
if "living" in t or "livingroom" in t:
return "living_room"
if "study" in t:
return "study"
return "default"
def _make_grid_25() -> np.ndarray:
u = np.linspace(0.1, 0.9, 5, dtype=np.float32)
v = np.linspace(0.1, 0.9, 5, dtype=np.float32)
uu, vv = np.meshgrid(u, v, indexing="ij")
return np.stack([uu.reshape(-1), vv.reshape(-1)], axis=-1) # (25, 2)
EXTRA_LIBERO_PATH: str | None = None
def _load_cotracker(checkpoint: str, device: torch.device):
hub_dir = Path(os.environ.get("TORCH_HUB_DIR", Path.home() / ".cache" / "torch" / "hub"))
torch.hub.set_dir(str(hub_dir))
local_repo = hub_dir / "facebookresearch_co-tracker_main"
if not local_repo.exists():
raise FileNotFoundError(f"Local CoTracker repo not found at {local_repo}. Clone it or set TORCH_HUB_DIR accordingly.")
model = torch.hub.load(str(local_repo), "cotracker2", checkpoint=checkpoint, source="local")
return model.eval().to(device)
def _track_queries(video: np.ndarray, queries_xy01: np.ndarray, model, device: torch.device) -> tuple[np.ndarray, np.ndarray]:
"""
video: [T, H, W, 3] uint8
queries_xy01: [N, 2] normalized [0,1]
"""
T, H, W, _ = video.shape
vid = rearrange(torch.from_numpy(video).float().to(device) / 255.0, "t h w c -> 1 t c h w")
q = torch.from_numpy(queries_xy01.copy()).float().to(device)
# CoTracker expects pixel coords with leading time dim; start at frame 0.
q_pix = q * torch.tensor([W, H], device=device).float()
q_with_t = torch.cat([torch.zeros_like(q_pix[:, :1]), q_pix], dim=1) # [N,3] (t,x,y)
pred_tracks, pred_vis = model(vid, queries=q_with_t[None], backward_tracking=True)
tracks = pred_tracks[0].detach().cpu().numpy() # [T, N, 2] in px
vis = pred_vis[0].detach().cpu().numpy() # [T, N]
tracks[:, :, 0] /= float(W)
tracks[:, :, 1] /= float(H)
return tracks, vis
def _track_queries_from_start(
video: np.ndarray,
queries_xy01: np.ndarray,
start_t: int,
model,
device: torch.device,
) -> tuple[np.ndarray, np.ndarray]:
"""
Track queries starting from an arbitrary frame `start_t`.
Returns tracks and vis over video[start_t:].
"""
if start_t < 0 or start_t >= video.shape[0]:
raise ValueError(f"start_t out of range: {start_t} for T={video.shape[0]}")
sub_video = video[start_t:]
T, H, W, _ = sub_video.shape
vid = rearrange(torch.from_numpy(sub_video).float().to(device) / 255.0, "t h w c -> 1 t c h w")
q = torch.from_numpy(queries_xy01.copy()).float().to(device)
q_pix = q * torch.tensor([W, H], device=device).float()
q_with_t = torch.cat([torch.zeros_like(q_pix[:, :1]), q_pix], dim=1) # t=0 in sub-video
pred_tracks, pred_vis = model(vid, queries=q_with_t[None], backward_tracking=True)
tracks = pred_tracks[0].detach().cpu().numpy()
vis = pred_vis[0].detach().cpu().numpy()
tracks[:, :, 0] /= float(W)
tracks[:, :, 1] /= float(H)
return tracks, vis
def _track_time_indexed_queries(
video: np.ndarray,
base_queries_xy01: np.ndarray,
model,
device: torch.device,
) -> tuple[np.ndarray, np.ndarray]:
"""
Track fixed query set launched at every frame in one CoTracker pass.
Returns:
tracks: [T, T*N, 2] (time, concatenated query-groups, xy01)
vis: [T, T*N]
Query-group k corresponds to queries initialized at frame k.
"""
T, H, W, _ = video.shape
N = base_queries_xy01.shape[0]
vid = rearrange(torch.from_numpy(video).float().to(device) / 255.0, "t h w c -> 1 t c h w")
q = torch.from_numpy(base_queries_xy01.copy()).float().to(device) # [N,2]
q_pix = q * torch.tensor([W, H], device=device).float() # [N,2]
q_rep = q_pix[None].repeat(T, 1, 1).reshape(T * N, 2) # [T*N,2]
t_idx = (
torch.arange(T, device=device, dtype=torch.float32)[:, None]
.repeat(1, N)
.reshape(T * N, 1)
) # [T*N,1]
q_with_t = torch.cat([t_idx, q_rep], dim=1) # [T*N,3]
pred_tracks, pred_vis = model(vid, queries=q_with_t[None], backward_tracking=True)
tracks = pred_tracks[0].detach().cpu().numpy() # [T, T*N, 2] px
vis = pred_vis[0].detach().cpu().numpy() # [T, T*N]
tracks[:, :, 0] /= float(W)
tracks[:, :, 1] /= float(H)
return tracks, vis
def _build_features() -> dict[str, dict]:
return {
"image": {"dtype": "image", "shape": (256, 256, 3), "names": ["height", "width", "channel"]},
"wrist_image": {"dtype": "image", "shape": (256, 256, 3), "names": ["height", "width", "channel"]},
"state": {"dtype": "float32", "shape": (8,), "names": ["state"]},
"actions": {"dtype": "float32", "shape": (7,), "names": ["actions"]},
"agentview_tracks": {"dtype": "float32", "shape": (7, 2), "names": ["points", "xy"]},
"agentview_vis": {"dtype": "float32", "shape": (7,), "names": ["points"]},
"wrist_tracks": {"dtype": "float32", "shape": (32, 2), "names": ["points", "xy"]},
"wrist_vis": {"dtype": "float32", "shape": (32,), "names": ["points"]},
"track_targets_raw": {"dtype": "float32", "shape": (16, 39, 2), "names": ["h", "points", "xy"]},
"agentview_grid_mesh_32": {"dtype": "float32", "shape": (32, 2), "names": ["points", "xy"]},
"wrist_grid_mesh_32": {"dtype": "float32", "shape": (32, 2), "names": ["points", "xy"]},
"agentview_mesh_vertices_2d": {"dtype": "float32", "shape": (7, 2), "names": ["points", "xy"]},
"wrist_mesh_vertices_2d": {"dtype": "float32", "shape": (7, 2), "names": ["points", "xy"]},
"has_track_mesh": {"dtype": "float32", "shape": (1,), "names": ["flag"]},
}
def _track_window(tracks: np.ndarray, start: int, horizon: int) -> np.ndarray:
end = min(start + horizon, tracks.shape[0])
window = tracks[start:end]
if window.shape[0] == 0:
return np.zeros((horizon, tracks.shape[1], tracks.shape[2]), dtype=tracks.dtype)
if window.shape[0] < horizon:
pad = np.repeat(window[-1:], horizon - window.shape[0], axis=0)
window = np.concatenate([window, pad], axis=0)
return window
def _to_hwc_uint8(image: np.ndarray) -> np.ndarray:
img = np.asarray(image)
if img.ndim == 3 and img.shape[0] in (1, 3) and img.shape[-1] not in (1, 3):
img = np.transpose(img, (1, 2, 0))
if img.dtype != np.uint8:
if np.issubdtype(img.dtype, np.floating):
img = np.clip(img, 0.0, 1.0)
img = (img * 255.0).astype(np.uint8)
else:
img = img.astype(np.uint8)
return img
@dataclass
class EpisodeBounds:
start: int
end: int
task: str
def _episode_bounds(ds: LeRobotDataset, ep_idx: int) -> EpisodeBounds:
ep_from = int(ds.episode_data_index["from"][ep_idx])
ep_to = int(ds.episode_data_index["to"][ep_idx])
task = ds.meta.episodes[ep_idx]["tasks"][0]
return EpisodeBounds(ep_from, ep_to, task)
def _episode_file_from_global_index(ds_root: Path, global_idx: int) -> Path:
chunk = global_idx // 1000
chunk_dir = ds_root / "data" / f"chunk-{chunk:03d}"
# Prefer global naming (episode_001000.parquet), but also support chunk-local naming
# (episode_000000.parquet in chunk-001) for compatibility with older exports.
global_name = chunk_dir / f"episode_{global_idx:06d}.parquet"
local_name = chunk_dir / f"episode_{(global_idx % 1000):06d}.parquet"
if global_name.exists():
return global_name
if local_name.exists():
return local_name
raise FileNotFoundError(
"Episode parquet not found for global_idx="
f"{global_idx}: tried {global_name} and {local_name}"
)
def _coerce_mesh7(arr: np.ndarray, *, key: str, frame_idx: int) -> np.ndarray:
a = np.asarray(arr, dtype=np.float32)
# Accept exact (7,2) or larger point sets where first 7 are canonical mesh points.
if a.ndim != 2 or a.shape[-1] != 2:
raise ValueError(f"{key} at frame={frame_idx} must be [P,2], got {a.shape}")
if a.shape[0] < 7:
raise ValueError(f"{key} at frame={frame_idx} has <7 points: {a.shape}")
return a[:7]
def _mesh_seeds_for_scene(scene: str) -> tuple[list[int], list[int]]:
ids = SCENE_VERTEX_IDS.get(scene, DEFAULT_VERTEX_IDS)
return ids["agentview"], ids["eye_in_hand"]
def _normalize_task_text(text: str) -> str:
return re.sub(r"[^a-z0-9]+", " ", str(text).lower()).strip()
def _resolve_bddl_for_task(
task_name: str | None,
benchmark_mod,
bddl_roots: list[Path],
) -> str | None:
if not task_name:
return None
target = _normalize_task_text(task_name)
if not target:
return None
best_score = -1
best_path: Path | None = None
benchmark_dict = benchmark_mod.get_benchmark_dict()
for _, suite_ctor in benchmark_dict.items():
try:
suite = suite_ctor()
except Exception:
continue
for i in range(getattr(suite, "n_tasks", 0)):
try:
t = suite.get_task(i)
except Exception:
continue
language = _normalize_task_text(getattr(t, "language", ""))
if not language:
continue
score = 0
if language == target:
score = 3
elif target in language or language in target:
score = 2
elif set(target.split()) & set(language.split()):
score = 1
if score < best_score:
continue
rel = Path(t.problem_folder) / t.bddl_file
for root in bddl_roots:
cand = root / rel
if cand.exists():
best_score = score
best_path = cand
break
return str(best_path) if best_path is not None else None
def _select_dynamic_ids_for_camera(
env,
camera_name: str,
fallback_agent_ids: list[int],
img_hw: tuple[int, int],
) -> list[int]:
if not hasattr(env, "vertex_mapping"):
return fallback_agent_ids
base_env = env
while hasattr(base_env, "env") and not hasattr(base_env, "sim"):
base_env = base_env.env
if not hasattr(base_env, "sim"):
return fallback_agent_ids
model = base_env.sim.model
def _group_for_geom(name: str) -> str:
n = (name or "").lower()
if "finger1" in n:
return "finger1"
if "finger2" in n:
return "finger2"
if "hand" in n or "gripper" in n:
return "base"
return "other"
candidates: dict[int, str] = {}
for vid, mapping in env.vertex_mapping.items():
geom_id = int(mapping.get("geom_id", -1))
if geom_id < 0:
continue
geom_name = model.geom_id2name(geom_id)
group = _group_for_geom(geom_name)
if group != "other":
candidates[int(vid)] = group
if not candidates:
return fallback_agent_ids
vids = list(candidates.keys())
if hasattr(env, "_update_vertex_positions"):
env._update_vertex_positions()
tracked = env._get_tracked_vertices(camera_name, vids)
id_to_xy = {int(i): p for i, p in zip(tracked["vertex_ids"].tolist(), tracked["vertices_2d"])}
rows = []
w, h = float(img_hw[1]), float(img_hw[0])
for vid in vids:
if vid not in id_to_xy:
continue
x = float(id_to_xy[vid][0]) / w
y = float(id_to_xy[vid][1]) / h
if not (0.0 <= x <= 1.0 and 0.0 <= y <= 1.0):
continue
rows.append({"vid": vid, "group": candidates[vid], "x": x, "y": y})
def _pick_three(group_rows: list[dict]) -> list[dict]:
if len(group_rows) <= 3:
return group_rows
ys = np.array([r["y"] for r in group_rows], dtype=np.float32)
lo = int(np.argmin(ys))
hi = int(np.argmax(ys))
mid_target = float((ys[lo] + ys[hi]) * 0.5)
mid = min(range(len(group_rows)), key=lambda i: abs(float(group_rows[i]["y"]) - mid_target))
idx = []
for i in (lo, mid, hi):
if i not in idx:
idx.append(i)
for i in range(len(group_rows)):
if len(idx) >= 3:
break
if i not in idx:
idx.append(i)
return [group_rows[i] for i in idx[:3]]
f1 = _pick_three([r for r in rows if r["group"] == "finger1"])
f2 = _pick_three([r for r in rows if r["group"] == "finger2"])
b = [r for r in rows if r["group"] == "base"]
if len(f1) < 2 or len(f2) < 2 or not b:
return fallback_agent_ids
mx = (float(np.mean([r["x"] for r in f1])) + float(np.mean([r["x"] for r in f2]))) / 2.0
my = (float(np.mean([r["y"] for r in f1])) + float(np.mean([r["y"] for r in f2]))) / 2.0
base = sorted(b, key=lambda r: abs(r["x"] - mx) + 0.5 * abs(r["y"] - my))[0]
picked = (f1 + f2 + [base])[:7]
picked_ids = [int(r["vid"]) for r in picked]
return picked_ids if len(picked_ids) == 7 else fallback_agent_ids
def _get_mesh_seeds_from_sim(scene: str, state_vec: np.ndarray, img_hw=(256, 256)) -> tuple[np.ndarray, np.ndarray]:
last_exc = None
OffScreenRenderEnv = None
MeshVertexWrapper = None
MeshVertexTracker = None
benchmark_mod = None
get_libero_path = None
for mod_name in ("libero.envs.env_wrapper", "libero.libero.envs.env_wrapper"):
try:
env_mod = importlib.import_module(mod_name)
OffScreenRenderEnv = env_mod.OffScreenRenderEnv # type: ignore
base_pkg = env_mod.__package__.rsplit(".env_wrapper", 1)[0]
for benchmark_mod_name in ("libero.benchmark", "libero.libero.benchmark"):
try:
benchmark_mod = importlib.import_module(benchmark_mod_name)
break
except Exception:
benchmark_mod = None # type: ignore
continue
if benchmark_mod is None:
raise ModuleNotFoundError("libero.benchmark not found")
try:
get_libero_path = importlib.import_module("libero").get_libero_path # type: ignore
except AttributeError:
get_libero_path = importlib.import_module("libero.libero").get_libero_path # type: ignore
mesh_vertex_wrapper = None
mesh_vertex_tracker = None
for wrapper_mod in (
f"{base_pkg}.mesh_vertex_wrapper",
"libero.envs.mesh_vertex_wrapper",
"libero.libero.envs.mesh_vertex_wrapper",
):
try:
mesh_vertex_wrapper = importlib.import_module(wrapper_mod)
break
except Exception:
mesh_vertex_wrapper = None # type: ignore
continue
for tracker_mod in (
f"{base_pkg}.mesh_vertex_tracker",
"libero.envs.mesh_vertex_tracker",
"libero.libero.envs.mesh_vertex_tracker",
):
try:
mesh_vertex_tracker = importlib.import_module(tracker_mod)
break
except Exception:
mesh_vertex_tracker = None # type: ignore
continue
if mesh_vertex_wrapper is None or mesh_vertex_tracker is None:
fallback_root = EXTRA_LIBERO_PATH or ""
wrapper_path = Path(fallback_root) / "libero" / "envs" / "mesh_vertex_wrapper.py"
tracker_path = Path(fallback_root) / "libero" / "envs" / "mesh_vertex_tracker.py"
if mesh_vertex_wrapper is None and wrapper_path.exists():
mesh_vertex_wrapper = SourceFileLoader("libero.envs.mesh_vertex_wrapper_fallback", str(wrapper_path)).load_module() # type: ignore
if mesh_vertex_tracker is None and tracker_path.exists():
mesh_vertex_tracker = SourceFileLoader("libero.envs.mesh_vertex_tracker_fallback", str(tracker_path)).load_module() # type: ignore
if mesh_vertex_wrapper is None or mesh_vertex_tracker is None:
raise ModuleNotFoundError("mesh vertex wrapper/tracker not found in LIBERO installation")
MeshVertexWrapper = mesh_vertex_wrapper.MeshVertexWrapper # type: ignore
MeshVertexTracker = mesh_vertex_tracker.MeshVertexTracker # type: ignore
break
except Exception as exc: # noqa: BLE001
last_exc = exc
OffScreenRenderEnv = None # type: ignore
continue
if OffScreenRenderEnv is None:
raise ImportError(
"LIBERO simulator not available. Ensure libero is on PYTHONPATH and provides libero/envs/env_wrapper.py. "
"Set --extra-libero-path to the LIBERO package path (e.g., /mnt/kevin/code/wmrl/Dual-Dynamics-Models/ATM/libero)."
) from last_exc
benchmark_dict = benchmark_mod.get_benchmark_dict()
suite = benchmark_dict["libero_goal"]() # suite choice not critical for camera layout
# Heuristic: pick first task to init cameras; real scene choice drives vertex ids.
task = suite.get_task(0)
rel_bddl = Path(task.problem_folder) / task.bddl_file
bddl_roots: list[Path] = []
try:
bddl_roots.append(Path(get_libero_path("bddl_files")))
except Exception:
pass
for module_name in ("libero.libero", "libero"):
try:
mod = importlib.import_module(module_name)
mod_file = getattr(mod, "__file__", None)
if mod_file:
bddl_roots.append(Path(mod_file).resolve().parent / "bddl_files")
except Exception:
continue
if env_mod and getattr(env_mod, "__file__", None):
bddl_roots.append(Path(env_mod.__file__).resolve().parents[1] / "bddl_files")
if EXTRA_LIBERO_PATH:
bddl_roots.append(Path(EXTRA_LIBERO_PATH) / "libero" / "bddl_files")
bddl_roots.append(Path(EXTRA_LIBERO_PATH) / "bddl_files")
seen = set()
uniq_roots = []
for root in bddl_roots:
k = str(root)
if k not in seen:
seen.add(k)
uniq_roots.append(root)
task_bddl_file = None
for root in uniq_roots:
cand = root / rel_bddl
if cand.exists():
task_bddl_file = str(cand)
break
if task_bddl_file is None:
raise FileNotFoundError(
f"Could not resolve BDDL file {rel_bddl} from candidates: {[str(p) for p in uniq_roots]}"
)
task_mapping = getattr(env_mod, "TASK_MAPPING", {})
known_problems = set(task_mapping.keys()) if isinstance(task_mapping, dict) else set()
def _problem_name_from_bddl(path: str) -> str | None:
try:
txt = Path(path).read_text(encoding="utf-8", errors="ignore")
except Exception:
return None
m = re.search(r"\(problem\s+([^\s\)]+)\)", txt)
return m.group(1) if m else None
candidate_bddls = [task_bddl_file]
for root in uniq_roots:
if not root.exists():
continue
for cand in root.rglob("*.bddl"):
prob = _problem_name_from_bddl(str(cand))
if prob is None or (known_problems and prob in known_problems):
candidate_bddls.append(str(cand))
seen = set()
candidate_bddls = [p for p in candidate_bddls if not (p in seen or seen.add(p))]
env = None
env_exc = None
for cand in candidate_bddls:
try:
env = OffScreenRenderEnv(
bddl_file_name=cand,
camera_heights=img_hw[0],
camera_widths=img_hw[1],
render_gpu_device_id=0,
)
break
except Exception as exc: # noqa: BLE001
env_exc = exc
continue
if env is None:
raise RuntimeError("Failed to create OffScreenRenderEnv from candidate BDDL files") from env_exc
agent_ids, wrist_ids = _mesh_seeds_for_scene(scene)
env = MeshVertexWrapper(env, max_points=10000, include_non_mesh_geoms=True)
env = MeshVertexTracker(env, agentview_vertex_ids=agent_ids, eye_in_hand_vertex_ids=wrist_ids, img_height=img_hw[0], img_width=img_hw[1])
env.reset()
base_env = env
while hasattr(base_env, "env") and not hasattr(base_env, "sim"):
base_env = base_env.env
expected_qpos = int(getattr(getattr(base_env, "sim", None).model, "nq", 0)) if hasattr(base_env, "sim") else 0
state_arr = np.asarray(state_vec, dtype=np.float32).reshape(-1)
if expected_qpos > 0 and state_arr.shape[0] >= expected_qpos:
try:
env.set_init_state(state_arr)
except Exception as exc:
print(f"[warn] set_init_state failed ({exc}); using reset state for mesh seeds")
else:
print(
f"[warn] simulator state unavailable or mismatched (got {state_arr.shape[0]}, expected >= {expected_qpos}); "
"using reset state for mesh seeds"
)
if hasattr(env, "_update_vertex_positions"):
env._update_vertex_positions()
av = env._get_tracked_vertices("agentview", env.agentview_vertex_ids)["vertices_2d"]
wr = env._get_tracked_vertices("robot0_eye_in_hand", env.eye_in_hand_vertex_ids)["vertices_2d"]
env.close()
av = np.asarray(av, dtype=np.float32) / float(img_hw[1])
wr = np.asarray(wr, dtype=np.float32) / float(img_hw[1])
return av, wr
def _get_mesh_sequence_from_sim(
scene: str,
state_vec: np.ndarray,
actions: np.ndarray,
task_name: str | None = None,
img_hw=(256, 256),
) -> tuple[np.ndarray, np.ndarray]:
"""Return per-frame simulator mesh vertices by stepping the simulator with episode actions."""
last_exc = None
OffScreenRenderEnv = None
MeshVertexWrapper = None
MeshVertexTracker = None
benchmark_mod = None
get_libero_path = None
env_mod = None
for mod_name in ("libero.envs.env_wrapper", "libero.libero.envs.env_wrapper"):
try:
env_mod = importlib.import_module(mod_name)
OffScreenRenderEnv = env_mod.OffScreenRenderEnv # type: ignore
base_pkg = env_mod.__package__.rsplit(".env_wrapper", 1)[0]
for benchmark_mod_name in ("libero.benchmark", "libero.libero.benchmark"):
try:
benchmark_mod = importlib.import_module(benchmark_mod_name)
break
except Exception:
benchmark_mod = None # type: ignore
continue
if benchmark_mod is None:
raise ModuleNotFoundError("libero.benchmark not found")
try:
get_libero_path = importlib.import_module("libero").get_libero_path # type: ignore
except AttributeError:
get_libero_path = importlib.import_module("libero.libero").get_libero_path # type: ignore
mesh_vertex_wrapper = None
mesh_vertex_tracker = None
for wrapper_mod in (
f"{base_pkg}.mesh_vertex_wrapper",
"libero.envs.mesh_vertex_wrapper",
"libero.libero.envs.mesh_vertex_wrapper",
):
try:
mesh_vertex_wrapper = importlib.import_module(wrapper_mod)
break
except Exception:
mesh_vertex_wrapper = None # type: ignore
continue
for tracker_mod in (
f"{base_pkg}.mesh_vertex_tracker",
"libero.envs.mesh_vertex_tracker",
"libero.libero.envs.mesh_vertex_tracker",
):
try:
mesh_vertex_tracker = importlib.import_module(tracker_mod)
break
except Exception:
mesh_vertex_tracker = None # type: ignore
continue
if mesh_vertex_wrapper is None or mesh_vertex_tracker is None:
fallback_root = EXTRA_LIBERO_PATH or ""
wrapper_path = Path(fallback_root) / "libero" / "envs" / "mesh_vertex_wrapper.py"
tracker_path = Path(fallback_root) / "libero" / "envs" / "mesh_vertex_tracker.py"
if mesh_vertex_wrapper is None and wrapper_path.exists():
mesh_vertex_wrapper = SourceFileLoader("libero.envs.mesh_vertex_wrapper_fallback", str(wrapper_path)).load_module() # type: ignore
if mesh_vertex_tracker is None and tracker_path.exists():
mesh_vertex_tracker = SourceFileLoader("libero.envs.mesh_vertex_tracker_fallback", str(tracker_path)).load_module() # type: ignore
if mesh_vertex_wrapper is None or mesh_vertex_tracker is None:
raise ModuleNotFoundError("mesh vertex wrapper/tracker not found in LIBERO installation")
MeshVertexWrapper = mesh_vertex_wrapper.MeshVertexWrapper # type: ignore
MeshVertexTracker = mesh_vertex_tracker.MeshVertexTracker # type: ignore
break
except Exception as exc: # noqa: BLE001
last_exc = exc
OffScreenRenderEnv = None # type: ignore
continue
if OffScreenRenderEnv is None:
raise ImportError(
"LIBERO simulator not available. Ensure libero is on PYTHONPATH and provides libero/envs/env_wrapper.py. "
"Set --extra-libero-path to the LIBERO package path (e.g., /mnt/kevin/code/wmrl/Dual-Dynamics-Models/ATM/libero)."
) from last_exc
benchmark_dict = benchmark_mod.get_benchmark_dict()
suite = benchmark_dict["libero_goal"]()
task = suite.get_task(0)
rel_bddl = Path(task.problem_folder) / task.bddl_file
bddl_roots: list[Path] = []
try:
bddl_roots.append(Path(get_libero_path("bddl_files")))
except Exception:
pass
for module_name in ("libero.libero", "libero"):
try:
mod = importlib.import_module(module_name)
mod_file = getattr(mod, "__file__", None)
if mod_file:
bddl_roots.append(Path(mod_file).resolve().parent / "bddl_files")
except Exception:
continue
if env_mod and getattr(env_mod, "__file__", None):
bddl_roots.append(Path(env_mod.__file__).resolve().parents[1] / "bddl_files")
if EXTRA_LIBERO_PATH:
bddl_roots.append(Path(EXTRA_LIBERO_PATH) / "libero" / "bddl_files")
bddl_roots.append(Path(EXTRA_LIBERO_PATH) / "bddl_files")
seen = set()
uniq_roots = []
for root in bddl_roots:
k = str(root)
if k not in seen:
seen.add(k)
uniq_roots.append(root)
task_bddl_file = None
for root in uniq_roots:
cand = root / rel_bddl
if cand.exists():
task_bddl_file = str(cand)
break
if task_bddl_file is None:
raise FileNotFoundError(
f"Could not resolve BDDL file {rel_bddl} from candidates: {[str(p) for p in uniq_roots]}"
)
task_mapping = getattr(env_mod, "TASK_MAPPING", {})
known_problems = set(task_mapping.keys()) if isinstance(task_mapping, dict) else set()
def _problem_name_from_bddl(path: str) -> str | None:
try:
txt = Path(path).read_text(encoding="utf-8", errors="ignore")
except Exception:
return None
m = re.search(r"\(problem\s+([^\s\)]+)\)", txt)
return m.group(1) if m else None
resolved_bddl = _resolve_bddl_for_task(task_name, benchmark_mod, uniq_roots)
candidate_bddls = [resolved_bddl] if resolved_bddl else [task_bddl_file]
for root in uniq_roots:
if not root.exists():
continue
for cand in root.rglob("*.bddl"):
prob = _problem_name_from_bddl(str(cand))
if prob is None or (known_problems and prob in known_problems):
candidate_bddls.append(str(cand))
seen = set()
candidate_bddls = [p for p in candidate_bddls if not (p in seen or seen.add(p))]
env = None
env_exc = None
for cand in candidate_bddls:
try:
env = OffScreenRenderEnv(
bddl_file_name=cand,
camera_heights=img_hw[0],
camera_widths=img_hw[1],
render_gpu_device_id=0,
)
break
except Exception as exc: # noqa: BLE001
env_exc = exc
continue
if env is None:
raise RuntimeError("Failed to create OffScreenRenderEnv from candidate BDDL files") from env_exc
agent_ids, wrist_ids = _mesh_seeds_for_scene(scene)
env = MeshVertexWrapper(env, max_points=10000, include_non_mesh_geoms=True)
env = MeshVertexTracker(
env,
agentview_vertex_ids=agent_ids,
eye_in_hand_vertex_ids=wrist_ids,
img_height=img_hw[0],
img_width=img_hw[1],
)
env.reset()
base_env = env
while hasattr(base_env, "env") and not hasattr(base_env, "sim"):
base_env = base_env.env
expected_qpos = int(getattr(getattr(base_env, "sim", None).model, "nq", 0)) if hasattr(base_env, "sim") else 0
state_arr = np.asarray(state_vec, dtype=np.float32).reshape(-1)
if expected_qpos > 0 and state_arr.shape[0] >= expected_qpos:
try:
env.set_init_state(state_arr)
except Exception as exc:
print(f"[warn] set_init_state failed ({exc}); using reset state for mesh seeds")
else:
print(
f"[warn] simulator state unavailable or mismatched (got {state_arr.shape[0]}, expected >= {expected_qpos}); "
"using reset state for mesh seeds"
)
agent_ids = _select_dynamic_ids_for_camera(
env,
camera_name="agentview",
fallback_agent_ids=agent_ids,
img_hw=img_hw,
)
wrist_ids = _select_dynamic_ids_for_camera(
env,
camera_name="robot0_eye_in_hand",
fallback_agent_ids=wrist_ids,
img_hw=img_hw,
)
env.agentview_vertex_ids = agent_ids
env.eye_in_hand_vertex_ids = wrist_ids
actions_arr = np.asarray(actions, dtype=np.float32)
if actions_arr.ndim != 2:
raise ValueError(f"Expected actions shape [T, A], got {actions_arr.shape}")
av_seq = []
wr_seq = []
for t in range(actions_arr.shape[0]):
if hasattr(env, "_update_vertex_positions"):
env._update_vertex_positions()
av = env._get_tracked_vertices("agentview", env.agentview_vertex_ids)["vertices_2d"]
wr = env._get_tracked_vertices("robot0_eye_in_hand", env.eye_in_hand_vertex_ids)["vertices_2d"]
av_seq.append(np.asarray(av, dtype=np.float32) / float(img_hw[1]))
wr_seq.append(np.asarray(wr, dtype=np.float32) / float(img_hw[1]))
if t < actions_arr.shape[0] - 1:
env.step(actions_arr[t])
env.close()
return np.stack(av_seq, axis=0), np.stack(wr_seq, axis=0)
def process_episode(
ds: LeRobotDataset,
ep_idx: int,
horizon: int,
cotracker_model,
device: torch.device,
grid25: np.ndarray,
mesh_source: str = "sim",
mesh_ds: LeRobotDataset | None = None,
) -> Iterable[dict]:
bnds = _episode_bounds(ds, ep_idx)
frames = []
offline_agent_mesh = []
offline_wrist_mesh = []
# Pull source frames for observations/state/actions.
for i in range(bnds.start, bnds.end):
row = ds[i]
agent_mesh_row = row.get("agentview_mesh_vertices_2d", None)
wrist_mesh_row = row.get("wrist_mesh_vertices_2d", None)
frames.append(
(
np.asarray(row["image"]),
np.asarray(row["wrist_image"]),
np.asarray(row["state"], dtype=np.float32),
np.asarray(row["actions"], dtype=np.float32),
row.get("task", bnds.task),
None if agent_mesh_row is None else np.asarray(agent_mesh_row, dtype=np.float32),
None if wrist_mesh_row is None else np.asarray(wrist_mesh_row, dtype=np.float32),
)
)
offline_agent_mesh.append(agent_mesh_row is not None)
offline_wrist_mesh.append(wrist_mesh_row is not None)
images = np.stack([f[0] for f in frames], axis=0)
wrist_images = np.stack([f[1] for f in frames], axis=0)
states = np.stack([f[2] for f in frames], axis=0)
actions = np.stack([f[3] for f in frames], axis=0)
task = frames[0][4]
# Normalize to HWC uint8
images = np.stack([_to_hwc_uint8(f[0]) for f in frames], axis=0)
wrist_images = np.stack([_to_hwc_uint8(f[1]) for f in frames], axis=0)
states = np.stack([f[2] for f in frames], axis=0)
actions = np.stack([f[3] for f in frames], axis=0)
task = frames[0][4]
if images.shape[1] != 256 or images.shape[2] != 256:
raise ValueError(f"Expected 256x256 images, got {images.shape[1:3]}")
if mesh_source == "dataset":
if not all(offline_agent_mesh) or not all(offline_wrist_mesh):
raise ValueError(
"mesh_source=dataset requires source dataset columns "
"`agentview_mesh_vertices_2d` and `wrist_mesh_vertices_2d` for all frames."
)
agent_mesh_seq = np.stack([f[5] for f in frames], axis=0)
wrist_mesh_seq = np.stack([f[6] for f in frames], axis=0)
elif mesh_source == "mesh_repo":
if mesh_ds is None:
raise ValueError("mesh_source=mesh_repo requires mesh_ds to be provided.")
mesh_root = Path(str(mesh_ds.repo_id))
mesh_ep_file = _episode_file_from_global_index(mesh_root, ep_idx)
mesh_table = pq.read_table(mesh_ep_file, columns=["agentview_mesh_vertices_2d", "wrist_mesh_vertices_2d"])
src_len = bnds.end - bnds.start
if mesh_table.num_rows != src_len:
raise ValueError(
f"mesh_repo episode length mismatch for ep={ep_idx}: "
f"source_len={src_len} mesh_len={mesh_table.num_rows}"
)
mesh_agent_raw = mesh_table["agentview_mesh_vertices_2d"].to_pylist()
mesh_wrist_raw = mesh_table["wrist_mesh_vertices_2d"].to_pylist()
mesh_agent = [_coerce_mesh7(v, key="agentview_mesh_vertices_2d", frame_idx=i) for i, v in enumerate(mesh_agent_raw)]
mesh_wrist = [_coerce_mesh7(v, key="wrist_mesh_vertices_2d", frame_idx=i) for i, v in enumerate(mesh_wrist_raw)]
agent_mesh_seq = np.stack(mesh_agent, axis=0).astype(np.float32)
wrist_mesh_seq = np.stack(mesh_wrist, axis=0).astype(np.float32)
else:
scene = _scene_from_task(bnds.task)
# Mesh vertices from simulator for each frame (step simulator with episode actions).
agent_mesh_seq, wrist_mesh_seq = _get_mesh_sequence_from_sim(
scene, states[0], actions, task_name=task, img_hw=(images.shape[1], images.shape[2])
)
# Dense wrist grid tracking: launch fixed grid25 at every frame.
dense_grid_tracks, dense_grid_vis = _track_time_indexed_queries(
wrist_images, grid25, cotracker_model, device
) # [T, T*25, 2], [T, T*25]
# Build frames
for local_t, global_t in enumerate(range(bnds.start, bnds.end)):
# Dense per-frame logic:
# - agent window uses simulator mesh7 sequence from current frame onward
# - wrist window uses fixed grid25 queries launched at current frame, plus simulator mesh7 sequence
agent_window = _track_window(agent_mesh_seq, local_t, horizon) # (H,7,2)
q0 = local_t * grid25.shape[0]
q1 = q0 + grid25.shape[0]
wrist_grid_future_src = dense_grid_tracks[:, q0:q1, :] # (T,25,2) for queries started at local_t
wrist_grid_window = _track_window(wrist_grid_future_src, local_t, horizon) # (H,25,2)
wrist_mesh_window = _track_window(wrist_mesh_seq, local_t, horizon) # (H,7,2)
wrist_window = np.concatenate([wrist_grid_window, wrist_mesh_window], axis=1) # (H,32,2)
track_targets_raw = np.concatenate([agent_window, wrist_window], axis=1) # (H,39,2)
wrist_grid_curr = dense_grid_tracks[local_t, q0:q1, :] # (25,2)
wrist_grid_vis_curr = np.asarray(dense_grid_vis[local_t, q0:q1], dtype=np.float32) # (25,)
wrist_tracks_curr = np.concatenate([wrist_grid_curr, wrist_mesh_seq[local_t]], axis=0) # (32,2)
wrist_vis_curr = np.concatenate(
[wrist_grid_vis_curr, np.ones((7,), dtype=np.float32)],
axis=0,
)
yield {
"image": images[local_t],
"wrist_image": wrist_images[local_t],
"state": states[local_t],
"actions": actions[local_t],
"task": task,
"agentview_tracks": agent_mesh_seq[local_t],
"agentview_vis": np.ones((7,), dtype=np.float32),
"wrist_tracks": wrist_tracks_curr,
"wrist_vis": wrist_vis_curr,
"track_targets_raw": track_targets_raw,
"agentview_mesh_vertices_2d": agent_mesh_seq[local_t],
"wrist_mesh_vertices_2d": wrist_mesh_seq[local_t],
"agentview_grid_mesh_32": np.concatenate([grid25, agent_mesh_seq[local_t]], axis=0),
"wrist_grid_mesh_32": np.concatenate([grid25, wrist_mesh_seq[local_t]], axis=0),
"has_track_mesh": np.asarray([1.0], dtype=np.float32),
}
def main():
p = argparse.ArgumentParser(description="Reprocess LIBERO RLDS to LeRobot with CoTracker and simulator mesh seeds.")
p.add_argument("--source-repo-id", default="/mnt/kevin/data/physical-intelligence/libero")
p.add_argument("--target-repo-id", default="/mnt/kevin/data/physical-intelligence/libero_with_tracks_mesh_cotracker")
p.add_argument("--cotracker-checkpoint", default="/mnt/kevin/vlm_models/cotracker/scaled_offline.pth")
p.add_argument("--horizon", type=int, default=16)
p.add_argument(
"--mesh-source",
choices=["sim", "dataset", "mesh_repo"],
default="sim",
help=(
"Mesh timeline source: sim=simulator rollout; dataset=source dataset mesh columns; "
"mesh_repo=load mesh columns from --mesh-repo-id aligned by episode/frame."
),
)
p.add_argument(
"--mesh-repo-id",
type=str,
default=None,
help="LeRobot dataset path used when --mesh-source=mesh_repo.",
)
p.add_argument("--overwrite", action="store_true")
p.add_argument("--max-episodes", type=int, default=None)
p.add_argument("--start-episode", type=int, default=0, help="Start episode index (inclusive).")
p.add_argument("--end-episode", type=int, default=None, help="End episode index (exclusive).")
p.add_argument("--libero-root", type=str, default=None, help="Path to LIBERO codebase (adds to sys.path)")
p.add_argument("--libero-data", type=str, default=None, help="Path to LIBERO assets (sets LIBERO_PATH)")
p.add_argument(
"--extra-libero-path",
type=str,
default="/mnt/kevin/code/wmrl/Dual-Dynamics-Models/ATM/libero",
help="LIBERO code path to add to sys.path (expects libero/envs/env_wrapper.py).",
)
args = p.parse_args()
global EXTRA_LIBERO_PATH
EXTRA_LIBERO_PATH = args.extra_libero_path
if args.libero_root and (Path(args.libero_root) / "libero" / "envs" / "mesh_vertex_wrapper.py").exists():
EXTRA_LIBERO_PATH = args.libero_root
sys_paths = []
if args.libero_root:
sys_paths.append(args.libero_root)
if args.extra_libero_path:
sys_paths.append(args.extra_libero_path)
for pth in reversed(sys_paths):
if pth and pth not in sys.path and Path(pth).exists():
sys.path.insert(0, pth)
if args.libero_data:
os.environ.setdefault("LIBERO_PATH", args.libero_data)
os.environ.setdefault("MUJOCO_GL", "egl")
os.environ.setdefault("PYOPENGL_PLATFORM", "egl")
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
grid25 = _make_grid_25()
print(f"Loading source LeRobot dataset: {args.source_repo_id}")
src = LeRobotDataset(args.source_repo_id)
mesh_ds = None
if args.mesh_source == "mesh_repo":
if not args.mesh_repo_id:
raise ValueError("--mesh-repo-id is required when --mesh-source=mesh_repo.")
print(f"Loading mesh LeRobot dataset: {args.mesh_repo_id}")
mesh_ds = LeRobotDataset(args.mesh_repo_id)
if len(mesh_ds.meta.episodes) < len(src.meta.episodes):
raise ValueError(
f"Mesh dataset has fewer episodes ({len(mesh_ds.meta.episodes)}) than source ({len(src.meta.episodes)})."
)
total_eps = len(src.meta.episodes)
start_ep = max(0, int(args.start_episode))
end_ep = total_eps if args.end_episode is None else min(int(args.end_episode), total_eps)
if args.max_episodes is not None:
end_ep = min(end_ep, start_ep + int(args.max_episodes))
if end_ep <= start_ep:
raise ValueError(f"Invalid episode range [{start_ep}, {end_ep}) for total {total_eps}")
n_eps = end_ep - start_ep
target_path = Path(args.target_repo_id)
if not target_path.is_absolute():
target_path = Path(os.environ.get("HF_HOME", Path.home() / ".cache" / "huggingface")) / args.target_repo_id
if target_path.exists() and args.overwrite:
import shutil
shutil.rmtree(target_path)
dst = LeRobotDataset.create(
repo_id=str(target_path),
robot_type="panda",
fps=src.fps,
features=_build_features(),
image_writer_threads=10,
image_writer_processes=5,
)
print(f"Loading CoTracker checkpoint from {args.cotracker_checkpoint} on {device}")
cot_model = _load_cotracker(args.cotracker_checkpoint, device)
for ep_idx in range(start_ep, end_ep):
print(f"[ep {ep_idx-start_ep}/{n_eps}] abs={ep_idx}")
for frame in process_episode(
src,
ep_idx,
args.horizon,
cot_model,
device,
grid25,
mesh_source=args.mesh_source,
mesh_ds=mesh_ds,
):
dst.add_frame(frame)
dst.save_episode()
print(f"Done. Wrote {n_eps} episodes ({start_ep}:{end_ep}) to {args.target_repo_id}")
if __name__ == "__main__":
main()