| import os |
| import json |
| import time |
| import random |
| from typing import Dict, Any, List, Optional, Tuple |
|
|
| from openai import OpenAI |
| import multiprocessing as mp |
|
|
| |
| |
| |
| def load_json(file_path: str): |
| with open(file_path, "r", encoding="utf-8") as f: |
| return json.load(f) |
|
|
| def append_jsonl(path: str, obj: Dict[str, Any]): |
| os.makedirs(os.path.dirname(path), exist_ok=True) |
| with open(path, "a", encoding="utf-8") as f: |
| f.write(json.dumps(obj, ensure_ascii=False) + "\n") |
|
|
| def load_done_keys(jsonl_path: str): |
| done = set() |
| if not os.path.exists(jsonl_path): |
| return done |
| with open(jsonl_path, "r", encoding="utf-8") as f: |
| for line in f: |
| line = line.strip() |
| if not line: |
| continue |
| try: |
| obj = json.loads(line) |
| k = obj.get("key") |
| if k is not None: |
| done.add(k) |
| except Exception: |
| continue |
| return done |
|
|
| |
| |
| |
| def safe_json_parse(text: str) -> Optional[Dict[str, Any]]: |
| text = (text or "").strip() |
| if text.startswith("```"): |
| text = text.strip("`").strip() |
| if text.lower().startswith("json"): |
| text = text[4:].strip() |
| try: |
| return json.loads(text) |
| except Exception: |
| return None |
|
|
| def call_chat_json( |
| client: OpenAI, |
| model: str, |
| system: str, |
| user: str, |
| max_tokens: int, |
| temperature: float, |
| top_p: float = 0.9, |
| max_retries: int = 4, |
| ) -> Dict[str, Any]: |
| last_err = None |
| for attempt in range(1, max_retries + 1): |
| try: |
| resp = client.chat.completions.create( |
| model=model, |
| messages=[ |
| {"role": "system", "content": system}, |
| {"role": "user", "content": user}, |
| ], |
| temperature=temperature, |
| top_p=top_p, |
| max_tokens=max_tokens, |
| ) |
| text = resp.choices[0].message.content |
| obj = safe_json_parse(text) |
| if obj is None: |
| raise ValueError(f"JSON parse failed. Raw: {text[:200]}...") |
| return obj |
| except Exception as e: |
| last_err = str(e) |
| time.sleep(1.0 * attempt) |
| raise RuntimeError(f"call_chat_json failed after {max_retries} retries. Last error: {last_err}") |
|
|
| def normalize_text(s: str) -> str: |
| return " ".join((s or "").lower().split()) |
|
|
|
|
| |
| |
| |
| _BINARY_ANS = {"yes", "no", "true", "false"} |
| _AUX_START = { |
| "is", "are", "was", "were", |
| "do", "does", "did", |
| "can", "could", |
| "will", "would", "should", |
| "has", "have", "had", |
| "may", "might", "must", |
| } |
|
|
| def is_binary_qa(question: str, gt_answer: str) -> bool: |
| a = normalize_text(gt_answer) |
| if a in _BINARY_ANS: |
| return True |
| q = (question or "").strip().lower() |
| |
| if q.endswith("?"): |
| first = q.split()[0] if q.split() else "" |
| if first in _AUX_START: |
| return True |
| return False |
|
|
| |
| |
| |
| SYSTEM_STATEMENT = r""" |
| You generate ONE correct answer sentence given a situation and a QA pair. |
| |
| CRITICAL: |
| - The output must be a single, continuous sentence that integrates the situation text with a short answer clause that resolves the question (e.g., “When …, …”). |
| |
| Return STRICT JSON: {"statement": "..."} only. |
| |
| Rules: |
| - Must be accurate given the answer. |
| - Natural, concise, factual. |
| - Preserve referenced entity labels if any (e.g., "chair A"). |
| - JSON only. |
| """ |
|
|
| USER_STATEMENT = """Situation: {situation} |
| Question: {question} |
| Ground-truth answer: {answer} |
| |
| Return JSON only. |
| """ |
|
|
| def gen_statement(client: OpenAI, model: str, situation: str, question: str, answer: str) -> str: |
| obj = call_chat_json( |
| client=client, |
| model=model, |
| system=SYSTEM_STATEMENT, |
| user=USER_STATEMENT.format(situation=situation, question=question, answer=answer), |
| max_tokens=512, |
| temperature=0.2, |
| ) |
| st = obj.get("statement", "") |
| if not isinstance(st, str) or not st.strip(): |
| raise ValueError(f"Bad statement: {obj}") |
| return st.strip() |
|
|
| |
| |
| |
| SYSTEM_WRONG_9 = r""" |
| You generate distractors for a ground-truth statement. |
| |
| CRITICAL: |
| - EVERY sentence MUST integrates the situation text with a wrong answer clause (e.g., “When …, …”). |
| |
| OUTPUT (STRICT JSON ONLY): |
| {{"distractors": ["<s1>", "<s2>", "<s3>", "<s4>", "<s5>", "<s6>", "<s7>", "<s8>", "<s9>"]}} |
| |
| Rules: |
| - The situation should be fixed in each sentence. |
| - Exactly 9 strings, all unique. |
| - Each is a single declarative sentence. |
| - Each must be WRONG given the ground-truth answer. |
| - Preserve the same main subject/entity (same instance label if present). |
| - The distractors should be diverse, non-ambiguous, realistic. The difference between ground-truth and distractors should be VERY clear. |
| - No markdown, no extra keys. |
| """ |
|
|
| USER_WRONG_9 = """Situation: {situation} |
| Question: {question} |
| Ground-truth answer: {answer} |
| Correct statement: {correct_statement} |
| |
| TASK |
| Generate 9 WRONG but plausible alternative declarative statements. |
| |
| Return JSON only: |
| {{"distractors": ["...","...","...","...","...","...","...","...","..."]}} |
| """ |
|
|
| SYSTEM_WRONG_1 = r""" |
| You generate distractors for a ground-truth statement. |
| |
| CRITICAL: |
| - EVERY sentence MUST integrates the situation text with a wrong answer clause (e.g., “When …, …”). |
| |
| OUTPUT (STRICT JSON ONLY): |
| {{"distractors": ["<s1>"]}} |
| |
| Rules: |
| - The situation should be fixed in the sentence. |
| - Exactly 1 string. |
| - The sentence must be WRONG given the ground-truth answer. |
| - Preserve the same main subject/entity (same instance label if present). |
| - Make the wrong answer VERY clear (not ambiguous). |
| - No markdown, no extra keys. |
| """ |
|
|
| USER_WRONG_1 = """Situation: {situation} |
| Question: {question} |
| Ground-truth answer: {answer} |
| Correct statement: {correct_statement} |
| |
| TASK |
| Generate 1 WRONG but plausible alternative declarative statement. |
| |
| Return JSON only: |
| {{"distractors": ["..."]}} |
| """ |
|
|
| def _add_unique(collected: List[str], seen: set, candidates: List[Any], situation: str) -> None: |
| for x in candidates: |
| if not isinstance(x, str): |
| continue |
| x = x.strip() |
| if not x: |
| continue |
| x = x |
| nx = normalize_text(x) |
| if nx in seen: |
| continue |
| seen.add(nx) |
| collected.append(x) |
|
|
| def gen_wrong_sentences( |
| client: OpenAI, |
| model: str, |
| situation: str, |
| question: str, |
| answer: str, |
| correct_statement: str, |
| n: int = 9, |
| batch_rounds: int = 5, |
| ) -> List[str]: |
| correct_norm = normalize_text(correct_statement) |
|
|
| if n == 1: |
| system = SYSTEM_WRONG_1 |
| user_tmpl = USER_WRONG_1 |
| max_tokens = 256 |
| else: |
| system = SYSTEM_WRONG_9 |
| user_tmpl = USER_WRONG_9 |
| max_tokens = 1024 |
|
|
| for _ in range(batch_rounds): |
| obj = call_chat_json( |
| client=client, |
| model=model, |
| system=system, |
| user=user_tmpl.format( |
| situation=situation, |
| question=question, |
| answer=answer, |
| correct_statement=correct_statement, |
| ), |
| max_tokens=max_tokens, |
| temperature=0.95, |
| ) |
|
|
| ds = obj.get("distractors", None) |
| if not isinstance(ds, list): |
| continue |
|
|
| seen = {correct_norm} |
| out: List[str] = [] |
| for s in ds: |
| if not isinstance(s, str): |
| continue |
| s = s.strip() |
| ns = normalize_text(s) |
| if ns in seen: |
| continue |
| seen.add(ns) |
| out.append(s) |
| if len(out) >= n: |
| return out[:n] |
|
|
| raise ValueError(f"Could not collect {n} unique distractors in {batch_rounds} batch rounds.") |
|
|
| |
| |
| |
| def build_mcq(correct: str, wrongs: List[str], seed: Optional[int] = None): |
| options = [correct] + list(wrongs) |
| if seed is not None: |
| rnd = random.Random(seed) |
| rnd.shuffle(options) |
| else: |
| random.shuffle(options) |
| label = options.index(correct) |
| return options, label |
|
|
| |
| |
| |
| _WORKER_CLIENT = None |
| _WORKER_MODEL = None |
|
|
| def _init_worker(base_url: str, model_name: str, timeout: int = 3600): |
| global _WORKER_CLIENT, _WORKER_MODEL |
| _WORKER_MODEL = model_name |
| _WORKER_CLIENT = OpenAI(api_key="EMPTY", base_url=base_url, timeout=timeout) |
|
|
| def _process_one(args: Tuple[str, int, str, str, str, str]) -> Dict[str, Any]: |
| |
| key, idx, scene_id, situation, question, gt_answer = args |
| try: |
| correct_stmt = gen_statement(_WORKER_CLIENT, _WORKER_MODEL, situation, question, gt_answer) |
|
|
| |
| n_wrong = 1 if is_binary_qa(question, gt_answer) else 9 |
| wrongs = gen_wrong_sentences( |
| _WORKER_CLIENT, |
| _WORKER_MODEL, |
| situation, |
| question, |
| gt_answer, |
| correct_stmt, |
| n=n_wrong, |
| batch_rounds=6 if n_wrong == 9 else 4, |
| ) |
|
|
| seed = abs(hash(key)) % (2**32) |
| options, label = build_mcq(correct_stmt, wrongs, seed=seed) |
|
|
| out = { |
| "key": key, |
| "scene_id": scene_id, |
| "situation": situation, |
| "question": question, |
| "gt_answer": gt_answer, |
| "correct": correct_stmt, |
| "options": options, |
| "label": label, |
| "is_binary": (n_wrong == 1), |
| } |
| return {"ok": True, "out": out} |
|
|
| except Exception as e: |
| return {"ok": False, "err": { |
| "key": key, |
| "scene_id": scene_id, |
| "situation": situation, |
| "question": question, |
| "gt_answer": gt_answer, |
| "error": str(e), |
| }} |
|
|
| |
| |
| |
| def main(): |
| base_url = "http://lrc-alpha-sg-gpu06:22002/v1" |
| model_name = "Qwen/Qwen3-VL-8B-Instruct" |
|
|
| in_path = "/home/m50048399/transfered/ye_project/PointMapVerse/existing_datasets/ScanNet/annotations/sqa_task/balanced/v1_balanced_questions_test_scannetv2.json" |
| ann_path = "/home/m50048399/transfered/ye_project/PointMapVerse/existing_datasets/ScanNet/annotations/sqa_task/balanced/v1_balanced_sqa_annotations_test_scannetv2.json" |
|
|
| out_jsonl = "/home/m50048399/transfered/ye_project/PointMapVerse/derived/sqa_sentence_mcq_test_scannetv2.jsonl" |
| err_jsonl = out_jsonl + ".errors.jsonl" |
|
|
| qdata = load_json(in_path)["questions"] |
| adata = load_json(ann_path)["annotations"] |
|
|
| done = load_done_keys(out_jsonl) |
| print(f"Loaded {len(qdata)} items. Already done: {len(done)}") |
|
|
| tasks = [] |
| for idx, item in enumerate(qdata): |
| scene_id = item.get("scene_id", "") |
| question = (item.get("question") or "").strip() |
| situation = (item.get("situation") or "").strip() |
|
|
| gt_answer = "" |
| try: |
| gt_answer = (adata[idx]["answers"][0]["answer"] or "").strip() |
| except Exception: |
| gt_answer = "" |
|
|
| if not situation or not question or not gt_answer: |
| continue |
|
|
| key = f"{scene_id}::{idx}" |
| if key in done: |
| continue |
|
|
| tasks.append((key, idx, scene_id, situation, question, gt_answer)) |
|
|
| print(f"To process: {len(tasks)}") |
|
|
| ctx = mp.get_context("spawn") |
| with ctx.Pool( |
| processes=8, |
| initializer=_init_worker, |
| initargs=(base_url, model_name, 3600), |
| maxtasksperchild=50, |
| ) as pool: |
| processed = 0 |
| ok_cnt = 0 |
| err_cnt = 0 |
|
|
| for res in pool.imap_unordered(_process_one, tasks, chunksize=4): |
| processed += 1 |
| if res["ok"]: |
| append_jsonl(out_jsonl, res["out"]) |
| ok_cnt += 1 |
| else: |
| append_jsonl(err_jsonl, res["err"]) |
| err_cnt += 1 |
|
|
| if processed % 100 == 0: |
| print(f"Finished {processed}/{len(tasks)} | ok={ok_cnt} err={err_cnt}") |
|
|
| print("Done.") |
|
|
| if __name__ == "__main__": |
| main() |
|
|