STET

task_detail

validation/gpt-5-3-codex/flux-pr-1241/task_detail.v1.json

16805 bytes

Back to adjudication
{
  "version": "v1",
  "materialization": {
    "mode": "canonical"
  },
  "identity": {
    "variant_id": "gpt-5.3-codex",
    "requested_model": "gpt-5.3-codex",
    "model_key": "gpt-5-3-codex",
    "run_id": "2026-02-28__13-38-10__gpt-5-3-codex",
    "task_id": "flux-pr-1241",
    "trial_name": "flux-pr-1241.1-of-1.2026-02-28__13-38-10__gpt-5-3-codex"
  },
  "decision": {
    "status": "pass",
    "primary_signal": "tests",
    "matrix_status": "pass_with_warn",
    "reason_codes": [
      "primary_tests",
      "tests_passed",
      "equivalence_non_equivalent",
      "review_fail"
    ],
    "flags": [
      "equiv_warn",
      "review_warn"
    ],
    "publish": {
      "include_in_leaderboard": true,
      "weak_signal_risk": false
    },
    "rescue": {
      "candidate": false,
      "eligible": false,
      "decision": "not_candidate",
      "required_artifacts_ok": true
    }
  },
  "tests": {
    "outcome": "pass",
    "strength": "strong",
    "commands": [
      "go test -C v2 ./... -count=1 -timeout=300s"
    ],
    "command_results": [
      {
        "command": "go test -C v2 ./... -count=1 -timeout=300s",
        "gold_outcome": "pass",
        "gold_tb_trial_name": "flux-pr-1241.1-of-1.validate-gold-1772290480654267000",
        "gold_failure_mode": "unset",
        "agent_outcome": "pass",
        "agent_tb_trial_name": "flux-pr-1241.1-of-1.validate-override-1772290528319228000",
        "agent_failure_mode": "unset",
        "agent_retried": false
      }
    ],
    "partial_score": {
      "score": 1,
      "numerator": 1,
      "denominator": 1,
      "level": "command",
      "provenance": "fallback_command_level",
      "reason": "test_case_detail_unavailable"
    },
    "tb": {
      "trial_name": "flux-pr-1241.1-of-1.2026-02-28__13-38-10__gpt-5-3-codex",
      "is_resolved": true,
      "failure_mode": "unset",
      "duration_sec": 101.604170917,
      "time_budget_sec": 1200
    },
    "cache": {
      "cache_hit": false,
      "cache_miss_reason": "missing_pinned_dependencies",
      "environment_group_id": "eg_46b6a6c82a4890b6e6e87c39575fd02c"
    }
  },
  "quality": {
    "equivalence": {
      "status": "used",
      "outcome": "non_equivalent",
      "confidence": 0.89,
      "issue_type": "behavioral",
      "notes": "The patch adds limit checks in parser traversal (`ParseWithOptions`) rather than in tokenization, so tokenization still processes full documents without limits. It also introduces a different API/contract (no `ParseWithLimits(limits) (stats, error)` and no tokenizer-level limit errors), which misses a core part of the intended change and likely expected behavior."
    },
    "code_review": {
      "status": "used",
      "signal": "fail",
      "confidence": 0.87,
      "issue_type": "unclear",
      "source_field": "code_review",
      "dimension_scores": {
        "correctness": {
          "score_0_4": 1,
          "confidence_0_1": 0.89,
          "rationale": "The patch implements a different API/flow than the requested change: it adds parser-level `ParseWithOptions`/`ParseOptions` instead of tokenizer-level limits and errors (`ParseWithLimits`, tokenizer stats/errors), so it likely misses intended behavior and compatibility expectations."
        },
        "edge_case_handling": {
          "score_0_4": 2,
          "confidence_0_1": 0.76,
          "rationale": "It handles fragment spreads and accumulates depth per executable definition, but enforcement is tied to parser traversal and not tokenizer streaming limits, so extreme token streams remain an uncovered edge case."
        },
        "introduced_bug_risk": {
          "score_0_4": 1,
          "confidence_0_1": 0.83,
          "rationale": "Risk is high because limits are checked only during parse callbacks, after full tokenization, which does not provide early tokenizer cutoff against pathological inputs."
        },
        "maintainability_idioms": {
          "score_0_4": 2,
          "confidence_0_1": 0.72,
          "rationale": "The code is readable, but introduces parallel option/stats APIs in both `pkg` and `v2` parser layers rather than the focused tokenizer extension, increasing divergence from expected architecture."
        }
      },
      "findings": [
        {
          "id": "f1",
          "dimension": "correctness",
          "severity": "major",
          "title": "Implements different public API than requested tokenizer-based limits",
          "detail": "The patch introduces `ParseOptions`/`ParseWithOptions` and parser-side stats, but does not add tokenizer limit types/errors or a `ParseWithLimits` tokenizer-integrated path, so expected API/behavior from the task is likely missing.",
          "confidence_0_1": 0.91,
          "citations": [
            {
              "path": "v2/pkg/astparser/parse_options.go",
              "side": "new",
              "start_line": 1,
              "end_line": 14,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            },
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 58,
              "end_line": 84,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            }
          ]
        },
        {
          "id": "f2",
          "dimension": "introduced_bug_risk",
          "severity": "major",
          "title": "Depth/field limits are enforced after tokenization, not during tokenization",
          "detail": "The parser calls `tokenize()` unconditionally before limit checks in `parseSelectionSet`/`parseField`; this misses the intended early rejection during tokenization for DoS mitigation.",
          "confidence_0_1": 0.88,
          "citations": [
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 116,
              "end_line": 128,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            },
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 1412,
              "end_line": 1419,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            },
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 1503,
              "end_line": 1510,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            }
          ]
        },
        {
          "id": "f3",
          "dimension": "correctness",
          "severity": "major",
          "title": "Error surface differs from expected typed limit errors",
          "detail": "Instead of returning typed tokenizer errors, the patch writes generic external report messages (`parser depth/field limit exceeded`), which can break callers/tests expecting specific limit error types/messages.",
          "confidence_0_1": 0.8,
          "citations": [
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 157,
              "end_line": 171,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            },
            {
              "path": "v2/pkg/astparser/parser.go",
              "side": "new",
              "start_line": 186,
              "end_line": 200,
              "anchor_kind": "diff",
              "head_sha": "723be02771e716060c49a769e1e1cee215c0536b"
            }
          ]
        }
      ],
      "overall": {
        "confidence_overall": 0.87,
        "summary": "The patch partially addresses complexity limits but diverges materially from the intended tokenizer-level design and API, so it is unlikely to fully satisfy the requested change despite passing the provided test command."
      },
      "notes": "The patch partially addresses complexity limits but diverges materially from the intended tokenizer-level design and API, so it is unlikely to fully satisfy the requested change despite passing the provided test command."
    },
    "behavioral_robustness": {
      "status": "used",
      "probe_status": "used",
      "gold_pass_candidate_pass_count": 1,
      "agreement_rate": 1,
      "coverage_delta_status": "unavailable",
      "coverage_delta_reason": "coverage_adapter_unavailable",
      "mutation_lite_status": "unavailable",
      "mutation_lite_reason": "mutation_runner_unavailable",
      "notes": "derived from command-level outcomes; no accepted generated probes"
    },
    "footprint_risk": {
      "status": "used",
      "reason": "none",
      "version": "v1",
      "level": "medium",
      "score": 0.6211208909896463,
      "flag": false,
      "severe_flag": false
    },
    "cost": {
      "status": "tokens_only",
      "token_status": "present",
      "cache_token_status": "present",
      "total_input_tokens": 1544316,
      "total_output_tokens": 13775,
      "total_tokens": 1558091,
      "uncached_input_tokens": 68732,
      "cache_read_input_tokens": 1475584,
      "cached_input_tokens": 1475584,
      "total_cost_usd": 4.070856,
      "token_source": "openai_cached_tokens_usage",
      "pricing_model_key": "gpt-5.3-codex",
      "pricing_mode": "cache_aware"
    }
  },
  "summary": {
    "task": {
      "matrix_status": "pass_with_warn",
      "tests_outcome": "pass",
      "tests_unknown_cause": null,
      "lane_report_lane": null,
      "lane_report_source": null,
      "lane_report_reasons": null,
      "cache_hit": false,
      "cache_miss_reason": "missing_pinned_dependencies",
      "setup_ms_saved": null,
      "pinned_bytes": null,
      "environment_group_id": "eg_46b6a6c82a4890b6e6e87c39575fd02c",
      "equivalence_status": "used",
      "equivalence_outcome": "non_equivalent",
      "code_review_status": "used",
      "code_review_signal": "fail",
      "behavioral_robustness_status": "used",
      "coverage_delta_status": "unavailable",
      "mutation_lite_status": "unavailable",
      "probe_accepted_candidates": null,
      "probe_accepted_commands": null,
      "probe_agreement_rate": 1,
      "probe_gold_pass_candidate_pass_count": 1,
      "probe_gold_pass_candidate_fail_count": null,
      "probe_review_required_count": null,
      "flags": [
        "equiv_warn",
        "review_warn"
      ],
      "tb_is_resolved": true,
      "tb_failure_mode": "unset",
      "tb_total_input_tokens": 1544316,
      "tb_total_output_tokens": 13775,
      "tb_total_tokens": 1558091,
      "tb_uncached_input_tokens": 68732,
      "tb_cache_creation_input_tokens": null,
      "tb_cache_read_input_tokens": 1475584,
      "tb_cached_input_tokens": 1475584,
      "token_status": "present",
      "cache_token_status": "present",
      "token_source": "openai_cached_tokens_usage",
      "task_cost": 4.070856,
      "cost_status": "present",
      "cost_pricing_mode": "cache_aware",
      "pricing_model_key": "gpt-5.3-codex",
      "equiv_rescue_policy": "on",
      "rescue_candidate": false,
      "rescue_eligible": false,
      "rescue_decision": "not_candidate",
      "publish_include_in_leaderboard": true,
      "publish_exclusion_reasons": [],
      "publish_weak_signal_risk": false,
      "tests_only_outcome": 1,
      "rescue_aware_outcome": 1,
      "partial_score": 1,
      "partial_score_numerator": 1,
      "partial_score_denominator": 1,
      "partial_score_level": "command",
      "partial_score_provenance": "fallback_command_level",
      "partial_score_reason": "test_case_detail_unavailable",
      "partial_score_unknown_count": 0,
      "footprint_risk_status": "used",
      "footprint_risk_reason": "none",
      "footprint_risk_level": "medium",
      "footprint_risk_score": 0.6211208909896463,
      "footprint_risk_flag": false,
      "footprint_risk_severe_flag": false
    },
    "run": {
      "model": "gpt-5.3-codex",
      "requested_model": "gpt-5.3-codex",
      "run_id": "2026-02-28__13-38-10__gpt-5-3-codex",
      "passRate": 0.9310344827586207,
      "ciLow": 0.8275862068965517,
      "ciHigh": 1,
      "effectiveN": 29,
      "tier": 1,
      "validation_counts": {
        "fail_guardrail": 0,
        "fail_high_conf": 1,
        "fail_infra": 0,
        "fail_likely_equiv": 1,
        "fail_with_diag": 0,
        "missing": 0,
        "pass": 6,
        "pass_with_warn": 21
      }
    },
    "run_metadata": {
      "agent_name": "codex",
      "model_name": "gpt-5.3-codex",
      "start_time": "2026-02-28T13:38:12.733292+00:00",
      "end_time": "2026-02-28T14:43:03.470912+00:00",
      "commit_hash": "99ae882e59281de74856a3e7eeb1b7c986b466af",
      "accuracy": 0.896551724137931,
      "reasoning_effort": "medium"
    }
  },
  "evidence": {
    "task_detail": {
      "path": "validation/gpt-5-3-codex/flux-pr-1241/task_detail.v1.json",
      "exists": true
    },
    "trajectory": {
      "path": "validation/gpt-5-3-codex/flux-pr-1241/trajectory.v1.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 15254
    },
    "validation": {
      "path": "validation/gpt-5-3-codex/flux-pr-1241/validation.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 160905
    },
    "results": {
      "path": "runs/2026-02-28__13-38-10__gpt-5-3-codex/results.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 62248
    },
    "run_metadata": {
      "path": "runs/2026-02-28__13-38-10__gpt-5-3-codex/run_metadata.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 1634
    },
    "agent_patch": {
      "path": "runs/2026-02-28__13-38-10__gpt-5-3-codex/flux-pr-1241/flux-pr-1241.1-of-1.2026-02-28__13-38-10__gpt-5-3-codex/agent-logs/agent.patch",
      "exists": true,
      "non_empty": true,
      "size_bytes": 18964
    },
    "summary": {
      "path": "reports/summary.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 269968
    },
    "manifest": {
      "path": "manifest.json",
      "exists": true,
      "non_empty": true,
      "size_bytes": 677
    },
    "validation_inputs": {
      "required_for_rescue_ok": true,
      "gold_patch": {
        "exists": true,
        "non_empty": true,
        "size_bytes": 6312,
        "sha256": "d6f21e8593865339582d014b1e0e6843ce77f999eb6b3b68430b76e0d9bac323",
        "required_for_rescue": true
      },
      "repo_tarball": {
        "exists": true,
        "non_empty": true,
        "size_bytes": 4319535,
        "sha256": "37aee773dfb332627541f509062e33041fb9fb2ecbd45f46b13c5c83275c6873",
        "required_for_rescue": true
      },
      "agent_patch": {
        "path": "runs/2026-02-28__13-38-10__gpt-5-3-codex/flux-pr-1241/flux-pr-1241.1-of-1.2026-02-28__13-38-10__gpt-5-3-codex/agent-logs/agent.patch",
        "exists": true,
        "non_empty": true,
        "size_bytes": 18964,
        "sha256": "6032c26ead7739c265d1c0c1ad09b053c100b0aaf0525b81cacd674c25576ea2",
        "required_for_rescue": true
      }
    }
  },
  "provenance": {
    "identity": [
      {
        "source": "manifest",
        "available": true
      },
      {
        "source": "tb_results",
        "available": true
      }
    ],
    "decision": [
      {
        "source": "validation",
        "available": true
      }
    ],
    "tests": [
      {
        "source": "validation",
        "available": true
      },
      {
        "source": "tb_results",
        "available": true
      }
    ],
    "quality": {
      "equivalence": [
        {
          "source": "validation",
          "available": true
        }
      ],
      "code_review": [
        {
          "source": "validation",
          "available": true
        }
      ],
      "behavioral_robustness": [
        {
          "source": "validation",
          "available": true
        }
      ],
      "footprint_risk": [
        {
          "source": "validation",
          "available": true
        }
      ],
      "cost": [
        {
          "source": "validation",
          "available": true
        },
        {
          "source": "summary",
          "available": true
        }
      ]
    },
    "summary": {
      "task": [
        {
          "source": "summary",
          "available": true
        }
      ],
      "run": [
        {
          "source": "summary",
          "available": true
        }
      ],
      "run_metadata": [
        {
          "source": "run_metadata",
          "available": true
        }
      ]
    },
    "evidence": [
      {
        "source": "filesystem",
        "available": true
      }
    ]
  }
}