Agent Skills: Runtime Behavior Probe

Plan and execute runtime-behavior investigations with temporary probe scripts, validation matrices, state controls, and findings-first reports. Use only when the user explicitly invokes this skill to verify actual runtime behavior beyond normal code-level checks, especially to uncover edge cases, undocumented behavior, or common failure modes in local or live integrations. A baseline smoke check is fine as an entry point, but do not stop at happy-path confirmation.

UncategorizedID: openai/openai-agents-python/runtime-behavior-probe

Install this agent skill to your local

pnpm dlx add-skill https://github.com/openai/openai-agents-python/tree/HEAD/.agents/skills/runtime-behavior-probe

Skill Files

Browse the full folder contents for runtime-behavior-probe.

Download Skill

Loading file tree…

.agents/skills/runtime-behavior-probe/SKILL.md

Skill Metadata

Name
runtime-behavior-probe
Description
Plan and execute runtime-behavior investigations with temporary probe scripts, validation matrices, state controls, and findings-first reports. Use only when the user explicitly invokes this skill to verify actual runtime behavior beyond normal code-level checks, especially to uncover edge cases, undocumented behavior, or common failure modes in local or live integrations. A baseline smoke check is fine as an entry point, but do not stop at happy-path confirmation.

Runtime Behavior Probe

Overview

Use this skill to investigate real runtime behavior, not to restate code or documentation. Start by planning the investigation, then execute a case matrix, record observed behavior, and report both the findings and the method used to obtain them.

Core Rules

  • Treat this skill as manual-only. Do not rely on implicit invocation.
  • A baseline success or smoke case is often the right entry point, but do not stop there when the real question involves edge cases, drift, or failure behavior.
  • Plan before running anything. Write the case matrix first, then fill it in with observed results. The matrix can live in a scratch note, a temporary file, or the probe script header.
  • Default to local or read-only probes. Consider a live service only when it is clearly relevant, then apply the lightweight gates below before you run it.
  • Size the probe to the decision. Start with the smallest matrix that can disqualify or validate the current hypothesis, then expand only when uncertainty remains.
  • Before a live probe, apply three lightweight gates:
    • Destination gate. Use only a live destination that is clearly allowed for the task.
    • Intent gate. Run the live probe only when the user explicitly wants runtime verification on that integration, or explicitly approves it after you propose the probe.
    • Data gate. If the probe will read environment variables, mutate remote state, incur material cost, or exercise non-public or user data, name the exact variable names or data class and get explicit approval first.
  • Classify each case as read-only, mutating, or costly before execution. For mutating or costly cases, or for any live case that will read environment variables, define cleanup or rollback before running the probe.
  • Use temporary files or a temporary directory for one-off probe scripts.
  • Keep temporary artifacts until the final response is drafted. Then delete them by default unless the user asked to keep them or they are needed for follow-up. Even when artifacts are deleted, keep a short run summary of the command shape, runtime context, and artifact status in the report.
  • Before executing a live probe that will read environment variables, tell the user the exact variable names you plan to use and why, then wait for explicit approval. Examples include OPENAI_API_KEY and other expected default names for the system under test.
  • Never print secrets, even when they come from standard environment variables that this skill may use.
  • For OpenAI API or OpenAI platform probes in this repository, use $openai-knowledge early to confirm contract-sensitive details such as supported parameters, field names, and limits. Use runtime probing to validate or challenge the documented behavior, not to skip the documentation pass entirely. If the docs MCP is unavailable, fall back to the official OpenAI docs and say that you used the fallback in the report.
  • For benchmark or comparison probes, make parity explicit before execution. Record what is held constant, what variable is under test, which response-shape constraints keep the comparison fair, and any usage or token counters that matter for interpreting latency or cost.
  • For OpenAI hosted tool probes, remove setup ambiguity before attributing a negative result to runtime behavior:
    • Force the tool path with the matching tool_choice when the question depends on tool invocation.
    • Treat container_auto and container_reference as separate cases, not interchangeable setup details.
    • Clear unsupported model or tool options first so they do not invalidate the probe.

Workflow

  1. Restate the investigation target in operational terms. Name the runtime surface, the key uncertainty, and the highest-risk behaviors to test.
  2. Do a short preflight. Check the relevant code or docs first, decide whether the question needs local or live validation, and note any repo, baseline, or release boundary that matters.
  3. Create a validation matrix before executing probes. Cover both baseline behavior and the most relevant failure or drift cases. The matrix can live in a scratch note, a temporary file, or a structured header inside the probe script.
  4. For each case, choose an execution mode up front:
    • single-shot for deterministic one-run checks.
    • repeat-N for cache, retry, streaming, interruption, rate-limit, concurrency, or other run-to-run-sensitive behavior.
    • warm-up + repeat-N when first-run cold-start effects could distort the result. Use these defaults unless the task clearly needs something else:
    • Quick screen of a repeat-sensitive question: repeat-3.
    • Decision-grade latency or release recommendation: warm-up + repeat-10.
    • Costly live cases: start at repeat-3, then expand only if the answer remains unclear. If it is genuinely unclear whether extra runs are worth the time or cost, ask the user before expanding the probe.
  5. When the question is benchmark-like or comparative, run in phases. Start with a high-signal pilot matrix against a control, then expand only the surviving candidates or unresolved cases.
  6. If the question is about a suspected regression or behavior change, add at least one known-good control case such as origin/main, the latest release, or the same request without the suspected option.
  7. For comparative probes, define parity before execution. Record prompt or input shape, tool-choice setup, model-settings parity, state reuse rules, and any response-shape constraint that keeps the comparison fair. If materially different output length could bias the result, record usage or token notes too.
  8. If the question asks whether one option has the same intelligence or quality as another, decide whether the matrix supports only example-pattern parity or a broader quality claim. For broader claims, add at least one harder or more open-ended case. Otherwise say explicitly that the result is limited to the covered patterns.
  9. Plan state controls before execution when hidden state could affect the result. Record whether each case uses fresh or reused state, how cache reuse or cache busting is handled, what unique IDs isolate repeated runs, and how cleanup is verified.
  10. If any live case will read environment variables, list the exact variable names and purpose for each case, then ask the user for approval before execution. Keep the approval ask short and include destination, read-only versus mutating or costly risk, exact variable names, and cleanup or rollback if relevant.
  11. Build task-specific probe scripts in a temporary location. Keep the script small, observable, and easy to discard.
  12. In openai-agents-python, make the runtime context explicit:
  • Run Python probes from the repository root with uv run python when practical.
  • Record the current commit, working directory, Python executable, and Python version.
  • Avoid accidental imports from a different checkout or site-packages location. If you must deviate from uv run python, say exactly why and what interpreter or environment was used instead.
  1. Execute the matrix and capture evidence. Record request shape, setup, observation summary, unexpected or negative result, error details, timing, runtime context, approved environment-variable names, repeat counts, warm-up handling, variance when relevant, cleanup behavior, and for comparisons note what was held constant plus any response-shape or usage notes that affect interpretation.
  2. Update the matrix with actual outcomes, not guesses.
  3. Keep temporary artifacts until the final response is drafted. Then delete them unless the user asked to keep them or they are needed for follow-up. Benchmark and repeat-heavy probes often need follow-up, so keeping artifacts is normal when the result may be revisited. If deleted, retain and report a short run summary.
  4. Report findings first, with unexpected or negative findings first. Then summarize how the validation was performed and which cases were covered.
  5. If the probe isolates one clear defect, you may include a short implementation hypothesis or minimal repro direction. Do not expand into a larger next-step plan unless the user asked for it.

Validation Matrix

Use a matrix that makes the news easy to scan. Start from the runtime question and the observation summary, not just from expected and pass or fail.

Use a matrix with at least these columns:

  • case_id
  • scenario
  • mode
  • question
  • setup
  • observation_summary
  • result_flag
  • evidence

Add these columns when they materially improve the investigation:

  • comparison_basis
  • variable_under_test
  • held_constant
  • output_constraint
  • status
  • confidence
  • state_setup
  • repeats
  • warm_up
  • variance
  • usage_note
  • risk_profile
  • env_vars
  • approval
  • control

Treat result_flag as a fast scan field such as unexpected, negative, expected, or blocked. Use status only when there is a credible comparison basis, baseline, or documented contract to compare against.

Always consider whether the matrix should include these categories:

  • Baseline success.
  • Control or baseline comparison when a regression is suspected.
  • Boundary input or parameter variation.
  • Invalid or unsupported input.
  • Missing or incorrect configuration.
  • Transient external failure such as timeout, network interruption, or rate limiting.
  • Retry, idempotence, or cleanup behavior.
  • Concurrency or overlapping operations when shared state or ordering may matter.
  • Open-ended quality or intelligence samples when the question is broader than pattern parity.

Open validation-matrix.md when you need a stronger prioritization model or a reusable case template.

Temporary Probe Scripts

Write one-off scripts in a temporary file or temporary directory such as one created by mktemp -d or Python tempfile. Keep the script outside the repository by default, even when it imports code from the repository.

If the probe needs repository code:

  • Run it with the repository as the working directory, or
  • Set PYTHONPATH or the equivalent import path explicitly.
  • In openai-agents-python, prefer uv run python /tmp/probe.py from the repository root.

Design the probe to maximize observability:

  • Print or log the exact scenario being exercised.
  • Capture runtime context such as git SHA, working directory, Python executable and version, relevant package versions, model or deployment name, endpoint or base URL alias, and any retry or tool options that materially affect behavior.
  • For live probes, record only the names of environment variables that were approved for use. Never print their values.
  • Capture structured outputs when possible.
  • Preserve raw error type, message, and status code.
  • For repeat-sensitive cases, capture the attempt index, warm-up status, and any stable identifiers that help compare runs.
  • For repeated or benchmark-style probes, write both raw results and a compact summary artifact when practical.
  • Keep branching minimal so each script answers a narrow question.

Before deleting the temporary script or directory, keep a short run summary of the script path, command used, runtime context, and whether the evidence was kept or deleted.

Open python_probe.py when you want a lightweight disposable Python probe scaffold.

Reporting

Report in this order:

  1. Findings. Put unexpected or negative findings first. If there was no real news, say that explicitly.
  2. Validation approach. Summarize the code used, the runtime surface exercised, the execution modes, and the case matrix coverage.
  3. Case results. Include the matrix or a condensed version of it when the case count is large.
  4. Artifact status and brief run summary. State whether temporary artifacts were deleted or kept, and provide kept paths or the retained summary.
  5. Optional implementation note. Include this only when one clear defect was isolated and a short implementation direction would help.

For comparative probes, the report should also say what was held constant, what variable was under test, and whether the result supports only pattern parity or a broader quality claim.

Open reporting-format.md for the recommended response template.

Resources