"thinking" string.
This gives you the accuracy benefits of reasoning while keeping outputs auditable, compact, and machine-parseable.
Use case
You need explainable classifications for internal reviewers. Each classification should include the specific evidence from the input that drove the decision, plus a short summary a reviewer can scan in seconds.Schema pattern
Prompt snippet
Example output
Why this works
Theevidence array (1-3 items, each 10-140 characters) forces the model to ground its reasoning in specific observations from the input, not generate vague explanations. The length bounds prevent evidence bullets from becoming mini-essays.
decision_summary (20-240 characters) gives reviewers a one-sentence rationale without scrolling. Combined with confidence, it lets your application route low-confidence decisions to human review while auto-approving high-confidence ones.
The fixed property order also helps. By consistently generating label, then evidence, then decision_summary, then confidence, the model follows a predictable structure instead of deciding its own output sequence on the fly.
Because all reasoning fields are bounded and typed, you can store them in structured logs, aggregate them in dashboards, and search across them. None of that is practical with freeform reasoning text.