Skip to content

Conversation

@robtinn
Copy link
Contributor

@robtinn robtinn commented Nov 24, 2025

Summary

  • Adding in some functions to generate synthetic data in this example as examples from HealthBench might not always be suitable for fine-tuning.
  • Also tidied up some of the wording and visualization notebook cells.

Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

Codex can also answer questions or update the PR. Try commenting "@codex address that feedback".

Comment on lines +273 to 277
" \"content\": f\"Produce twenty examples.\"\n",
" }\n",
" ],\n",
"text_format=SyntheticData\n",
")\n",

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P1 Badge responses.parse called with unsupported argument

The new synthetic-data cell calls client.responses.parse(..., text_format=SyntheticData) but the responses API expects a response_format argument for Pydantic parsing; text_format is not a valid parameter and will raise a TypeError before any data is generated. With this change, the notebook cannot produce the synthetic train/val/test JSONL files and the fine-tuning walk-through stops at this cell.

Useful? React with 👍 / 👎.

@robtinn robtinn merged commit f249969 into main Nov 24, 2025
1 check passed
@robtinn robtinn deleted the rtinn/rft_updates branch November 24, 2025 09:40
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants