Skip to content

Conversation

@michael-L-i
Copy link

Code tailored to SFT training on Dipam's dataset using Qwen3-1.7B. The Colab is a similar replica to the Qwen SFT example already present, but adjusted for our chess case.

After installing dependencies, simply run the same torchrun command pointing to the finetune_chess_model.py file. In that file, you can control the train/eval dataset accordingly.

This pipeline I verified by training a small 5000 position dataset on a trn1.2xlarge instance with relatively good results already.

@EmilyWebber
Copy link
Contributor

Awesome, thanks Mike!

@EmilyWebber EmilyWebber requested a review from jlonge4 November 18, 2025 17:56
Copy link
Contributor

@jlonge4 jlonge4 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The one issue is that the max_model_len from the compilation and inference should match (512!=2048). Otherwise LGTM!

Copy link
Contributor

@EmilyWebber EmilyWebber left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey Mike - we need the max_model_len from the compilation and inference to match (512!=2048)

@michael-L-i michael-L-i changed the title init code for SFT on Dipam's dataset formatted for model input SFT pipeline integrated for aicrowd/ChessExplained dataset Nov 19, 2025
@EmilyWebber EmilyWebber merged commit b95f725 into aws-neuron:main Nov 20, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants