Skip to content

Conversation

@georgedouzas
Copy link

Description

This PR adds minimal support for the unsloth library without making it a required dependency of Oumi. The idea is to let users load models and tokenizers through Unsloth’s API if it’s available, without changing the trainer abstraction or adding a new TrainerType. Since Unsloth patches transformers and trl under the hood, it can work with the existing setup without needing a brand-new trainer.

Feedback on whether this approach makes sense would be appreciated, or let me know if you would prefer a more explicit integration — like introducing a dedicated trainer type for Unsloth even though it’s patching everything internally. There are also some dependency conflicts between Unsloth and Oumi at the moment.

Related issues

#1322

Before submitting

  • This PR only changes documentation. (You can ignore the following checks in that case)
  • Did you read the contributor guideline Pull Request guidelines?
  • Did you link the issue(s) related to this PR in the section above?
  • Did you add / update tests where needed?

Reviewers

At least one review from a member of oumi-ai/oumi-staff is required.

@georgedouzas georgedouzas force-pushed the georgedouzas/unsloth-fine-tuning-support branch from f4832e3 to acf918d Compare April 27, 2025 18:39
@georgedouzas georgedouzas force-pushed the georgedouzas/unsloth-fine-tuning-support branch from acf918d to 7788ed7 Compare April 28, 2025 18:35
@georgedouzas georgedouzas force-pushed the georgedouzas/unsloth-fine-tuning-support branch from 7788ed7 to a34a3ea Compare May 1, 2025 18:10
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant