-
Notifications
You must be signed in to change notification settings - Fork 26
⏪ revert back to 3-layer micro model #497
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
Signed-off-by: Prashant Gupta <[email protected]>
bot:test |
|
||
- name: "Download HF models" | ||
if: ( steps.changed-src-files.outputs.any_changed == 'true' && steps.cache_restore.outputs.cache-hit != 'true' ) | ||
if: ( steps.changed-src-files.outputs.any_changed == 'true' && steps.cache_restore.outputs.cache-hit == 'true' ) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
temporary change to get GH to download the older revision
FYI bot test seems to be fine with the 3 layer model! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If it is just tiny granite that we need to change revisions for (occasionally), then your quick fix here is probably sufficient -- after merging your PR and deleting the old cache blobs
|
||
- name: "Download HF models" | ||
if: ( steps.changed-src-files.outputs.any_changed == 'true' && steps.cache_restore.outputs.cache-hit != 'true' ) | ||
if: ( steps.changed-src-files.outputs.any_changed == 'true' && steps.cache_restore.outputs.cache-hit == 'true' ) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This means you will have both revisions:
- 4 layer model from the cache
- 3 layer model downloaded with revision
Making all the tests work. If you bypass the cache, I reckon some tests will break like in PR #499
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Once the 3 layer model is verified, I don't think we need the 4 layer anymore, so we should only be using one revision of the model at all places
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ideally the fix has to be that we delete the old cache and populate it with the 3 layer revision and merge this PR right?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it needs more code/test changes to make sure we don't use the latest/main revision of the tiny granite model anymore/anywhere.
Kicking of a test run here that is not also still using the old cached model: #502
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yup:
E huggingface_hub.errors.LocalEntryNotFoundError: Cannot find an appropriate cached snapshot folder for the specified revision on the local disk and outgoing traffic has been disabled. To enable repo look-ups and downloads online, pass 'local_files_only=False' as input.
head_call_error = OfflineModeIsEnabled("Cannot access file since 'local_files_only=True' as been set. (repo_id: ibm-ai-platform/micro-g3.3-8b-instruct-1b, repo_type: model, revision: main, filename: config.json)") force_download = False, local_files_only = True
Also, need to not use the same revision key to download the FP8 model 🙄
ValueError: Unrecognized model in ibm-ai-platform/micro-g3.3-8b-instruct-1b-FP8
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Description
Revert back to 3-layer micro model.
Needs a fix from @ckadner (#499) where the HF models cache also contains the revision so that we can cache the right model.