-
Notifications
You must be signed in to change notification settings - Fork 230
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Create unified script and workflow for llama-fast models validationin #94
Conversation
4e3c734
to
8a57c22
Compare
eb57f3f
to
d949b12
Compare
It's not fully ready for review as the newly added models will require running the hg_converter first, but the CI part should be okay. My main goal is to run model valdiation jobs on macOS x86 (supported), macOS M1/ARM (supportted), Linux x86 (supported), Linux ARM (?). @huydhn @kit1980 I know pytorch test-infra has lots of runners, but not sure if those are self-hosted and whether we can use one here. Also is it possible if we'd like to run the Android/iOS emulator on an host platform for the correctness validation? |
9c19e53
to
8b0b17a
Compare
The inference error using ExecuTorch is fixed in #122 |
The macos x86 issue is same as what Scott reported early when build ExecuTorch. Let's disable it in this PR and move on. Will have to re-enable it once the issue is cleared. |
To run the workflow across different platforms for supported models, simply run the top-level script
bash ./scripts/workflow.sh [cpu | cuda]
, which relies on bunch of scripts under.ci/scripts
. So CI and local dev are running pretty much same scripts. For the platforms that are not supported in GitHub CI, we can manually run the script on the target platforms