-
Notifications
You must be signed in to change notification settings - Fork 12
[ModelSuite] Add Toy Models #183
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Summary: Here we introduce model suite (model.py). The idea here to start and codify the ideas from jiannanWang/BackendBenchExamples. Specifically this PR adds some example models / configs which are to be loaded + a Readme. (It may be useful to look at the PR above this as well since it's the model loading logic). This PR adds two toy models to model suite SmokeTestModel - This is simple model that uses aten.ops.mm as we can implement a correct version of this op ToyCoreOpsModel - This is a model which explicitly calls the backwards passes which are both in torchbench + core. Test Plan: the test infra is in the pr above, so tests passing on the PR above should be sufficient here ### Future work with Model Suite #181
Should we really be having model definitions in this repo? It seems like this is very much the scope of projects like torchbench that are much further along |
I was thinking maybe we can do something similar to this: Instead of saving the model files, we can import them from HF. |
Longer form answer from internal chat I guess your code looks fine but it's solving a very tangential problem in my mind to your stated goals in the proposal of solving training support in backendbench. I don't think it's a good idea for us to get into the model zoo business, there are better repos suited for this like torchbench. Torchbench has other problems like it's not particularly up to date so if we want yet another zoo that's the problem I'd like to see solved head on Ideally our scope should be focused on kernel binding and correctness checks and we've been heavily reliant on other repos to determine what is important whether that's OpInfo or torchbench - it's totally valid to say I think those projects have deficiencies we'd like to see solved but then that's a very separate discussion And a more minimal PR we could merge fast would be what we need to run correctness checks on backwards passes using the same input/output pair idea we've been using already |
Hi @PaliC! Thank you for your pull request. We require contributors to sign our Contributor License Agreement, and yours needs attention. You currently have a record in our system, but the CLA is no longer valid, and will need to be resubmitted. ProcessIn order for us to review and merge your suggested changes, please sign at https://code.facebook.com/cla. If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA. Once the CLA is signed, our tooling will perform checks and validations. Afterwards, the pull request will be tagged with If you have received this in error or have any questions, please contact us at cla@meta.com. Thanks! |
Summary:
Here we introduce model suite (model.py). The idea here to start and codify the ideas from jiannanWang/BackendBenchExamples. Specifically this PR adds some example models / configs which are to be loaded + a Readme. (It may be useful to look at the PR above this as well since it's the model loading logic).
This PR adds two toy models to model suite
SmokeTestModel - This is simple model that uses aten.ops.mm as we can implement a correct version of this op
ToyCoreOpsModel - This is a model which explicitly calls the backwards passes which are both in torchbench + core.
Test Plan:
the test infra is in the pr above, so tests passing on the PR above should be sufficient here
Future work with Model Suite
#181
Stack created with Sapling. Best reviewed with ReviewStack.