-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[VTA][Relay] Extending Vision model coverage compilation for VTA #3740
Conversation
@@ -61,6 +62,8 @@ | |||
# Make sure that TVM was compiled with RPC=1 | |||
assert tvm.module.enabled("rpc") | |||
|
|||
# Increase python recursion limit to traverse Relay program | |||
sys.setrecursionlimit(10000) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@jroesch is there a better place to put this line?
@tmoreau89 maybe we should set limit in tvm init.py |
@MarisaKirisame If we set it in init.py, then do we not revert it back to its original value?? |
@tmoreau89 why revert at all? tvm is supposed to be used with a high recursion limit. if user wish otherwise she should pick a better constant herself. |
Ok, which init.py should we include the lines in? the top level? |
@tmoreau89 I think relay is fine for now. If tvm also need it, we can move it up lateron. |
…che#3740) * adding support for graphpack over multiply op * increasing resnet model coverage * fix indentation * lint * moving recursion limit fix into graphpack pass * moving recursionlimit to relay init * pooling on NCHWnc format * adding more models * deploy_resnet_on_vta.py * trailing line * generalizing to vision models * merge conflicts * fix, apply quantization to VTA only * improving comments * trimming models that have runtime issues for the moment * lint * lint * lint
…che#3740) * adding support for graphpack over multiply op * increasing resnet model coverage * fix indentation * lint * moving recursion limit fix into graphpack pass * moving recursionlimit to relay init * pooling on NCHWnc format * adding more models * deploy_resnet_on_vta.py * trailing line * generalizing to vision models * merge conflicts * fix, apply quantization to VTA only * improving comments * trimming models that have runtime issues for the moment * lint * lint * lint
…che#3740) * adding support for graphpack over multiply op * increasing resnet model coverage * fix indentation * lint * moving recursion limit fix into graphpack pass * moving recursionlimit to relay init * pooling on NCHWnc format * adding more models * deploy_resnet_on_vta.py * trailing line * generalizing to vision models * merge conflicts * fix, apply quantization to VTA only * improving comments * trimming models that have runtime issues for the moment * lint * lint * lint
…che#3740) * adding support for graphpack over multiply op * increasing resnet model coverage * fix indentation * lint * moving recursion limit fix into graphpack pass * moving recursionlimit to relay init * pooling on NCHWnc format * adding more models * deploy_resnet_on_vta.py * trailing line * generalizing to vision models * merge conflicts * fix, apply quantization to VTA only * improving comments * trimming models that have runtime issues for the moment * lint * lint * lint
…che#3740) * adding support for graphpack over multiply op * increasing resnet model coverage * fix indentation * lint * moving recursion limit fix into graphpack pass * moving recursionlimit to relay init * pooling on NCHWnc format * adding more models * deploy_resnet_on_vta.py * trailing line * generalizing to vision models * merge conflicts * fix, apply quantization to VTA only * improving comments * trimming models that have runtime issues for the moment * lint * lint * lint
This fixes Relay compilation for ResNet18_v2, ResNet34_v2, ResNet50_v2, ResNet101_v2, ResNet152_v2, Alexnet, VGG11, VGG13, VGG16, VGG19.
Currently the following models will work in HW/SIM (white-listed in tutorial until more models are supported): ResNet18_v2, ResNet34_v2, ResNet50_v2, ResNet101_v2
Performance on
pynq
IoT platform: 371.68ms, 470.86ms, 662.13ms and 916.26ms for ResNet-18, 34, 50, and 101 respectively.Bug fixes:
1<<24
to avoid out of memory errors on the pynq[resolved]: getting quantization to pass on these models is pending on #3543 being merged. I've addressed this dependence with a TODO.