Mobile deployment #102
-
Hi again! I saw that in the paper on which this is built upon, Supervised Compression for Resource-Constrained Edge Computing Systems from 2021, mentions the deployment of the encoder on mobile devices such at the Raspberry Pi 4 and NVIDIA Jetson TX2. Is there an example of how someone can deploy the head model on the mobile device and the tail model on the edge server, as to effectively replicate what was done in the study? |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
Hi @banciuadrian , Is your question about how we split For deployment, this repo does not offer such a script at this moment. In the WACV '22 paper, we manually split models into 2 pipelines (i.e., 2 Let me know if you need further clarifications. By the way, I happened to find one of your repos using both |
Beta Was this translation helpful? Give feedback.
-
Thank you for the heads up regarding the repository, I've made the suggested changes (to be noted at the moment the repository is really messy and not of any use). I do have a set of questions I've gathered while trying to split the modules, I hope this isn't too much!
|
Beta Was this translation helpful? Give feedback.
Hi @banciuadrian ,
Is your question about how we split
nn.Module
instances of the models in the WACV '22 paper into head and tail models?For deployment, this repo does not offer such a script at this moment. In the WACV '22 paper, we manually split models into 2 pipelines (i.e., 2
nn.Module
instances): 1) model input -> encoder output for mobile device and 2) tail model input (encoder output) -> tail model output for edge server.Let me know if you need further clarifications.
By the way, I happened to find one of your repos using both
sc2bench
andtorchdistill
https://github.com/banciuadrian/torchdistill_dense_gcn
Like
torchdistill
, I'd suggest removesc2bench/
folder in the repo and ju…