Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Exploiting unstructured sparsity via tiramisu_pytorch #1

Open
Msabih opened this issue Jun 9, 2020 · 1 comment
Open

Exploiting unstructured sparsity via tiramisu_pytorch #1

Msabih opened this issue Jun 9, 2020 · 1 comment

Comments

@Msabih
Copy link

Msabih commented Jun 9, 2020

Hi,

If I have a model that has a number of weights set to zero via unstructured pruning, its it possible to utilize that from tiramisu_pytorch ?

@rbaghdadi
Copy link

Hi @Msabih , at this stage we still do not support sparse DNNs in this integration. If you are interested in helping with this please let us know. @IHIaadj is leading the integration of Tiramisu in Pytorch.

In order to have full support of sparse DNNs in tiramisu_pytorch we need to:

  • Include the sparse DNN operators in the tiramisu_pytorch package (we already have the sparse operators but they are not included in the package).
  • We need to add support for any data layout transformation that might be necessary between dense and sparse operators (if some operators do not have a sparse equivalent).

Feel to reach out by email, we can setup a zoom call to give you more details.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants