Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Explore: Efficient Transformers #20

Open
msrepo opened this issue May 7, 2023 · 0 comments
Open

Explore: Efficient Transformers #20

msrepo opened this issue May 7, 2023 · 0 comments
Assignees

Comments

@msrepo
Copy link
Collaborator

msrepo commented May 7, 2023

Description

Currently, we are struggling to train transformer models for inputs larger than 128^3 volume. Within the volumetric grid representation of shape, the resolution achievable from this may not be enough. For example, for large bone shapes such as hip and ribs, we have had to use voxel resolution of ~2.5mm which is very coarse.
We might be able to output a more high resolution image if the transformers were lighter in terms of memory usage.

Proposal

  1. Use a 2D encoder and a 3D decoder. How do we execute this using transformer? A case in point how do we port the ideas from TransVert, 3DReconNet into pure transformers. How do we concatenate features from the two parallel x-ray branches and decode a 3D output?
  2. Implement/port Conv+transformer
    These might be more efficient in terms of memory usage.
    Try:
  3. implement/port various lightweight transformers and other interesting ideas
  • 3D Medical Axial transformer, CoTr and AFTer-UNet seem promising
@msrepo msrepo self-assigned this May 7, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant