Implementation of PointTransformer #3169
QuanticDisaster
started this conversation in
Show and tell
Replies: 1 comment 3 replies
-
Thanks. This is great. Are you interested in contributing this layer directly? We can work on an example for it in addition or later. Note that we can usually simply pass in dummy node features in case none exists, e.g., a simply |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
While looking for recent architectures to try for 3D semantic segmentation, I came across PointTransformer, so I gave it a try to implement it with the Message Passing scheme.
The implementation is based on PointTransformer from Hengshuang Zhao and all (2020) https://arxiv.org/abs/2012.09164, more specifically in the 3.1, 3.2 and 3.3 sections and features an attention layer.
Unfortunately, there is no official implementation, so some details are not known such as the intermediate dimensions for the PointTransformer layer or when features cardinality is augmented or reduced.
This is why I also had a look at these github repositories /~https://github.com/POSTECH-CVLab/point-transformer and /~https://github.com/qq456cvb/Point-Transformers as well as /~https://github.com/lucidrains/point-transformer-pytorch (PointTransformer layer only)
The transformer block itself :
However, the network does not seem to support points only input, at least one feature is mandatory (or I didn't implement it correctly, which could be the case feel free to correct it !). While I tried to implement the full architecture, I didn't try it on the S3DIS or shapenet dataset.
I would be glad to hear your thoughts on it and see if someones wants to do a full example !
Beta Was this translation helpful? Give feedback.
All reactions