Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What is the relationship between mvsplat and unimatch? #84

Open
dodododddo opened this issue Dec 17, 2024 · 1 comment
Open

What is the relationship between mvsplat and unimatch? #84

dodododddo opened this issue Dec 17, 2024 · 1 comment

Comments

@dodododddo
Copy link

Hello! Excellent work! I am trying to understand more details about depth estimation in MVSplat from the code. MVSplat supports handling cases with more than three views, but its depth estimation module relies on the Unimatch framework (even using its pre-trained weights for initialization). However, Unimatch seems to only support handling two-view cases (at least based on its official implementation code). How does MVSplat enable Unimatch to support more views? Is it possible to provide code for a depth Unimatch module that supports multiple views? If I have misunderstood the relationship between MVSplat and Unimatch or how they work, please kindly correct me. Thank you!

@donydchen
Copy link
Owner

Hi @dodododddo, thanks for your appreciation.

Our multi-view Transformer is adopted from UniMatch. To extend from two-view to multi-view, we set the attention K, V as N-1 views rather than just one view. This essentially changes the token length of K, V from H*W*1 to H*W*(N-1), which does not affect any trainable parameters and allows us to use the two-view UniMatch pre-trained weight. For more details, you can compare the implementations between two-view attention and multi-view attention.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants