Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2D to 3D attention training question #11

Open
andrewbutts opened this issue Sep 21, 2023 · 3 comments
Open

2D to 3D attention training question #11

andrewbutts opened this issue Sep 21, 2023 · 3 comments

Comments

@andrewbutts
Copy link

Hiya, nice work! I just had a clarifying question:

When replicating the existing model’s 2D attention for 3D, do you freeze the portion of the 3D attention that relates a particular camera’s view to itself? Or do you let the training process affect it as well? If so, what’s the rationale and what effect does that have compared to leaving that part frozen?

Thanks! I hope my question makes sense!

@seasonSH
Copy link
Collaborator

Nothing is frozen in our model during training. We fine-tuned the whole model.
In fact, we tried to freeze some parts (e.g. conv and crossattn) in our experiments but saw no clear improvement. Therefore, we just fine-tune the whole model for simplicity.

@Friedrich-M
Copy link

I was wondering whether you directly finetune the whole unet parameters or use the Lora.

@seasonSH
Copy link
Collaborator

whole unet

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants