Skip to content
Discussion options

You must be logged in to vote

Actually, dimensions must be integer when divided by 64.
Thus intead of "[4096, 2160].", it would be better to use [4096, 2176].

In my own code I also adapt the auto encoder "class Encoder(nn.Module):" in "models/image/efficient_ad/torch_model.py".
If you use larger images and do not adapt the auto-encoder; depth or the kernel size of the final layer, it will not see the whole image at once.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by samet-akcay
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #2635 on June 16, 2025 09:40.