-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training PixelCNN unclear #5
Comments
Hi! It has been a while since I've worked on this project, so my memory is not too sharp. As far as I can see/remember, the input to the PixelCNN is a list of 3 dimensional one-hot encoded matrices (tensors), see how I unpack them in the PixelCNN: 3D-VQ-VAE-2/pixel_model/pixelcnn.py Lines 106 to 120 in 0b2148f
The whole pickling/txn context etc is just fluff needed for LMDB to work. The reason I use LMDB was that at the time it was the only database implementation available to support both memmapped arrays and concurrent reads (which is important for computational efficiency when running multi-node, which I did for sampling the full 512x512x128 volumes) As said I'm not entirely up-to-date on these kinds of workloads anymore, but two thoughts:
If you have more questions let me know. Robert Jan |
Hi,
I'm using your implementation to generate MRIs. I have trained a VQ-VAE to reconstruct 3D MRIs, but I am unsure about which vectors to use for training the PixelCNN for sampling.
I attempted to understand your LMDB implementation, but it would take me a significant amount of time to fully grasp it. I'm not clear on what exactly is being stored in the LMDB database.
Given that the VQ-VAE encoder outputs multiple quantization vectors (one for each encoding block), what should be the specific input for the PixelCNN?
I think i'll have to modify the LMDB data module part.
Thank you!
The text was updated successfully, but these errors were encountered: