-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
vgg16 model batch size 64 OOM on old docker image #74
Comments
Here note some conclusions.
|
A lesson learned from the vgg16 model. I vaguely remember previous V2 implementation version only can support 128 batch size with 4 12G cards. When it goes to Fluid, batch size 32 make the system reach the memory peak seems reasonable. However, a crucial fact that the mxnet only cost 7G GPU memory even with 200 layers. To be more concrete, let's do some math caculate. 128 * 3 * 224 * 224. Before do the convolution, we will do im2col. It's feature map shape equals (assume Same shape, kernel=3) 224 * 224 * (3 * 3 * 3) This im2col only used by one image, then We can raughly caculate the result, it can not reach a horrific 1.5G. |
bisect rollback to 1.11 image/ci build, but it makes nonsense. |
The text was updated successfully, but these errors were encountered: