-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fluid benchmark & book validation #6208
Comments
Task Lists 2 : compare results and performance with TensorFlowWe select some typical tasks, to validate our performance and results with TensorFlow.
Task Lists 2 How to doscripts: both cpu tests can be done in docker image, but you need to install tensorflow-gpu by yourself: |
recomend docker command. |
|
This feature will be delayed to |
We have released the Fluid 0.11.0 two days ago. Next stage will be introducing our new design to users. We need more solid metrics to compare with other frameworks, find the potential flaw in Fluid. git clone https://github.com/dzhwinter/benchmark
docker pull dzhwinter/benchmark:latest
nvidia-docker run -it --name mnist_gpu --security-opt seccomp=unconfined -v $PWD/benchmark:/benchmark -v /usr/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu dzhwinter/benchmark:latest /bin/bash https://github.com/dzhwinter/benchmark/blob/master/HowToDoBenchmark.md Follow the guide step by step, then you will get the same environment. To make the benchmark more convincing and general testing, there are some decisions need to make.
https://www.tensorflow.org/performance/benchmarks TensorFlow and Pytorch focus on huge image related models, namely, CNN models. I think their test case is not general enough, but also can be a comparable sign. |
如所提出的问题,讨论了以上5个方面的问题。防止英文翻译有偏,会议记录中文如下:
分别选取了图像领域的经典模型VGG, Resnet,属于CNN,和NLP领域的经典模型stacked-lstm, seq2seq,属于RNN。 benchmark的镜像为当次发版后paddlepaddle/paddle:VERSION
benchmark需要和TensorFlow对齐指标,其中表格中test accuracy是迭代相同轮数后的结果,instance/second衡量训练速度,GPU memory size不同架构可能不同,作为参考。
NOTES: |
In the 0.11.0 version, we will release the book chapters written with
fluid
, there are some tasks need to be done.Task Lists 1 : compare results with Paddle books V2
Need to validate these books can convergence to the approximate same result with books chapters.
book.03 image classification CPU loss validation @jacquesqiao @qingqing01 @kuke
book.03 image classification GPU loss validation @jacquesqiao @qingqing01 @kuke
book.04 word2vec CPU loss validation @peterzhang2029
book.04 word2vec GPU loss validation @peterzhang2029
book.05 recommendation systems CPU loss validation @typhoonzero
book.05 recommendation systems GPU loss validation @typhoonzero
Need to note that we have three different implementation of understand_sentiment, only test the lstm one in this chapter.
book.06 understand_sentiment lstm CPU loss validation @ranqiu92
book.06 understand_sentiment lstm GPU loss validation @ranqiu92
book.07 label semantic roles CPU loss validation @chengduoZH
We do not have GPU version label semantic roles implementation.
book.08 machine translation CPU loss validation @jacquesqiao @ChunweiYan
book.08 machine translation GPU loss validation @jacquesqiao @ChunweiYan
Task Lists How to do
We have benchmark scripts and docker image. So these things should be done quickly and report a bug if you find any issue. (operator implement, convergence result).
Because we are still finetuning the performance, so if you find any magnitude gap in performance, please file an issue without hesitation.
scripts are put under this directory, please find the correct chapter name:
https://github.com/PaddlePaddle/Paddle/blob/develop/python/paddle/v2/fluid/tests/book
old books docker image:
paddlepaddle/book:latest-gpu
new books docker image:
dzhwinter/benchmark:latest
The text was updated successfully, but these errors were encountered: