-
Notifications
You must be signed in to change notification settings - Fork 5.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Auto Parallel] Upgrade fluid comm operators to be compatible with new comm library #56088
[Auto Parallel] Upgrade fluid comm operators to be compatible with new comm library #56088
Conversation
fluid operator support new comm library
…uce_scatter and c_scatter.
你的PR提交成功,感谢你对开源项目的贡献! |
python/paddle/distributed/auto_parallel/static/process_group.py
Outdated
Show resolved
Hide resolved
python/paddle/distributed/auto_parallel/static/process_group.py
Outdated
Show resolved
Hide resolved
1. remove useless loggings. 2. Fix conditional compilation for HIP. 3. Fix problems of test_pass_generation_pipeline.py. It calls paddle.distributed.init_parallel_env() at first, then auto.Engine calls _init_comm(), which will calls process_group.instantiate(). However, init_parallel_env() will call paddle.distributed.barrier(), it will call CreateNCCLEnvCache and create corresponding NCCLCommContext. But dev_id is not set, as a result, NCCLCommContext's dev_ctx is not initialized.
… static_compat_upgrade
will be submitted in another PR.
… pure_static_compat_upgrade
… pure_static_compat_upgrade
…Screaming/Paddle into static_compat_upgrade
… static_compat_upgrade
… static_compat_upgrade
… static_compat_upgrade
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
… static_compat_upgrade
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM for flags
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
Others
PR changes
Others
Description
Pcard-70448
Based on compatible new communication library upgrade PR #56604. Old communication library is used in default, export
FLAGS_dynamic_static_unified_comm=1
to use new communication library.Following operators are upgraded compatibly now:
alltoall
,barrier
,c_concat
,c_allreduce
,c_sync_stream_op
,c_wait_comm_op
,c_wait_compute_op
,recv_v2_op
,send_v2_op
,c_reduce_op
,c_allgather_op
,c_scatter_op
,c_reducescatter_op
.