-
Notifications
You must be signed in to change notification settings - Fork 5.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[AutoParallel] Support shard parameter #57278
[AutoParallel] Support shard parameter #57278
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
… ap/support_shard_parameter
@@ -276,6 +276,7 @@ void InitDistTensorWithTensor(TensorObject* self, | |||
paddle::platform::errors::InvalidArgument( | |||
"DistTensor can only initialize by DenseTensor")); | |||
self->tensor.set_name(name); | |||
VLOG(4) << "Do TensorCopy from DenseTensor to DistTensor."; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这条日志应该可以移除,后续PR再移除
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* support shard parameter * add deep copy impl * dix adamax test failed
PR types
New features
PR changes
APIs
Description
Pcard-73145
[AutoParallel] Support shard parameter
paddle.distributed.shard_tensor需要支持对Parameter的切分,以满足当时设计的写法需求
此处Layer的Parameter被shard之后,仍然赋值给Layer的Parameter,因此shard_tensor不能改变Parameter的属性,仍然需要是Parameter