-
Notifications
You must be signed in to change notification settings - Fork 876
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[v5.0.x] backport bugfixes created during mtt bug bash #11821
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
The two pre-defined groups: group_null and group_empty must be the 0th and 1st group in the group_table, for MPI_Group_f2c to be able to convert fortran group index to c_group. However, prior to this patch ompi_group_all_failed_procs was inserted to group table as the 0th group, which broke MPI_Group_f2c. This patch moved the insertion of ompi_group_all_failed_procs to after group_null and group_empty. Signed-off-by: Wei Zhang <wzam@amazon.com> (cherry picked from commit 75ea1df)
In MCA_PML_CM_HVY_SEND_REQUEST_BSEND_ALLOC(), after call to opal_convert_pack() will changed convertor's status, the convertor need to be reset to original state. This is achieved by calling opal_convertor_prepare_for_send(), and it should be called with original send buffer provided by application, which is sendreq->req_addr. However, prior to this change, the function was called with sendreq->req_buff, which is the temprary buffer used for send. As a result, when the same request is used the 2nd time, wrong data was copied to outgoing buffer, and caused memory corrupiton. This patch addressed the issue. Signed-off-by: Wei Zhang <wzam@amazon.com> (cherry picked from commit d71fe93)
ompi_mpi_compat_mpi3 controls whether MPI default behavior fix MPI-3 standard or MPI-4 standard. Because the main branch is following MPI-3 standard, this parameter's default value should be true, but it is false prior to this patch This patch addressed the issue. Signed-off-by: Wei Zhang <wzam@amazon.com> (cherry picked from commit 67a71fc)
Piror to this patch, the reduce code try to applied arithematic operations on a sendbuf, even when it is MPI_IN_PLACE. this patch addressed the issue. Signed-off-by: Wei Zhang <wzam@amazon.com> (cherry picked from commit d520921)
bosilca
approved these changes
Jul 12, 2023
bot:nvidia:retest |
Previous commit 2d68804 removed "errhandler_type" from communicator, and replaced it with "errhandler->eh_mpi_object_type". However, for an errhandler to be invoked on a communicator, errhandler_type must always be OMPI_ERRHANDLER_TYPE_COMM. But, errhandler->eh_mpi_object_type can be OMPI_ERRHANDLER_TYPE_PREDEFINED for predefined error handlers such as MPI_ERRORS_ARE_FATAL. This patch added "errhandler_type" back to communicator to address the issue. Signed-off-by: Wei Zhang <wzam@amazon.com> (cherry picked from commit 216c221)
added a commit "communicator: add errhandler_type back", which fix "MPI_Errhandler_fatal` |
jsquyres
approved these changes
Jul 13, 2023
This was referenced Jul 14, 2023
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR back port 4 patches that addressed issues in mtt:
coll/han: fix bug in reduce in place.
This commit fix
reduce_big_in_place
runtime/params: set ompi_mpi_compat_mpi3 to true by default.
This commit fix
MPI_Errhandler_set/get/free
pml/cm: fix buffer usage in MCA_PML_CM_HVY_SEND_REQUEST_BSEND_ALLOC()
This commit fix
MPI_Bsend_init
group: defer the insertion of ompi_group_all_failed_procs to group_table
This commit fix
MPI_Group_union/intersection
communicator: add errhandler_type back
This commit fix "MPI_Errhandler_fatal`