-
Notifications
You must be signed in to change notification settings - Fork 572
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Set up new Intel 17.x build to use as auto PR build #2463
Comments
Found a problem with this plan for the Intel 17.x build. It looks like the SEMS env does not provide any builds of OpenMPI 2.x. All that seems to be available is:
Now they do have:
Seems like it would be a good idea to us a different MPI implementation for one of the builds. I know in CASL that we switched to MPICH because it caught more errors than OpenMPI at the time. Is this still true? As for Intel 17.x, it looks like the only version supported by SEMS is:
Is it a problem then if we go with Intel 17.0.1 + OpenMPI 1.8.7? Is there really any value in going with OpenMPI 1.10.1? I can try it if people think that is useful. Otherwise, should we ask SEMS to install a software stack with OpenMPI 2.x? That could take a while and it seems like they need to retire an OpenMPI version (like 1.6.5) before they add another OpenMPI version. |
@bartlettroscoe - we have generally found OpenMPI to be every bit as good as MPICH. More important, we use OpenMPI on the testbeds, CTS and it will underpin IBM Spectrum MPI on ATS2. In an ideal world we would test both MPI variants, but if we have to pick one I would select OpenMPI because of the CTS use. |
I just looked and it seems that |
@nmhamster, how do you define "good"? In the CASL case, we found that MPICH found errors in the usage of MPI that OpenMPI did not. I would have to dig up what versions those were that that was our experience. We did not care if OpenMPI ran faster than MPICH or visa versa because this was just for our test env. If I remember right (since it was many years ago), there was a defect in Tpetra MPI usage that OpenMPI let pass but when we ran CASL VERA on another machine, it bombed. It took a long time to debug to find the issue. Anyway, given that OpenMPI is the target for ATS-2, it seems like a good choice for our testing. |
OpenMPI 1.10.x implements the bits of MPI 3 that Tpetra optionally uses (with macros for the MPI version). For GPU builds, it's better to use newer versions of OpenMPI, but for CPU builds, I'm less worried about that for now. |
I can't remember all of the CASL errors, but one of the easier ones to diagnose was that openmpi allowed for aliasing of arrays which was technically not allowed in the standard and mpich automatically flagged those uses in casl code. |
@rppawlo That's a good point -- it would be helpful to have an extra Dashboard test for other MPI implementations. |
So should we try MPICH for this Intel 17.0.1 build or the GCC 4.8.4 build? Note that OpenMPI 1.8.7 causes 30 test timeouts with the GCC 4.8.4 build as described in #2462 (comment). I am currently testing OpenMPI 1.10.1 with that GCC 4.8.4 build. |
@bartlettroscoe I very deliberately said "Dashboard" not necessarily PR ;-) . I would welcome more MPI options for PR testing, but I would rather have mandatory PR testing sooner than have multiple MPIs in PR testing later :-) . I would say, OpenMPI 1.10.x w/ GCC 4.8.4, and MPICH w/ Intel 17.0.1. |
Working on this now - notes Matches our setup:
New stuff to look at. Why experimental code in PR instead of specialized or experimental tracks?
I will take a few days to get this much set up and working as I want to refactor the driver script as it is. Paul |
@prwolfe I don't think we actually need the MueLu and Xpetra "experimental" options enabled: but I'm not sure if the ATDM builds have disabled these options (yet). |
Thanks for the reference to that discussion @mhoemmen. That matches my instincts as well! |
This could be used, for example, for the Intel 17 build in trilinos#2463.
Kokkos is not using Pthread so don't name the build 'Pthread'. The Pthread TPL is enabled to allow other testing.
This could be used, for example, for the Intel 17 build in trilinos#2463.
It was requested that we use GCC 4.9.3 heades with Intel 17.0.1 builds of Trilinos (see trilinos#2317 and trilinos#2463).
One option for this Intel build is to use the SEMS Dev Env setup documented in: You basically just source:
and then configure Trilinos using the option:
Using the new aggregate file
and that is it! If we want to allow for tweaks (like some specific tests that need to be temporarily disabled), then we might create a new file called something like Yesterday I tested the full set of Primary Tested packages and TPLs on my machine crf450 with an experimental all-at-once build, test, and summit (with CMake 3.10.1) and it submitted to: This showed 7 failing tests for the packages:
That build took:
That is nearly 4 hours to run. Is that too long for an auto PR build? That ran the test categories See details below. We could set up a "Specialized" build for this that runs nightly and then get this cleaned up. DETAILS: (click to expand)
That submitted to: |
This has been done since about 6/1/2018 as shown in this query run just now. Closing as complete. |
CC: @trilinos/framework, @mhoemmen, @rppawlo, @ibaned, @crtrott
Next Action Status
Intel 17.0.1 PR builds running since 6/1/2018
Description
This Issue is to scope out and track efforts to create an Intel 17.x build that matches the auto PR build described in #2317 (comment).
The settings for this build are:
TPL_ENABLE_MPI=ON
(OpenMPI 2.x)BUILD_SHARED_LIBS=ON
CMAKE_BUILD_TYPE=RELEASE
Trilinos_ENABLE_DEBUG=OFF
Trilinos_ENABLE_EXPLICIT_TEMPLATE_INSTANTIATION=ON
Xpetra_ENABLE_Experimental=ON
MueLu_ENABLE_Experimental=ON
Trilinos_TRACE_ADD_TEST=ON
Trilinos_TEST_CATEGORIES=BASIC
The existing GCC 4.8.4 CI build shown here that has been running for 1.5+ years may be a good foundation for this build since it has most of the options already set and the
Trilinos/cmake/load_sems_env.sh
script already allows setting different compilers.Tasks:
a. NOTE: SEMS only provides
sems-intel/17.0.1
.b. NOTE: The highest version of OpenMPI provided by SEMS in
sems-openmpi/1.10.1
.Related Issues:
The text was updated successfully, but these errors were encountered: