summaryrefslogtreecommitdiff
path: root/sci-chemistry/gromacs
diff options
context:
space:
mode:
Diffstat (limited to 'sci-chemistry/gromacs')
-rw-r--r--sci-chemistry/gromacs/Manifest7
-rw-r--r--sci-chemistry/gromacs/files/gromacs-2021-cuda-detection.patch339
-rw-r--r--sci-chemistry/gromacs/gromacs-2021.7.ebuild5
-rw-r--r--sci-chemistry/gromacs/gromacs-2023.1.ebuild331
4 files changed, 680 insertions, 2 deletions
diff --git a/sci-chemistry/gromacs/Manifest b/sci-chemistry/gromacs/Manifest
index 85834110c3d7..f1875f06cde3 100644
--- a/sci-chemistry/gromacs/Manifest
+++ b/sci-chemistry/gromacs/Manifest
@@ -1,5 +1,6 @@
AUX gromacs-2019.6-missing-include.patch 1052 BLAKE2B 01d7e592caa1e975a1fcea99bc37dd61ea7c039ed32243852f532c87d08911f9e83d0752d3a09e46643ca433339fe7c8d510c09e4af1070060735f9ee8a151d0 SHA512 8ed2e1de98cd35d87c0a0ea6867a99d4c1b33940040506feac09da929389afd1e19f34691c9032b1f59b156986b56c49aed1d0c8f8e7e64ad7a78e6dfd521912
AUX gromacs-2020-pytest.patch 1346 BLAKE2B 4b0b34bf4f46ec1acc3d0088ac11323938ee44b700beecc32b083885727d31904649798e38b824a6cf04ea7f9129dbd40e66bdde6e5a88fa4734db51bdf4cc46 SHA512 2c05e243b0e7bb8f9d25051d7bc86463dffa722502586799f0a8ce53bebd24fd0e367b6840365ac4ebc30595ffb8bbccb10a88d0c81148ae198941966cebddb0
+AUX gromacs-2021-cuda-detection.patch 21849 BLAKE2B 4976321e5f2c5bc9b45feeaa29206d77d04d6148ca910652bd3f7923b6e74b7b10f039b5bcc4430826745f6d64b232c536a8ef75cce61556275942e94945acc0 SHA512 6a1d1f0467d27267d214bc4df95c4b15438fffcae6b84e03763425f869e2551490c0f5e2d28774f9dda2daa07da8b5555617721d46e8b4f067eba2b2b86b353d
AUX gromacs-2021-musl-stdint.patch 601 BLAKE2B 2e871d22655950ee9b9951229e78bc022f74ce7a8aa8063bafa30e0fd97cd0451e0621ab48f985d813fd5d890c0f2d0c9ced6d4b26d86e9d1f6774c084f169af SHA512 d36497423beb86a26ad011432dd6da7169ae42dc00d15bc90ba14de6ecf5eb3217ac987260dce479e5f92d5f3a40fa28c5eb1b51816a92b6a37745f674581280
DIST gromacs-2018.8.tar.gz 29913703 BLAKE2B 8780032305928067fbfe1559efe9eedc4d47e27ab2f6ab54d6dba39edfcbeefbee4673d8910009048a850620b5a08ba8feb622db652c439cfa50a149ec5f0015 SHA512 6717895942f2ab7b54019511ed2aebadadde8e96d3c4b7414bb9168ffe418432d2dee330192e199f7b9d2f665c71f5d4f1ef0911aa7243d525a2ca182ef5b122
DIST gromacs-2019.6.tar.gz 33446147 BLAKE2B adc21fb6b841b06d499607f8c0166a673645ef5af0b40bd823d0fff5ea24397e6301c5e1e0070986ae1ce1deba8a42052b66da148b071c1e21f2fe3908fee275 SHA512 7c227a9539e5775d5d128ae9e872a24181d79cdcd2852b86453597987e8d7332ecec371da65d089f6e610c54aafbccc8824ce298743383dd6240719ebee692a8
@@ -8,6 +9,7 @@ DIST gromacs-2021.6.tar.gz 38012702 BLAKE2B 023ae37dc53643758afa050fd179cd66f3c6
DIST gromacs-2021.7.tar.gz 38028032 BLAKE2B 2dc1b15ca8692ca5cf4ad1c7a266b9e2ced064d34a63cef61fea69e48231dd0310cfd8484d29c56cdb599f795d7655959f40038ac933b46ad2841d57bf2b3646 SHA512 ca478533da3bcb377fda30682660db8eb13beef6456fac6013a1a97ec3be60b5594567412e1e2a31392e484f9a40a22da1f12336f4738cf0422ff4f3e8814609
DIST gromacs-2022.4.tar.gz 40352282 BLAKE2B 616ba05a70a483d7d8fadefb383ac9fc8825ac09ea92f55a0222c57ac91e4a045fb55405ec963303d8dbabb04b3d1f8c73062785b3812b3c6b7582e002dc5538 SHA512 15c642f2b63c1731a01a3b58c13d454f57c29607554a6693c2e62f38da574c6d596c5cb6f361e0feb5e1069b3b19bb49ceaa432c030da91f48e41654e3e6a62f
DIST gromacs-2022.5.tar.gz 40362776 BLAKE2B e77c96cd995fb6c8284bcf9728e2f95e265911f643c69d00cc5910c6d22ada41293a1cb74033e708957776551b253453f64dba46497d68b2f73c675e94f30ddd SHA512 79310d6fd39a160e7bbc661a04540c9bcb89f40c133d1031ca16998ef0a67fc4a023f57950b22793f0c91b5fe47fd78cb2e7afedb50fe9251b1c920c9267344d
+DIST gromacs-2023.1.tar.gz 41934072 BLAKE2B 6d04642ba37a8563fafc10e0be83e33aab5ce25ef43ec2c9555e84328096bbe61426558a1433b086028f3c6637a5b3ae82241e67f206c0dd0b5254f5ec27ae4c SHA512 4bddbbc56f26487db9d35f371d65725289d1355dc4c295d32698499e05eb74e31d4db579fc1f665edb55a68b002bddacf37c0f05d3f970b1111c149990d434b1
DIST gromacs-2023.tar.gz 41863526 BLAKE2B 81669eaf176e48855e168da81f410f949851a5b7d1fde43ee367f4bbaeef4a61b963e7c936c8a07d0eb51ad0e412e2826c6dd194c1df6d3ba39490decdf9fb21 SHA512 4593a0802d809ea797c3d3ba157851c279e9569e96c3aefa2dbcf97b5a1a74cd9840f0fcf0f032f49024691458941efb43e59e4e05c7ee44ee46d6b3621e52d5
DIST manual-2018.8.pdf 10025023 BLAKE2B afb0a6e6a72d78df743fcb57e7c1716848589e571dd35167b957a9b407ca27978ccfb6cb9e0df9c9439b888f352501a00ba32281ed1e0b4193bd606f1d77152e SHA512 b1972f7ce965bfc9377542993c5943ea4868a8ed23a969d4203264746d6bea3a7a65c6379196ece37fd6d68ec7ec80f827bde87d1049284af7082759a124f1d5
DIST manual-2019.6.pdf 12702376 BLAKE2B c350127bc06a9eb3dee73da39037c84daeb89500e23cb131bd19a150bf60602d4dde7611e0c6f0f344af4093e96a899303dac71b4df56fd0c44c3a48a56606b7 SHA512 dfdff67c2c1c9ce1f1c236b7686c5331b0fe86951da6c98eddd6a3e748815e0be0a9653ae4318469fcdd913c4e7c355d64a997cc80b9d9fbe6282ef6d98e61eb
@@ -16,6 +18,7 @@ DIST manual-2021.6.pdf 12256595 BLAKE2B 5d808e38571b38c1f6e0d115f6e604238206cc00
DIST manual-2021.7.pdf 12256526 BLAKE2B 6f298d7dac142f66914bbc2ea7d6ca70c4aeafaf33ccf12336d5cec64b1f1097b5332c767bb411ccef706ec6184479a84bcf59bda21c4249d45d8ce3dea08e79 SHA512 197765accb8ab6c5f1cc94b2697ff1817e157d85fa231dfe36a85036076e04da2a8035de66aba4f46c6dd7f553176357b94c92e1da5d1dabc574c627deee5279
DIST manual-2022.4.pdf 13080976 BLAKE2B 45685fa1d727aa30425e946ac72e0bd723528032e3fb45525d3c025316ee2682ea3c9cc2fd75b5cd4e900adbc6e7f577e3368d9e0681294fd1515df3d2e5d840 SHA512 a7ac3a80a90f14b6917e15633d77230d24cdcf19ade38d14b35c6451074848a7754b33c612e2a26a657f2edd3d6710e5a8130cacb107d4085d1bd934267ad623
DIST manual-2022.5.pdf 13081119 BLAKE2B 7a0ffd086d98527d42da4a22984d8351ad7861e32951e8b65c88e7d0df1559303750886ae31a7c158b4e22f616f6dd5deed9a63b0ed114bb1e0adfda9350a487 SHA512 707869ee6052f1b7f75376b0f78fcc4859035234957cc740bd770ea3bea3a591f7fc1fbc0b8526baab48eebfa506d03927df94ddc9c412c75dc2f2864574dd82
+DIST manual-2023.1.pdf 13526498 BLAKE2B a9aabcb722954c9c83a8f5945c85e634b0b82ace8b9b077c3f2b9ff432ca71d3e2d5ca2dd52d6ce489d77d3d4ca7b41d75e5566205bb8855caa650fa2722e5ec SHA512 b6ca6c822ac5b3cbe3c84d8fd51d3ed465268c00bb5f96e2afa557d45f9c534b0de278f33a0fdf43080c9c619344b98cd71847e17b297a789eec4c8b42dea7b2
DIST manual-2023.pdf 13516908 BLAKE2B 03a62190921c1a5ddf1ce6b0af5f8d83eaea26d31ac9b7e2f59f8b08e350d7147ee03ff8ec4e81e43777d04fa3d91c96d2a3dc8daa53a2565df02c39d6897d06 SHA512 2b16ca234941ede6415f89e0e31ee611ea1324454f5ddaf3cbcbdfbb5e864344faf012e5106131dcb9d3c1ffbbac358f1ea387166b1e45035e6427e344ff1964
DIST regressiontests-2018.8.tar.gz 67855469 BLAKE2B 34c9b339f6229f483afbb5192ee6ba8b8f72d5c26907a853af9c53dfece0d88739e48f6b44b78d1c010f988f9385d077285300522164f533a5861e9dda879275 SHA512 3642389d27bd1942cd0f091c940ae97b197b94856a387fe581dc516b0d4169480f16551d4ba357f9282b3337d605c286d51dd38112ac87c826dda634904836bb
DIST regressiontests-2019.6.tar.gz 67643195 BLAKE2B 1e054e24b187946f7ea28090d4f20cf8e1d79a26253f57ba07d130e0773d8541b8a1552a38023d31a68ce8bc62d8e0af1d98609234bb3e7d3e6d567307ebb386 SHA512 eacf1c55b982515a305c29459fe80a7cd558a8481e5689962aa956148af542568b1d1ce59c6784ecb0afd4768c2b664afa21e12af2d89ae2b06b1ba61ad72036
@@ -24,15 +27,17 @@ DIST regressiontests-2021.6.tar.gz 48537070 BLAKE2B 982fad6b23da5c1e6d32009f9f08
DIST regressiontests-2021.7.tar.gz 48537500 BLAKE2B b0c97fa652cb9b972c50f29822f530fcaa2ec9acb6d6442114fc6446c54aade5eb1bfa153c69a7e8bf64b66e8b041d7f8e0d51e7f7b0516f9e03f3d9ad626193 SHA512 7cbead90aa6d62346dcf7c2249c3cfa179884b987a749624444d2f59d93e0322f7721143bf6eb44d26102635dc70a73546ef82d4647558b6320bfdd459074017
DIST regressiontests-2022.4.tar.gz 48614599 BLAKE2B d41d59d4336455c87f3877be48ec49b2cdc989548c51ef9c8277b05bb6cf3dc05ebfa91a0248cc113c130a0ed19aae2c8316184caf37b72aa612b211fe41b144 SHA512 6495a286b201774ea2a922376ff5c0abaae7aaef1dfd28a1a61c817b5c6ac83db99300d43cc019745ca32a025e1f8a7f538342af539d9d753e494c550b429706
DIST regressiontests-2022.5.tar.gz 48615858 BLAKE2B 6655946155291eb24f76832c76e21b5c606905e94426fd057ecdaa44d74844613ba288b651f94120553ef7d7f938de345ddaac7c3738d67bdad148359b7b4500 SHA512 8dd0c8a9b30f4c047fb756dee8eee1cb667803c231fd114717e3156c5694a195dec2b57d80b50b417312949bc4cb7da90027cc98102d84fa154a93862bc69790
+DIST regressiontests-2023.1.tar.gz 48613205 BLAKE2B b64716e08e5597676794209a0a951d2a5a46aab2b05bc1615d6f4ba666d5c81a7fe9247fc90ce8d9a5da1a9d52eea73d3d3cd4a4072a8bad66b8de0e06dfa599 SHA512 c6924ec75e456950cef0809ea9718be1bf9d85fe6e9ba5553b2df86ef604b0c7f54d41476f45cbc0ccf801ee0290a829bf4f9eb8d0773586bb8385efc5242d4d
DIST regressiontests-2023.tar.gz 48620353 BLAKE2B 536550976dd761a59ff5684bc4928d534f7e45e69b0a86831e661887a3ad1ea72c1ddb72ab3a27237926083a711dd0f17ea60079fec201dd0992c1aee32ebd7e SHA512 f6b62fab39c77491698822b2f5b3f66c505e801266b31ea7ddc46378f69b524461187bba31b1b23d29f06bfc35630b0cae9a921aa92c0dfba46b2f1d904ce2b6
EBUILD gromacs-2018.8-r1.ebuild 6675 BLAKE2B 46cd37f680b71e001586e88974f7a69bd8e41d07de2464ab0d8d80f7755a68da45b77ad28e01444cb48b0156e6f4332dc8fc67d21944c76768b9cbd18804cc46 SHA512 ece906b1c1d3b401ea85121b280f99fafe92f791918c47aab57f1094b53e95a5786812e414928ec765ef44fbecf6219e094dfb74d5a52f0b5fccbdf736a37968
EBUILD gromacs-2019.6-r2.ebuild 7053 BLAKE2B b61fb0e08daec321b562e246275e1ff9768bf3e24428c127fbcc67ae36fbe6054b60f3cb0587fb711e9c050798086bcd7dece03a64f5bebd53db327db939acb5 SHA512 de3d96b9780667be98a519bb496eb72c2f02cb8aa9482d075d03fa1558ef2852c4daa6e8ef36879232604a7dc30722464e70886cb7673a371691bf93252655a2
EBUILD gromacs-2020.7.ebuild 9847 BLAKE2B 5ac1c696c996e4d281d8e90b858982f9f6a475867ec94209b4a42a1018282752878f6d8a62827a06af247bd00b766be6bb52176ebbccb1c616190a38ed507627 SHA512 03044916378006f3d26e55d56b2fecf637680c80b3eef85f5427e3ac92406926ff9d22ad5b1a1450477e579cb41030d39c787566a2ccf9829c78cf795365f3b1
EBUILD gromacs-2021.6.ebuild 10054 BLAKE2B 6a1a608a8996cff7b2d429cf4d7f8b2909271c6a83158671687619303904f5ead3cdce98d5ca968b05ac6fb8f16d764f08f111009a78d3132e88d64c65ca6c5a SHA512 43fe99bf2b2d901477675bb0fc30e244baeaf8b74b9736dc4e51d15c16544c4c8e801eaa2585335b32f92dd8a29d000aa1c3e511c4975570e573d28be401b324
-EBUILD gromacs-2021.7.ebuild 10057 BLAKE2B bed7c9d0cb7f7ac408070a4e115a172311a14031e221cde57a60a781419af95f6271aa523b1ade0fde41b49624926c52053c3a5b7fadb9f889d387c4fe93ed04 SHA512 b52de46729be145c4b762bd0b5b3bbcdeb63798f1658c32d8d2ae77f9b9fbe7843ad5b195e24ba3862057d11ecd3e464d93c2f28badc94e50e7e5458b76a5b3a
+EBUILD gromacs-2021.7.ebuild 10105 BLAKE2B a97146e3165e8ee528921d48375a7968ced93de69aad48a028227a8c418b03af044ec47c7674dd3e5adf89c85368cc53c18740ceccb9a4c59ec0631c23807bc7 SHA512 158aa5c265d3fd4a0b59a2c7abc467ef873599e12f07bf939eea56661b659b42d163e3c1363e772f86c112d41b1b8658df5499302f2584a5a69acde9cf57ee22
EBUILD gromacs-2022.4.ebuild 9596 BLAKE2B b3e675f96621add5f81b07ca1ec6a6c4f4f3f61f69051ba6f6319bd5f5044bacb67554bd7e85877485163c4695cee7dc3246b024b904a0f97d2d858065a4a098 SHA512 f81d9732b050c4cd8d453031e3122e5c0beeacda352493e0acc0b8cade05b36fd4671f87a89036bcc7dddd098ebd825d3c17af3c06b391d7514d75520ca9368b
EBUILD gromacs-2022.5.ebuild 9596 BLAKE2B b3e675f96621add5f81b07ca1ec6a6c4f4f3f61f69051ba6f6319bd5f5044bacb67554bd7e85877485163c4695cee7dc3246b024b904a0f97d2d858065a4a098 SHA512 f81d9732b050c4cd8d453031e3122e5c0beeacda352493e0acc0b8cade05b36fd4671f87a89036bcc7dddd098ebd825d3c17af3c06b391d7514d75520ca9368b
EBUILD gromacs-2022.9999.ebuild 9596 BLAKE2B b3e675f96621add5f81b07ca1ec6a6c4f4f3f61f69051ba6f6319bd5f5044bacb67554bd7e85877485163c4695cee7dc3246b024b904a0f97d2d858065a4a098 SHA512 f81d9732b050c4cd8d453031e3122e5c0beeacda352493e0acc0b8cade05b36fd4671f87a89036bcc7dddd098ebd825d3c17af3c06b391d7514d75520ca9368b
+EBUILD gromacs-2023.1.ebuild 9544 BLAKE2B a03201533e4d713e62a7a93d70e2d621c9afb3af832b2aa711bb3840359f7f79b9529552c2cfa22a7ea1cd53d2d2524bd8fd382c932ceb233f7bbf769349de1c SHA512 1017e32229be36967c1e1a2c9c883d80f84bb8d2dc59ead2d34cf9299b981f27fe5d868df583c74af881c1916519af265cb6b126d01a5f337a926772a2e13dcb
EBUILD gromacs-2023.9999.ebuild 9544 BLAKE2B a03201533e4d713e62a7a93d70e2d621c9afb3af832b2aa711bb3840359f7f79b9529552c2cfa22a7ea1cd53d2d2524bd8fd382c932ceb233f7bbf769349de1c SHA512 1017e32229be36967c1e1a2c9c883d80f84bb8d2dc59ead2d34cf9299b981f27fe5d868df583c74af881c1916519af265cb6b126d01a5f337a926772a2e13dcb
EBUILD gromacs-2023.ebuild 9544 BLAKE2B a03201533e4d713e62a7a93d70e2d621c9afb3af832b2aa711bb3840359f7f79b9529552c2cfa22a7ea1cd53d2d2524bd8fd382c932ceb233f7bbf769349de1c SHA512 1017e32229be36967c1e1a2c9c883d80f84bb8d2dc59ead2d34cf9299b981f27fe5d868df583c74af881c1916519af265cb6b126d01a5f337a926772a2e13dcb
EBUILD gromacs-9999.ebuild 9544 BLAKE2B a03201533e4d713e62a7a93d70e2d621c9afb3af832b2aa711bb3840359f7f79b9529552c2cfa22a7ea1cd53d2d2524bd8fd382c932ceb233f7bbf769349de1c SHA512 1017e32229be36967c1e1a2c9c883d80f84bb8d2dc59ead2d34cf9299b981f27fe5d868df583c74af881c1916519af265cb6b126d01a5f337a926772a2e13dcb
diff --git a/sci-chemistry/gromacs/files/gromacs-2021-cuda-detection.patch b/sci-chemistry/gromacs/files/gromacs-2021-cuda-detection.patch
new file mode 100644
index 000000000000..8b458e96cb53
--- /dev/null
+++ b/sci-chemistry/gromacs/files/gromacs-2021-cuda-detection.patch
@@ -0,0 +1,339 @@
+--- gromacs-2021.7/cmake/gmxManageNvccConfig.cmake 2023-01-31 12:45:45.000000000 +0100
++++ gromacs-2022.5/cmake/gmxManageNvccConfig.cmake 2023-02-03 12:53:34.000000000 +0100
+@@ -1,11 +1,9 @@
+ #
+ # This file is part of the GROMACS molecular simulation package.
+ #
+-# Copyright (c) 2012,2013,2014,2015,2016 by the GROMACS development team.
+-# Copyright (c) 2017,2018,2019,2020,2021, by the GROMACS development team, led by
+-# Mark Abraham, David van der Spoel, Berk Hess, and Erik Lindahl,
+-# and including many others, as listed in the AUTHORS file in the
+-# top-level source directory and at http://www.gromacs.org.
++# Copyright 2012- The GROMACS Authors
++# and the project initiators Erik Lindahl, Berk Hess and David van der Spoel.
++# Consult the AUTHORS/COPYING files and https://www.gromacs.org for details.
+ #
+ # GROMACS is free software; you can redistribute it and/or
+ # modify it under the terms of the GNU Lesser General Public License
+@@ -19,7 +17,7 @@
+ #
+ # You should have received a copy of the GNU Lesser General Public
+ # License along with GROMACS; if not, see
+-# http://www.gnu.org/licenses, or write to the Free Software Foundation,
++# https://www.gnu.org/licenses, or write to the Free Software Foundation,
+ # Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+ #
+ # If you want to redistribute modifications to GROMACS, please
+@@ -28,10 +26,10 @@
+ # consider code for inclusion in the official distribution, but
+ # derived work must not be called official GROMACS. Details are found
+ # in the README & COPYING files - if they are missing, get the
+-# official version at http://www.gromacs.org.
++# official version at https://www.gromacs.org.
+ #
+ # To help us fund GROMACS development, we humbly ask that you cite
+-# the research papers on the package. Check out http://www.gromacs.org.
++# the research papers on the package. Check out https://www.gromacs.org.
+
+ # Manage CUDA nvcc compilation configuration, try to be smart to ease the users'
+ # pain as much as possible:
+@@ -51,7 +49,7 @@
+ # glibc source shows that _FORCE_INLINES is only used in this string.h
+ # feature and performance of memcpy variants is unimportant for CUDA
+ # code in GROMACS. So this workaround is good enough to keep problems
+-# away from users installing GROMACS. See Issue #1942.
++# away from users installing GROMACS. See Issue #1982.
+ function(work_around_glibc_2_23)
+ try_compile(IS_GLIBC_2_23_OR_HIGHER ${CMAKE_BINARY_DIR} ${CMAKE_SOURCE_DIR}/cmake/TestGlibcVersion.cpp)
+ if(IS_GLIBC_2_23_OR_HIGHER)
+@@ -83,67 +81,158 @@
+ mark_as_advanced(CUDA_HOST_COMPILER CUDA_HOST_COMPILER_OPTIONS)
+ endif()
+
++# We would like to be helpful and reject the host compiler with a
++# clear error message at configure time, rather than let nvcc
++# later reject the host compiler as not supported when the first
++# CUDA source file is built. We've implemented that for current
++# nvcc running on Unix-like systems, but e.g. changes to nvcc
++# will further affect the limited portability of this checking
++# code. Set the CMake variable GMX_NVCC_WORKS on if you want to
++# bypass this check.
++if((_cuda_nvcc_executable_or_flags_changed OR CUDA_HOST_COMPILER_CHANGED OR NOT GMX_NVCC_WORKS) AND NOT WIN32)
++ message(STATUS "Check for working NVCC/C++ compiler combination with nvcc '${CUDA_NVCC_EXECUTABLE}'")
++ execute_process(COMMAND ${CUDA_NVCC_EXECUTABLE} --compiler-bindir=${CUDA_HOST_COMPILER} -c ${CUDA_NVCC_FLAGS} ${CUDA_NVCC_FLAGS_${_build_type}} ${CMAKE_SOURCE_DIR}/cmake/TestCUDA.cu
++ RESULT_VARIABLE _cuda_test_res
++ OUTPUT_VARIABLE _cuda_test_out
++ ERROR_VARIABLE _cuda_test_err
++ OUTPUT_STRIP_TRAILING_WHITESPACE)
++
++ if(${_cuda_test_res})
++ message(STATUS "Check for working NVCC/C compiler combination - broken")
++ message(STATUS "${CUDA_NVCC_EXECUTABLE} standard output: '${_cuda_test_out}'")
++ message(STATUS "${CUDA_NVCC_EXECUTABLE} standard error: '${_cuda_test_err}'")
++ if(${_cuda_test_err} MATCHES "nsupported")
++ message(FATAL_ERROR "NVCC/C++ compiler combination does not seem to be supported. CUDA frequently does not support the latest versions of the host compiler, so you might want to try an earlier C++ compiler version and make sure your CUDA compiler and driver are as recent as possible. Set the GMX_NVCC_WORKS CMake cache variable to bypass this check if you know what you are doing.")
++ elseif(CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_EQUAL 11.2 AND CUDA_VERSION VERSION_GREATER 11.4 AND CUDA_VERSION VERSION_LESS 11.7) # Issue #4574, #4641
++ # Above, we should be checking for VERSION_LESS 11.6.2, but CUDA_VERSION is only "major.minor"
++ message(FATAL_ERROR "CUDA versions 11.5-11.6.1 are known to be incompatible with some GCC 11.x. Use a different GCC or update your CUDA installation to at least CUDA 11.6.2")
++ else()
++ message(FATAL_ERROR "CUDA compiler does not seem to be functional or is not compatible with the host compiler. Set the GMX_NVCC_WORKS CMake cache variable to bypass this check if you know what you are doing.")
++ endif()
++ elseif(NOT GMX_CUDA_TEST_COMPILER_QUIETLY)
++ message(STATUS "Check for working NVCC/C++ compiler combination - works")
++ set(GMX_NVCC_WORKS TRUE CACHE INTERNAL "Nvcc can compile a trivial test program")
++ endif()
++endif() # GMX_CHECK_NVCC
++
++# Tests a single flag to use with nvcc.
++#
++# If the flags are accepted, they are appended to the variable named
++# in the first argument. The cache variable named in the second
++# argument is used to avoid rerunning the check in future invocations
++# of cmake. The list of flags to check follows these two required
++# arguments.
++#
++# Note that a space-separated string of flags, or a flag-value pair
++# separated by spaces will not work. Use the single-argument forms
++# accepted by nvcc, like "--arg=value".
++#
++# As this code is not yet tested on Windows, it always accepts the
++# flags in that case.
++function(gmx_add_nvcc_flag_if_supported _output_variable_name_to_append_to _flags_cache_variable_name)
++ # If the check has already been run, do not re-run it
++ if (NOT ${_flags_cache_variable_name} AND NOT WIN32)
++ message(STATUS "Checking if nvcc accepts flags ${ARGN}")
++ # See detailed comment about gcc 7 below
++ if (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 8)
++ set(_cache_variable_value TRUE)
++ message(STATUS "Checking if nvcc accepts flags ${ARGN} - Assuming success when using gcc 7")
++ else()
++ if(NOT(CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 11))
++ set(CCBIN "--compiler-bindir=${CUDA_HOST_COMPILER}")
++ endif()
++ execute_process(
++ COMMAND ${CUDA_NVCC_EXECUTABLE} ${ARGN} ${CCBIN} "${CMAKE_SOURCE_DIR}/cmake/TestCUDA.cu"
++ RESULT_VARIABLE _cuda_success
++ OUTPUT_QUIET
++ ERROR_QUIET
++ )
++ # Convert the success value to a boolean and report status
++ if (_cuda_success EQUAL 0)
++ set(_cache_variable_value TRUE)
++ message(STATUS "Checking if nvcc accepts flags ${ARGN} - Success")
++ else()
++ set(_cache_variable_value FALSE)
++ message(STATUS "Checking if nvcc accepts flags ${ARGN} - Failed")
++ endif()
++ endif()
++ set(${_flags_cache_variable_name} ${_cache_variable_value} CACHE BOOL "Whether NVCC supports flag(s) ${ARGN}")
++ endif()
++ # Append the flags to the output variable if they have been tested to work
++ if (${_flags_cache_variable_name} OR WIN32)
++ list(APPEND ${_output_variable_name_to_append_to} ${ARGN})
++ set(${_output_variable_name_to_append_to} ${${_output_variable_name_to_append_to}} PARENT_SCOPE)
++ endif()
++endfunction()
++
++# Versions of gcc 7 have differing behavior when executing
++#
++# nvcc $args --compiler-bindir=gcc TestCUDA.cu
++#
++# and would need e.g. adding -lstdc++ to the command line so that
++# linking of a C++/CUDA object by the C-compiler flavor of gcc works.
++# This means we can't reliably test compiler flags in this case
++# without risking creating other problems. Instead we assume (above)
++# that all compiler flags will work, and issue this warning.
++# We also want to skip this warning during GROMACS CI testing.
++if (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 8
++ AND NOT DEFINED ENV{GITLAB_CI})
++ message(WARNING "You are using gcc version 7 with the CUDA compiler nvcc. GROMACS cannot reliably test compiler arguments for this combination, so if you later experience errors in building GROMACS, please use a more recent version of gcc.")
++endif()
++
+ # If any of these manual override variables for target CUDA GPU architectures
+ # or virtual architecture is set, parse the values and assemble the nvcc
+ # command line for these. Otherwise use our defaults.
+ # Note that the manual override variables require a semicolon separating
+ # architecture codes.
++set(GMX_CUDA_NVCC_GENCODE_FLAGS)
+ if (GMX_CUDA_TARGET_SM OR GMX_CUDA_TARGET_COMPUTE)
+- set(GMX_CUDA_NVCC_GENCODE_FLAGS)
+ set(_target_sm_list ${GMX_CUDA_TARGET_SM})
+ foreach(_target ${_target_sm_list})
+- list(APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_${_target},code=sm_${_target}")
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_${_target} "--generate-code=arch=compute_${_target},code=sm_${_target}")
++ if (NOT NVCC_HAS_GENCODE_COMPUTE_AND_SM_${_target} AND NOT WIN32)
++ message(FATAL_ERROR "Your choice of ${_target} in GMX_CUDA_TARGET_SM was not accepted by nvcc, please choose a target that it accepts")
++ endif()
+ endforeach()
+ set(_target_compute_list ${GMX_CUDA_TARGET_COMPUTE})
+ foreach(_target ${_target_compute_list})
+- list(APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_${_target},code=compute_${_target}")
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_${_target} --generate-code=arch=compute_${_target},code=compute_${_target})
++ if (NOT NVCC_HAS_GENCODE_COMPUTE_${_target} AND NOT WIN32)
++ message(FATAL_ERROR "Your choice of ${_target} in GMX_CUDA_TARGET_COMPUTE was not accepted by nvcc, please choose a target that it accepts")
++ endif()
+ endforeach()
+ else()
+ # Set the CUDA GPU architectures to compile for:
+- # - with CUDA >=9.0 CC 7.0 is supported and CC 2.0 is no longer supported
+- # => compile sm_30, sm_35, sm_37, sm_50, sm_52, sm_60, sm_61, sm_70 SASS, and compute_35, compute_70 PTX
+- # - with CUDA >=10.0 CC 7.5 is supported
+- # => compile sm_30, sm_35, sm_37, sm_50, sm_52, sm_60, sm_61, sm_70, sm_75 SASS, and compute_35, compute_75 PTX
+ # - with CUDA >=11.0 CC 8.0 is supported
+ # => compile sm_35, sm_37, sm_50, sm_52, sm_60, sm_61, sm_70, sm_75, sm_80 SASS, and compute_35, compute_80 PTX
+
+ # First add flags that trigger SASS (binary) code generation for physical arch
+- if(CUDA_VERSION VERSION_LESS "11.0")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_30,code=sm_30")
+- endif()
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_35,code=sm_35")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_37,code=sm_37")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_50,code=sm_50")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_52,code=sm_52")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_60,code=sm_60")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_61,code=sm_61")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_70,code=sm_70")
+- if(NOT CUDA_VERSION VERSION_LESS "10.0")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_75,code=sm_75")
+- endif()
+- if(NOT CUDA_VERSION VERSION_LESS "11.0")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_80,code=sm_80")
+- # Requesting sm or compute 35, 37, or 50 triggers deprecation messages with
+- # nvcc 11.0, which we need to suppress for use in CI
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-Wno-deprecated-gpu-targets")
+- endif()
+- if(NOT CUDA_VERSION VERSION_LESS "11.1")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_86,code=sm_86")
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_35 --generate-code=arch=compute_35,code=sm_35)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_37 --generate-code=arch=compute_37,code=sm_37)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_50 --generate-code=arch=compute_50,code=sm_50)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_52 --generate-code=arch=compute_52,code=sm_52)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_60 --generate-code=arch=compute_60,code=sm_60)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_61 --generate-code=arch=compute_61,code=sm_61)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_70 --generate-code=arch=compute_70,code=sm_70)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_75 --generate-code=arch=compute_75,code=sm_75)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_80 --generate-code=arch=compute_80,code=sm_80)
++ # Don't attempt to add newest architectures with old GNU compiler, to avoid issues in CI
++ # related to being unable to test which flags are supported
++ if (NOT (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 8))
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_86 --generate-code=arch=compute_86,code=sm_86)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_89 --generate-code=arch=compute_89,code=sm_89)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_AND_SM_90 --generate-code=arch=compute_90,code=sm_90)
+ endif()
++ # Requesting sm or compute 35, 37, or 50 triggers deprecation messages with
++ # nvcc 11.0, which we need to suppress for use in CI
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_WARNING_NO_DEPRECATED_GPU_TARGETS -Wno-deprecated-gpu-targets)
+
+ # Next add flags that trigger PTX code generation for the
+ # newest supported virtual arch that's useful to JIT to future architectures
+ # as well as an older one suitable for JIT-ing to any rare intermediate arch
+ # (like that of Jetson / Drive PX devices)
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_35,code=compute_35")
+- if(CUDA_VERSION VERSION_LESS "11.0")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_32,code=compute_32")
+- else()
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_53,code=compute_53")
+- endif()
+- if(NOT CUDA_VERSION VERSION_LESS "11.0")
+- list (APPEND GMX_CUDA_NVCC_GENCODE_FLAGS "-gencode;arch=compute_80,code=compute_80")
+- endif()
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_53 --generate-code=arch=compute_53,code=sm_53)
++ gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_GENCODE_FLAGS NVCC_HAS_GENCODE_COMPUTE_80 --generate-code=arch=compute_80,code=sm_80)
+ endif()
+
+ if (GMX_CUDA_TARGET_SM)
+@@ -158,27 +247,20 @@
+ # FindCUDA.cmake is unaware of the mechanism used by cmake to embed
+ # the compiler flag for the required C++ standard in the generated
+ # build files, so we have to pass it ourselves
+-if (CUDA_VERSION VERSION_LESS 11.0)
+- # CUDA doesn't formally support C++17 until version 11.0, so for
+- # now host-side code that compiles with CUDA is restricted to
+- # C++14. This needs to be expressed formally for older CUDA
+- # version.
++
++# gcc-7 pre-dated C++17, so uses the -std=c++1z compiler flag for it,
++# which modern nvcc does not recognize. So we work around that by
++# compiling in C++14 mode. Clang doesn't have this problem because nvcc
++# only supports version of clang that already understood -std=c++17
++if (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 8)
+ list(APPEND GMX_CUDA_NVCC_FLAGS "${CMAKE_CXX14_STANDARD_COMPILE_OPTION}")
+ else()
+- # gcc-7 pre-dated C++17, so uses the -std=c++1z compiler flag for it,
+- # which modern nvcc does not recognize. So we work around that by
+- # compiling in C++14 mode. Clang doesn't have this problem because nvcc
+- # only supports version of clang that already understood -std=c++17
+- if (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 8)
+- list(APPEND GMX_CUDA_NVCC_FLAGS "${CMAKE_CXX14_STANDARD_COMPILE_OPTION}")
+- else()
+- list(APPEND GMX_CUDA_NVCC_FLAGS "${CMAKE_CXX17_STANDARD_COMPILE_OPTION}")
+- endif()
++ list(APPEND GMX_CUDA_NVCC_FLAGS "${CMAKE_CXX17_STANDARD_COMPILE_OPTION}")
+ endif()
+
+ # assemble the CUDA flags
+ list(APPEND GMX_CUDA_NVCC_FLAGS "${GMX_CUDA_NVCC_GENCODE_FLAGS}")
+-list(APPEND GMX_CUDA_NVCC_FLAGS "-use_fast_math")
++gmx_add_nvcc_flag_if_supported(GMX_CUDA_NVCC_FLAGS NVCC_HAS_USE_FAST_MATH -use_fast_math)
+
+ # assemble the CUDA host compiler flags
+ list(APPEND GMX_CUDA_NVCC_FLAGS "${CUDA_HOST_COMPILER_OPTIONS}")
+@@ -187,12 +269,18 @@
+ # CUDA header cuda_runtime_api.h in at least CUDA 10.1 uses 0
+ # where nullptr would be preferable. GROMACS can't fix these, so
+ # must suppress them.
+- GMX_TEST_CXXFLAG(CXXFLAGS_NO_ZERO_AS_NULL_POINTER_CONSTANT "-Wno-zero-as-null-pointer-constant" NVCC_CLANG_SUPPRESSIONS_CXXFLAGS)
++ GMX_TEST_CXXFLAG(HAS_WARNING_NO_ZERO_AS_NULL_POINTER_CONSTANT "-Wno-zero-as-null-pointer-constant" NVCC_CLANG_SUPPRESSIONS_CXXFLAGS)
+
+ # CUDA header crt/math_functions.h in at least CUDA 10.x and 11.1
+ # used throw() specifications that are deprecated in more recent
+ # C++ versions. GROMACS can't fix these, so must suppress them.
+- GMX_TEST_CXXFLAG(CXXFLAGS_NO_DEPRECATED_DYNAMIC_EXCEPTION_SPEC "-Wno-deprecated-dynamic-exception-spec" NVCC_CLANG_SUPPRESSIONS_CXXFLAGS)
++ GMX_TEST_CXXFLAG(HAS_WARNING_NO_DEPRECATED_DYNAMIC_EXCEPTION_SPEC "-Wno-deprecated-dynamic-exception-spec" NVCC_CLANG_SUPPRESSIONS_CXXFLAGS)
++
++ # CUDA headers cuda_runtime.h and channel_descriptor.h in at least
++ # CUDA 11.0 uses many C-style casts, which are ncessary for this
++ # header to work for C. GROMACS can't fix these, so must suppress
++ # the warnings they generate
++ GMX_TEST_CXXFLAG(HAS_WARNING_NO_OLD_STYLE_CAST "-Wno-old-style-cast" NVCC_CLANG_SUPPRESSIONS_CXXFLAGS)
+
+ # Add these flags to those used for the host compiler. The
+ # "-Xcompiler" prefix directs nvcc to only use them for host
+@@ -205,37 +293,6 @@
+ string(TOUPPER "${CMAKE_BUILD_TYPE}" _build_type)
+ gmx_check_if_changed(_cuda_nvcc_executable_or_flags_changed CUDA_NVCC_EXECUTABLE CUDA_NVCC_FLAGS CUDA_NVCC_FLAGS_${_build_type})
+
+-# We would like to be helpful and reject the host compiler with a
+-# clear error message at configure time, rather than let nvcc
+-# later reject the host compiler as not supported when the first
+-# CUDA source file is built. We've implemented that for current
+-# nvcc running on Unix-like systems, but e.g. changes to nvcc
+-# will further affect the limited portability of this checking
+-# code. Set the CMake variable GMX_NVCC_WORKS on if you want to
+-# bypass this check.
+-if((_cuda_nvcc_executable_or_flags_changed OR CUDA_HOST_COMPILER_CHANGED OR NOT GMX_NVCC_WORKS) AND NOT WIN32)
+- message(STATUS "Check for working NVCC/C++ compiler combination with nvcc '${CUDA_NVCC_EXECUTABLE}'")
+- execute_process(COMMAND ${CUDA_NVCC_EXECUTABLE} -ccbin ${CUDA_HOST_COMPILER} -c ${CUDA_NVCC_FLAGS} ${CUDA_NVCC_FLAGS_${_build_type}} ${CMAKE_SOURCE_DIR}/cmake/TestCUDA.cu
+- RESULT_VARIABLE _cuda_test_res
+- OUTPUT_VARIABLE _cuda_test_out
+- ERROR_VARIABLE _cuda_test_err
+- OUTPUT_STRIP_TRAILING_WHITESPACE)
+-
+- if(${_cuda_test_res})
+- message(STATUS "Check for working NVCC/C compiler combination - broken")
+- message(STATUS "${CUDA_NVCC_EXECUTABLE} standard output: '${_cuda_test_out}'")
+- message(STATUS "${CUDA_NVCC_EXECUTABLE} standard error: '${_cuda_test_err}'")
+- if(${_cuda_test_err} MATCHES "nsupported")
+- message(FATAL_ERROR "NVCC/C++ compiler combination does not seem to be supported. CUDA frequently does not support the latest versions of the host compiler, so you might want to try an earlier C++ compiler version and make sure your CUDA compiler and driver are as recent as possible.")
+- else()
+- message(FATAL_ERROR "CUDA compiler does not seem to be functional.")
+- endif()
+- elseif(NOT GMX_CUDA_TEST_COMPILER_QUIETLY)
+- message(STATUS "Check for working NVCC/C++ compiler combination - works")
+- set(GMX_NVCC_WORKS TRUE CACHE INTERNAL "Nvcc can compile a trivial test program")
+- endif()
+-endif() # GMX_CHECK_NVCC
+-
+
+ # The flags are set as local variables which shadow the cache variables. The cache variables
+ # (can be set by the user) are appended. This is done in a macro to set the flags when all
diff --git a/sci-chemistry/gromacs/gromacs-2021.7.ebuild b/sci-chemistry/gromacs/gromacs-2021.7.ebuild
index 8cb244317617..1a6dbcab47f4 100644
--- a/sci-chemistry/gromacs/gromacs-2021.7.ebuild
+++ b/sci-chemistry/gromacs/gromacs-2021.7.ebuild
@@ -85,7 +85,10 @@ DOCS=( AUTHORS README )
RESTRICT="!test? ( test )"
-PATCHES=( "${FILESDIR}/${PN}-2021-musl-stdint.patch" )
+PATCHES=(
+ "${FILESDIR}/${PN}-2021-musl-stdint.patch"
+ "${FILESDIR}/${PN}-2021-cuda-detection.patch"
+)
if [[ ${PV} != *9999 ]]; then
S="${WORKDIR}/${PN}-${PV/_/-}"
diff --git a/sci-chemistry/gromacs/gromacs-2023.1.ebuild b/sci-chemistry/gromacs/gromacs-2023.1.ebuild
new file mode 100644
index 000000000000..98c909cdab48
--- /dev/null
+++ b/sci-chemistry/gromacs/gromacs-2023.1.ebuild
@@ -0,0 +1,331 @@
+# Copyright 1999-2023 Gentoo Authors
+# Distributed under the terms of the GNU General Public License v2
+
+EAPI=8
+
+CMAKE_MAKEFILE_GENERATOR="ninja"
+
+PYTHON_COMPAT=( python3_{9..11} )
+
+DISTUTILS_OPTIONAL=1
+DISTUTILS_USE_PEP517=no
+DISTUTILS_SINGLE_IMPL=1
+
+inherit bash-completion-r1 cmake cuda distutils-r1 flag-o-matic readme.gentoo-r1 toolchain-funcs xdg-utils
+
+if [[ ${PV} = *9999* ]]; then
+ EGIT_REPO_URI="
+ https://gitlab.com/gromacs/gromacs.git
+ https://github.com/gromacs/gromacs.git
+ git://git.gromacs.org/gromacs.git"
+ [[ ${PV} = 9999 ]] && EGIT_BRANCH="master" || EGIT_BRANCH="release-${PV:0:4}"
+ inherit git-r3
+else
+ SRC_URI="
+ https://ftp.gromacs.org/gromacs/${PN}-${PV/_/-}.tar.gz
+ doc? ( https://ftp.gromacs.org/manual/manual-${PV/_/-}.pdf )
+ test? ( https://ftp.gromacs.org/regressiontests/regressiontests-${PV/_/-}.tar.gz )"
+ # since 2022 arm support was dropped (but not arm64)
+ KEYWORDS="~amd64 -arm ~arm64 ~x86 ~amd64-linux ~x86-linux ~x64-macos"
+fi
+
+ACCE_IUSE="cpu_flags_x86_sse2 cpu_flags_x86_sse4_1 cpu_flags_x86_fma4 cpu_flags_x86_avx cpu_flags_x86_avx2 cpu_flags_x86_avx512f cpu_flags_arm_neon"
+
+DESCRIPTION="The ultimate molecular dynamics simulation package"
+HOMEPAGE="https://www.gromacs.org/"
+
+# see COPYING for details
+# https://repo.or.cz/w/gromacs.git/blob/HEAD:/COPYING
+# base, vmd plugins, fftpack from numpy, blas/lapck from netlib, memtestG80 library, mpi_thread lib
+LICENSE="LGPL-2.1 UoI-NCSA !mkl? ( !fftw? ( BSD ) !blas? ( BSD ) !lapack? ( BSD ) ) cuda? ( LGPL-3 ) threads? ( BSD )"
+SLOT="0/${PV}"
+IUSE="blas clang clang-cuda cuda +custom-cflags +doc build-manual double-precision +fftw +gmxapi +gmxapi-legacy +hwloc lapack mkl mpi +offensive opencl openmp +python +single-precision test +threads +tng ${ACCE_IUSE}"
+
+CDEPEND="
+ blas? ( virtual/blas )
+ cuda? ( >=dev-util/nvidia-cuda-toolkit-11[profiler] )
+ opencl? ( virtual/opencl )
+ fftw? ( sci-libs/fftw:3.0= )
+ hwloc? ( sys-apps/hwloc:= )
+ lapack? ( virtual/lapack )
+ mkl? ( sci-libs/mkl )
+ mpi? ( virtual/mpi[cxx] )
+ sci-libs/lmfit:=
+ >=dev-cpp/muParser-2.3:=
+ ${PYTHON_DEPS}
+ "
+BDEPEND="${CDEPEND}
+ virtual/pkgconfig
+ clang? ( >=sys-devel/clang-6:* )
+ build-manual? (
+ app-doc/doxygen
+ $(python_gen_cond_dep '
+ dev-python/sphinx[${PYTHON_USEDEP}]
+ ')
+ media-gfx/mscgen
+ media-gfx/graphviz
+ dev-texlive/texlive-latex
+ dev-texlive/texlive-latexextra
+ media-gfx/imagemagick
+ )"
+RDEPEND="${CDEPEND}"
+
+REQUIRED_USE="
+ || ( single-precision double-precision )
+ doc? ( !build-manual )
+ cuda? ( single-precision )
+ opencl? ( single-precision )
+ cuda? ( !opencl )
+ clang-cuda? ( clang cuda )
+ mkl? ( !blas !fftw !lapack )
+ ${PYTHON_REQUIRED_USE}"
+
+DOCS=( AUTHORS README )
+
+RESTRICT="!test? ( test )"
+
+if [[ ${PV} != *9999 ]]; then
+ S="${WORKDIR}/${PN}-${PV/_/-}"
+fi
+
+pkg_pretend() {
+ [[ ${MERGE_TYPE} != binary ]] && use openmp && tc-check-openmp
+}
+
+pkg_setup() {
+ [[ ${MERGE_TYPE} != binary ]] && use openmp && tc-check-openmp
+ python-single-r1_pkg_setup
+}
+
+src_unpack() {
+ if [[ ${PV} != *9999 ]]; then
+ default
+ else
+ git-r3_src_unpack
+ if use test; then
+ EGIT_REPO_URI="git://git.gromacs.org/regressiontests.git" \
+ EGIT_BRANCH="${EGIT_BRANCH}" \
+ EGIT_CHECKOUT_DIR="${WORKDIR}/regressiontests"\
+ git-r3_src_unpack
+ fi
+ fi
+}
+
+src_prepare() {
+ #notes/todos
+ # -on apple: there is framework support
+
+ xdg_environment_reset #591952
+
+ # we can use clang as default
+ if use clang && ! tc-is-clang ; then
+ export CC=${CHOST}-clang
+ export CXX=${CHOST}-clang++
+ else
+ tc-export CXX CC
+ fi
+ # clang-cuda need to filter mfpmath
+ if use clang-cuda ; then
+ filter-mfpmath sse
+ filter-mfpmath i386
+ fi
+
+ cmake_src_prepare
+
+ use cuda && cuda_src_prepare
+
+ GMX_DIRS=""
+ use single-precision && GMX_DIRS+=" float"
+ use double-precision && GMX_DIRS+=" double"
+
+ if use test; then
+ for x in ${GMX_DIRS}; do
+ mkdir -p "${WORKDIR}/${P}_${x}" || die
+ cp -al "${WORKDIR}/regressiontests"* "${WORKDIR}/${P}_${x}/tests" || die
+ done
+ fi
+
+ DOC_CONTENTS="Gromacs can use sci-chemistry/vmd to read additional file formats"
+ if use build-manual; then
+ # try to create policy for imagemagik
+ mkdir -p ${HOME}/.config/ImageMagick
+ cat >> ${HOME}/.config/ImageMagick/policy.xml <<- EOF
+ <?xml version="1.0" encoding="UTF-8"?>
+ <!DOCTYPE policymap [
+ <!ELEMENT policymap (policy)+>
+ !ATTLIST policymap xmlns CDATA #FIXED ''>
+ <!ELEMENT policy EMPTY>
+ <!ATTLIST policy xmlns CDATA #FIXED '' domain NMTOKEN #REQUIRED
+ name NMTOKEN #IMPLIED pattern CDATA #IMPLIED rights NMTOKEN #IMPLIED
+ stealth NMTOKEN #IMPLIED value CDATA #IMPLIED>
+ ]>
+ <policymap>
+ <policy domain="coder" rights="read | write" pattern="PS" />
+ <policy domain="coder" rights="read | write" pattern="PS2" />
+ <policy domain="coder" rights="read | write" pattern="PS3" />
+ <policy domain="coder" rights="read | write" pattern="EPS" />
+ <policy domain="coder" rights="read | write" pattern="PDF" />
+ <policy domain="coder" rights="read | write" pattern="XPS" />
+ </policymap>
+ EOF
+ fi
+}
+
+src_configure() {
+ local mycmakeargs_pre=( ) extra fft_opts=( )
+ local acce="AUTO"
+
+ if use custom-cflags; then
+ #go from slowest to fastest acceleration
+ acce="None"
+ if (use amd64 || use x86); then
+ use cpu_flags_x86_sse2 && acce="SSE2"
+ use cpu_flags_x86_sse4_1 && acce="SSE4.1"
+ use cpu_flags_x86_fma4 && acce="AVX_128_FMA"
+ use cpu_flags_x86_avx && acce="AVX_256"
+ use cpu_flags_x86_avx2 && acce="AVX2_256"
+ use cpu_flags_x86_avx512f && acce="AVX_512"
+ elif (use arm); then
+ use cpu_flags_arm_neon && acce="ARM_NEON"
+ elif (use arm64); then
+ use cpu_flags_arm_neon && acce="ARM_NEON_ASIMD"
+ fi
+ else
+ strip-flags
+ fi
+
+ #to create man pages, build tree binaries are executed (bug #398437)
+ [[ ${CHOST} = *-darwin* ]] && \
+ extra+=" -DCMAKE_BUILD_WITH_INSTALL_RPATH=OFF"
+
+ if use fftw; then
+ fft_opts=( -DGMX_FFT_LIBRARY=fftw3 )
+ elif use mkl; then
+ local bits=$(get_libdir)
+ fft_opts=( -DGMX_FFT_LIBRARY=mkl
+ -DMKL_INCLUDE_DIR="$(echo /opt/intel/*/mkl/include)"
+ -DMKL_LIBRARIES="$(echo /opt/intel/*/mkl/lib/*${bits/lib}/libmkl_rt.so)"
+ )
+ else
+ fft_opts=( -DGMX_FFT_LIBRARY=fftpack )
+ fi
+
+ mycmakeargs_pre+=(
+ "${fft_opts[@]}"
+ "${lmfit_opts[@]}"
+ -DGMX_USE_LMFIT=EXTERNAL
+ -DGMX_USE_MUPARSER=EXTERNAL
+ -DGMX_EXTERNAL_BLAS=$(usex blas)
+ -DGMX_EXTERNAL_LAPACK=$(usex lapack)
+ -DGMX_OPENMP=$(usex openmp)
+ -DGMX_COOL_QUOTES=$(usex offensive)
+ -DGMX_USE_TNG=$(usex tng)
+ -DGMX_BUILD_MANUAL=$(usex build-manual)
+ -DGMX_HWLOC=$(usex hwloc)
+ -DGMX_DEFAULT_SUFFIX=off
+ -DGMX_SIMD="$acce"
+ -DGMX_VMD_PLUGIN_PATH="${EPREFIX}/usr/$(get_libdir)/vmd/plugins/*/molfile/"
+ -DBUILD_TESTING=$(usex test)
+ -DGMX_BUILD_UNITTESTS=$(usex test)
+ -DPYTHON_EXECUTABLE="${EPREFIX}/usr/bin/${EPYTHON}"
+ ${extra}
+ )
+
+ for x in ${GMX_DIRS}; do
+ einfo "Configuring for ${x} precision"
+ local suffix=""
+ #if we build single and double - double is suffixed
+ use double-precision && use single-precision && \
+ [[ ${x} = "double" ]] && suffix="_d"
+ local p
+ [[ ${x} = "double" ]] && p="-DGMX_DOUBLE=ON" || p="-DGMX_DOUBLE=OFF"
+ local gpu=( "-DGMX_GPU=OFF" )
+ [[ ${x} = "float" ]] && use cuda && gpu=( "-DGMX_GPU=CUDA" )
+ [[ ${x} = "float" ]] && use clang-cuda && gpu=( "-DGMX_GPU=CUDA" "-DGMX_CLANG_CUDA=ON" )
+ use opencl && gpu=( "-DGMX_GPU=OPENCL" )
+ local mycmakeargs=(
+ ${mycmakeargs_pre[@]} ${p}
+ -DGMX_MPI=$(usex mpi)
+ -DGMX_THREAD_MPI=$(usex threads)
+ -DGMXAPI=$(usex gmxapi)
+ -DGMX_INSTALL_LEGACY_API=$(usex gmxapi-legacy)
+ "${gpu[@]}"
+ "$(use test && echo -DREGRESSIONTEST_PATH="${WORKDIR}/${P}_${x}/tests")"
+ -DGMX_BINARY_SUFFIX="${suffix}"
+ -DGMX_LIBS_SUFFIX="${suffix}"
+ -DGMX_PYTHON_PACKAGE=$(usex python)
+ )
+ BUILD_DIR="${WORKDIR}/${P}_${x}" cmake_src_configure
+ [[ ${CHOST} != *-darwin* ]] || \
+ sed -i '/SET(CMAKE_INSTALL_NAME_DIR/s/^/#/' "${WORKDIR}/${P}_${x}/gentoo_rules.cmake" || die
+ done
+}
+
+src_compile() {
+ for x in ${GMX_DIRS}; do
+ einfo "Compiling for ${x} precision"
+ BUILD_DIR="${WORKDIR}/${P}_${x}"\
+ cmake_src_compile
+ if use python; then
+ BUILD_DIR="${WORKDIR}/${P}_${x}"\
+ cmake_src_compile python_packaging/all
+ BUILD_DIR="${WORKDIR}/${P}" \
+ distutils-r1_src_compile
+ fi
+ # not 100% necessary for rel ebuilds as available from website
+ if use build-manual; then
+ BUILD_DIR="${WORKDIR}/${P}_${x}"\
+ cmake_src_compile manual
+ fi
+ done
+}
+
+src_test() {
+ for x in ${GMX_DIRS}; do
+ BUILD_DIR="${WORKDIR}/${P}_${x}"\
+ cmake_src_compile check
+ done
+}
+
+src_install() {
+ for x in ${GMX_DIRS}; do
+ BUILD_DIR="${WORKDIR}/${P}_${x}" \
+ cmake_src_install
+ if use python; then
+ BUILD_DIR="${WORKDIR}/${P}_${x}" \
+ cmake_src_install python_packaging/install
+ fi
+ if use build-manual; then
+ newdoc "${WORKDIR}/${P}_${x}"/docs/manual/gromacs.pdf "${PN}-manual-${PV}.pdf"
+ fi
+
+ if use doc; then
+ if [[ ${PV} != *9999* ]]; then
+ newdoc "${DISTDIR}/manual-${PV/_/-}.pdf" "${PN}-manual-${PV}.pdf"
+ fi
+ fi
+ done
+
+ if use tng; then
+ insinto /usr/include/tng
+ doins src/external/tng_io/include/tng/*h
+ fi
+ # drop unneeded stuff
+ rm "${ED}"/usr/bin/GMXRC* || die
+ for x in "${ED}"/usr/bin/gmx-completion-*.bash ; do
+ local n=${x##*/gmx-completion-}
+ n="${n%.bash}"
+ cat "${ED}"/usr/bin/gmx-completion.bash "$x" > "${T}/${n}" || die
+ newbashcomp "${T}"/"${n}" "${n}"
+ done
+ rm "${ED}"/usr/bin/gmx-completion*.bash || die
+ readme.gentoo_create_doc
+}
+
+pkg_postinst() {
+ einfo
+ einfo "Please read and cite gromacs related papers from list:"
+ einfo "https://www.gromacs.org/Gromacs_papers"
+ einfo
+ readme.gentoo_print_elog
+}