Hi Mayukh,
I have tested compiling Amber24 with CUDA 11.7.0 and PyTorch 2.0.1. The
compilation and build process proceeded without any issues, and make
test.serial passed without any stopping errors. Please ensure you use the
version of libtorch that corresponds to the specific CUDA version. However,
note that the pbsa_test did not involve libtorch or MLSES. Libtorch should
still function even if you encounter errors with pbsa_test. Here are the
specific variables I used:
-DLIBTORCH=ON \
-DCUDA_HOME=$CUDA_PATH/cuda_11.7.0 \
-DTORCH_HOME=$LIBTORCH_PATH/libtorch \
-DCUDA_TOOLKIT_ROOT_DIR=/$CUDA_PATH/cuda_11.7.0 \
-DCUDNN=TRUE \
-DCUDNN_INCLUDE_PATH=$CUDNN_PATH/include \
-DCUDNN_LIBRARY_PATH=$CUDNN_PATH/lib/libcudnn.so \
Best regards,
Yongxian Wu
On Wed, May 8, 2024 at 2:26 PM Chakrabarti, Mayukh (NIH/NCI) [C] via AMBER <
amber.ambermd.org> wrote:
> Hi Ray and Yongxian,
>
> Thank you for your response. I don’t have CUDA 11.3 available on my
> system, but I do have CUDA 11.8, which I tried to use instead. Upon your
> advice, I tried to re-compile Amber 24 & AmberTools 24 using an updated
> version of libtorch, version 2.1.0 with CUDA 11.8. I also used
> “cudnn-linux-x86_64-8.7.0.84_cuda11-archive.tar.xz”. Unfortunately,
> following compilation, I still encounter the same segmentation fault errors
> (SIGSEGV) in the pbsa binaries. I will next try with PyTorch 1.12.1 and
> CUDA 10.2, and report back.
>
> Best,
>
> Mayukh Chakrabarti (he/him)
> COMPUTATIONAL SCIENTIST
>
> From: Ray Luo <rluo.uci.edu>
> Date: Wednesday, May 8, 2024 at 1:04 PM
> To: Chakrabarti, Mayukh (NIH/NCI) [C] <mayukh.chakrabarti.nih.gov>, AMBER
> Mailing List <amber.ambermd.org>
> Subject: [EXTERNAL] Re: [AMBER] Amber22/AmberTools23: Enabling of libtorch
> & cudnn libraries breaks pbsa binaries
> Hi Mayukh,
>
> The issue might be due to the version of PyTorch being used.
>
> I recommend using the same version of libtorch that we currently have,
> which is version 1.12.1 with CUDA 11.3. PyTorch 1.9 with CUDA 10.2 is
> outdated. However, PyTorch 1.12.1 also supports CUDA 10.2, although we
> recommend using CUDA 11 since CUDA 10 is outdated.
>
> Let me know if there are any other questions.
>
> Best,
> Yongxian and Ray
> --
> Ray Luo, Ph.D.
> Professor of Structural Biology/Biochemistry/Biophysics,
> Chemical and Materials Physics, Chemical and Biomolecular Engineering,
> Biomedical Engineering, and Materials Science and Engineering
> Department of Molecular Biology and Biochemistry
> University of California, Irvine, CA 92697-3900
>
>
> On Tue, May 7, 2024 at 8:57 AM Chakrabarti, Mayukh (NIH/NCI) [C] via AMBER
> <amber.ambermd.org<mailto:amber.ambermd.org>> wrote:
> Hello,
>
> I wanted to update my report below to mention that the issue with pbsa
> binaries breaking upon enabling the LibTorch & cudnn libraries still
> persists in Amber24 with AmberTools 24. I compiled with CUDA 10.2, gcc 8.5,
> OpenMPI 4.1.5, python 3.10, and cmake 3.25.2 on a Red Hat Enterprise Linux
> release 8.8 (Ootpa) system. I am not aware of any workaround or resolution
> for this issue.
>
> Best,
>
> Mayukh Chakrabarti (he/him)
> COMPUTATIONAL SCIENTIST
>
> From: Chakrabarti, Mayukh (NIH/NCI) [C] <mayukh.chakrabarti.nih.gov
> <mailto:mayukh.chakrabarti.nih.gov>>
> Date: Tuesday, October 31, 2023 at 11:28 AM
> To: amber.ambermd.org<mailto:amber.ambermd.org> <amber.ambermd.org<mailto:
> amber.ambermd.org>>
> Subject: Amber22/AmberTools23: Enabling of libtorch & cudnn libraries
> breaks pbsa binaries
> Hello,
>
> I am encountering a problem in which enabling LibTorch libraries with
> Amber22 and AmberTools23 causes segmentation fault errors (SIGSEGV) in the
> pbsa binaries upon running the serial tests (make test.serial). I have
> successfully compiled a version of Amber22/AmberTools23 in which this
> library is not enabled, and none of the pbsa tests in AmberTools break
> (i.e., no segmentation fault errors).
>
> Further details:
>
> I am running my compilation with CUDA 10.2, gcc 8.5, OpenMPI 4.1.5, python
> 3.6, and cmake 3.25.1 on a Red Hat Enterprise Linux release 8.8 (Ootpa)
> system. As per the manual, I have tried both “Built-in” mode and
> “User-installed” mode, both resulting in the same errors. For the
> “User-installed” mode, I manually downloaded and extracted
> “libtorch-shared-with-deps-1.9.1+cu102.zip” from the PyTorch website to
> correspond to CUDA 10.2, and
> “cudnn-linux-x86_64-8.7.0.84_cuda10-archive.tar.xz” directly from the
> NVIDIA website, and specified the following variables to CMAKE:
>
>
> -DLIBTORCH=ON \
>
> -DTORCH_HOME=/path_to_libtorch \
>
> -DCUDNN=TRUE \
>
> -DCAFFE2_USE_CUDNN=1 \
>
> -DCUDNN_INCLUDE_PATH=/path_to_cudnn_include \
>
> -DCUDNN_LIBRARY_PATH=/path_to_libcudnn.so \
>
>
> Building and compiling proceed without issue. However, when running the
> serial tests, I get errors akin to the following (example shown after
> sourcing amber.sh and running AmberTools pbsa_ligand test):
>
> Program received signal SIGSEGV: Segmentation fault - invalid memory
> reference.
>
> Backtrace for this error:
> #0 0x7f480cb65171 in ???
> #1 0x7f480cb64313 in ???
> #2 0x7f480c221c0f in ???
> #3 0x7f4869667219 in ???
> #4 0x7f480c224856 in ???
> #5 0x7f4869648722 in ???
> ./Run.t4bnz.min: line 34: 2436022 Segmentation fault (core dumped)
> $DO_PARALLEL $TESTpbsa -O -i min.in<
> https://urldefense.com/v3/__http://min.in/__;!!CzAuKJ42GuquVTTmVmPViYEvSg!NC0Jwxkjnbm_mXLPxsbm3fva29jev44srLmSv7hv0XZuaw716Egbv1itPhJek6uNYpgJ3GbSbeK6jyZvAA$
> > -o $output < /dev/null
> ./Run.t4bnz.min: Program error
>
> When running the exact same test with the version of Amber22 not
> containing the LibTorch libraries:
>
> diffing mdout.lig.min.save with mdout.lig.min
> PASSED
> ==============================================================
>
> I have run ‘ldd’ on the pbsa binary to try to identify any libraries that
> may be missing, but there don’t appear to be any missing libraries. Could
> anybody please provide any insight into how to fix this issue?
>
> Best,
>
> Mayukh Chakrabarti
> COMPUTATIONAL SCIENTIST
>
>
>
>
>
>
>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org<mailto:AMBER.ambermd.org>
>
> https://urldefense.com/v3/__http://lists.ambermd.org/mailman/listinfo/amber__;!!CzAuKJ42GuquVTTmVmPViYEvSg!PLN2_ZzXkAP-XJnmppjnXaq9ruj3FNFdRR163QR8ehrjX6YYBOUfKc8Eo0t0w7N8I3CtOwXatyYNzg$
> CAUTION: This email originated from outside of the organization. Do not
> click links or open attachments unless you recognize the sender and are
> confident the content is safe.
>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
>
> https://urldefense.com/v3/__http://lists.ambermd.org/mailman/listinfo/amber__;!!CzAuKJ42GuquVTTmVmPViYEvSg!NC0Jwxkjnbm_mXLPxsbm3fva29jev44srLmSv7hv0XZuaw716Egbv1itPhJek6uNYpgJ3GbSbeI1bOw3SQ$
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Thu May 09 2024 - 23:00:02 PDT