Message boards : Number crunching : New app update (acemd3)
Author | Message |
---|---|
I am testing the new acemd3 app. The app is entirely new: faster and more general. The idea is to replace the old one asap. We'll also try to make it more maintainable (a long standing issue) using the boinc wrapper. | |
ID: 51939 | Rating: 0 | rate: / Reply Quote | |
do you mean this one? crunched in 6 or 7 minutes. | |
ID: 51940 | Rating: 0 | rate: / Reply Quote | |
How to select work for the new app? "New version of ACEMD" app is not a choice under project preferences. The current choices are only these:
ACEMD long runs (8-12 hours on fastest GPU) ACEMD Beta Quantum Chemistry (CPU) Quantum Chemistry (CPU, beta) Python Runtime
| |
ID: 51942 | Rating: 0 | rate: / Reply Quote | |
So far we got one success, i.e. 20962989. The other 5 Test tasks seem "stuck". They have been in progress now for quite a while. They must be really long, have errored, or hosts have downloaded the tasks and then been turned off. Can our Linux crunchers check your Linux hosts for progress? EDIT: The successful task above has also been accepted by 2 Windows Hosts ("New version of ACEMD v1.19" but failed. Also failed on 2 Linux hosts "New version of ACEMD v2.00"). So it seems the Test tasks are being accepted by Windows and Linux hosts. The successful Linux Host has Nvidia driver v430.14. The failed hosts had Nvidia drivers ranging from v375.70 to v418.19. | |
ID: 51947 | Rating: 0 | rate: / Reply Quote | |
How to select work for the new app? "New version of ACEMD" app is not a choice under project preferences. The current choices are only these: I've just selected everything including test apps with only Use GPUs selected. Nothing yet though but I would think that should be enough. Devs can sneak in about anything under the test apps options. | |
ID: 51948 | Rating: 0 | rate: / Reply Quote | |
probably my next build (in 6-10months) will be 4, 5 or 6 rtx cards. hopefully is the app then mature enough for investing couple of thousand euro for gpugrid | |
ID: 51957 | Rating: 0 | rate: / Reply Quote | |
The number of failures, and the existence of one success, is odd. Doesn't seem to be explained by driver versions alone. | |
ID: 51966 | Rating: 0 | rate: / Reply Quote | |
Try sending out more experimental WUs and see if it is one driver version | |
ID: 51967 | Rating: 0 | rate: / Reply Quote | |
Recent changes: | |
ID: 51968 | Rating: 0 | rate: / Reply Quote | |
Errored WUs on multiple different drivers and OS's | |
ID: 51969 | Rating: 0 | rate: / Reply Quote | |
Multiple failures of this task on both windows and linux <message> process exited with code 195 (0xc3, -61)</message> <stderr_txt> 15:19:27 (30109): wrapper (7.7.26016): starting 15:19:27 (30109): wrapper (7.7.26016): starting 15:19:27 (30109): wrapper: running acemd3 (--boinc input --device 0) # Engine failed: Error launching CUDA compiler: 32512 sh: 1: : Permission denied 15:19:28 (30109): acemd3 exited; CPU time 0.186092 15:19:28 (30109): app exit status: 0x1 15:19:28 (30109): called boinc_finish(195) </stderr_txt> Why is the app launching CUDA compiler? | |
ID: 51970 | Rating: 0 | rate: / Reply Quote | |
My host 43404 got one of WU 16517259. 03/06/2019 15:36:58 | GPUGRID | Starting task a27-TONI_TEST3-0-1-RND0985_6 03/06/2019 15:36:58 | GPUGRID | [cpu_sched] Starting task a27-TONI_TEST3-0-1-RND0985_6 using acemd3 version 119 (cuda80) in slot 0 03/06/2019 15:36:59 | GPUGRID | [sched_op] Deferring communication for 00:01:03 03/06/2019 15:36:59 | GPUGRID | [sched_op] Reason: Unrecoverable error for task a27-TONI_TEST3-0-1-RND0985_6 03/06/2019 15:36:59 | GPUGRID | Computation for task a27-TONI_TEST3-0-1-RND0985_6 finished 03/06/2019 15:36:59 | GPUGRID | Output file a27-TONI_TEST3-0-1-RND0985_6_0 for task a27-TONI_TEST3-0-1-RND0985_6 absent 03/06/2019 15:36:59 | GPUGRID | Output file a27-TONI_TEST3-0-1-RND0985_6_9 for task a27-TONI_TEST3-0-1-RND0985_6 absent with no further information than Incorrect function. (0x1) - exit code 1 (0x1) But I did capture all the specifications and downloaded files between download and run, so I can recreate the attempt offline and see what additional crash information I can collect. May take me a little time... Windows 7/64, GTX 970, runs v9.22 just fine. Edit - all I can get in offline runs is ACEMD can run with Boinc only! - even when I supply a dummy init_data.xml file which has worked in other standalone test environments. I'll go out for a walk and see if that activates the little grey cells. | |
ID: 51971 | Rating: 0 | rate: / Reply Quote | |
Recent changes: Can you please explain which app we have to select in our project preferences to get these tasks? The app name "New version of ACEMD" is not a an option in the project preferences. ____________ Reno, NV Team: SETI.USA | |
ID: 51972 | Rating: 0 | rate: / Reply Quote | |
Should be called "ACEMD3 Beta". It's for Linux only (for now). | |
ID: 51973 | Rating: 0 | rate: / Reply Quote | |
Can you please explain which app we have to select in our project preferences to get these tasks? The app name "New version of ACEMD" is not a an option in the project preferences. The computer I got a test app on has If no work for selected applications is available, accept work from other applications? yesNothing else out of the ordinary. The app name appeared as 'acemd3'. | |
ID: 51974 | Rating: 0 | rate: / Reply Quote | |
I got 1 task but it failed.:-( | |
ID: 51975 | Rating: 0 | rate: / Reply Quote | |
All but 2 of the libraries that were downloaded are marked as executable. Should libgcc and libOpenCL also be executable? -rwxr-xr-x 1 boinc boinc 425056 Jun 3 11:52 libcudart.so.8.0.61.46fcfd92ffc5c805d076b5e2b17e9647
-rwxr-xr-x 1 boinc boinc 146772120 Jun 3 11:56 libcufft.so.8.0.61.b142ab8797d534b619ef19c7e98cffc7
-rwxr-xr-x 1 boinc boinc 1647707 Jun 3 11:53 libfftw3f.so.3.4.4.a4580ddf9efebaad56fab49847a8c899
-rwxr-xr-x 1 boinc boinc 31467 Jun 3 11:52 libfftw3f_threads.so.3.4.4.dd0c6fcfa550371acf730db2d9d5a270
-rw-r--r-- 1 boinc boinc 819744 Jun 3 11:52 libgcc_s.so.1.d7f787a9bf6c3633eaebb9015c6d9044
-rwxr-xr-x 1 boinc boinc 937656 Jun 3 11:52 libgomp.so.1.0.0.efdf718669edc7fff00e0c5f7f0b8791
-rwxr-xr-x 1 boinc boinc 9659424 Jun 3 11:54 libnvrtc-builtins.so.8.0.61.ef79235263e650333dd8c573faa47432
-rwxr-xr-x 1 boinc boinc 18517368 Jun 3 11:54 libnvrtc.so.8.0.61.1ac77468cd8086b8cd1a6c855da50f8c
-rw-r--r-- 1 boinc boinc 31696 Jun 3 11:52 libOpenCL.so.1.0.0.343dee45a7d7eb4b9016b6cd9d1bd8d5
-rwxr-xr-x 1 boinc boinc 655240 Jun 3 11:54 libOpenMMCPU.so.19849b4ff1cf4d33f75d9433b4d5c6bb
-rwxr-xr-x 1 boinc boinc 37096 Jun 3 11:53 libOpenMMCudaCompiler.so.aaed781fe4caa9d1099312d458a9b902
-rwxr-xr-x 1 boinc boinc 2774560 Jun 3 11:52 libOpenMMCUDA.so.8867021fdc0daf2e39f1b7228ece45af
-rwxr-xr-x 1 boinc boinc 2979224 Jun 3 11:52 libOpenMMOpenCL.so.6a31fa1ff5ae3a26ea64f2abfb5a66cc
-rwxr-xr-x 1 boinc boinc 80808 Jun 3 11:53 libOpenMMPME.so.3208e45e71567824e8390ab1c79c6a66
-rwxr-xr-x 1 boinc boinc 4062370 Jun 3 11:53 libOpenMM.so.5406dfd716045d08ad6369e2399a98e2
-rwxr-xr-x 1 boinc boinc 9536208 Jun 3 11:54 libstdc++.so.6.0.25.e344f48acfbd4f5abbf99b2c75cc5e50 | |
ID: 51976 | Rating: 0 | rate: / Reply Quote | |
regarding the error on my task: | |
ID: 51977 | Rating: 0 | rate: / Reply Quote | |
http://www.gpugrid.net/result.php?resultid=20974104 | |
ID: 51978 | Rating: 0 | rate: / Reply Quote | |
Hi Toni | |
ID: 51979 | Rating: 0 | rate: / Reply Quote | |
name a11-TONI_TEST3-0-1-RND0663 | |
ID: 51980 | Rating: 0 | rate: / Reply Quote | |
I think I debugged it (app version 201). 100 new WUs sent. Progress bar should also work (please report if not). | |
ID: 51981 | Rating: 0 | rate: / Reply Quote | |
There are many more successes now. | |
ID: 51982 | Rating: 0 | rate: / Reply Quote | |
There are many more successes now. Hello Toni, I have received many successes and when I typed "nvcc -V" to verify the CUDA Toolkit version, it says "The program 'nvcc' is currently not installed. You can install it by typing: sudo apt install nvidia-cuda-toolkit" My system seems to not have it installed. This is the list of the successful tasks: http://www.gpugrid.net/results.php?userid=306281 | |
ID: 51983 | Rating: 0 | rate: / Reply Quote | |
This is the list of the successful tasks: http://www.gpugrid.net/results.php?userid=306281 access denied :-( | |
ID: 51984 | Rating: 0 | rate: / Reply Quote | |
This is the list of the successful tasks: http://www.gpugrid.net/results.php?userid=306281 Perhaps you can view a single WU? http://www.gpugrid.net/result.php?resultid=20978809 | |
ID: 51985 | Rating: 0 | rate: / Reply Quote | |
I have 2 machines. Both have linux mint 19.1 installed, same nvidia driver (390.116), cuda toolkit release 9.1 (both tested as functional), same boinc version 7.14.2. | |
ID: 51986 | Rating: 0 | rate: / Reply Quote | |
Question for anybody who has seen a success: do you have the CUDA Toolkit installed? No. I installed the Nvidia 430.14 drivers as Linux metapackages. According to the Synaptic Package Manager I do not have the CUDA Toolkit installed. ____________ | |
ID: 51987 | Rating: 0 | rate: / Reply Quote | |
I have 2 machines. Both have linux mint 19.1 installed, same nvidia driver (390.116), cuda toolkit release 9.1 (both tested as functional), same boinc version 7.14.2. I can't find anything in the logs. I was running Rosetta on the machine that had the failed GPUGrid tasks. There was no other project running on the machine that had the successful GPUGrid tasks. | |
ID: 51988 | Rating: 0 | rate: / Reply Quote | |
There are many more successes now. Even though nvcc is actually present on my Jetson Nano, nvcc -V yielded program not found. It is located at /usr/local/cuda-10.0/bin/nvcc I had to export the directory where nvcc was located for it to be found. That enabled a program to find nvcc. keith@Nano:~$ nvcc -V nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2018 NVIDIA Corporation Built on Sun_Sep_30_21:09:22_CDT_2018 Cuda compilation tools, release 10.0, V10.0.166 But as soon as I rebooted, nvcc could not be found. So I ended up adding the library directory as an export in .bashrc and then I could find nvcc after reboots. | |
ID: 51989 | Rating: 0 | rate: / Reply Quote | |
I completed one while 5 others had errors. | |
ID: 51990 | Rating: 0 | rate: / Reply Quote | |
Another option is to place the cuda library path in a file in /etc/ld.so.conf.d. you could name the file cuda.conf then: sudo ldconfig | |
ID: 51991 | Rating: 0 | rate: / Reply Quote | |
Correct. That is the other method I researched as a popular solution. So am I correct in understanding now is that one has to install the CUDA toolkit to run the new acemd application? That the wrapper download itself is insufficient? | |
ID: 51992 | Rating: 0 | rate: / Reply Quote | |
Hi all, thanks for the reports. | |
ID: 51993 | Rating: 0 | rate: / Reply Quote | |
Well, I see I attempted to run a task that failed on one host. I looked over all the downloaded files and thought to do a sanity check on the executable. This is what ldd showed. | |
ID: 51994 | Rating: 0 | rate: / Reply Quote | |
You don't (shouldn't) need to install any additional software, if everything works as intended (not the wrapper, nor the cuda toolkit). You may need to update the drivers, though. | |
ID: 51995 | Rating: 0 | rate: / Reply Quote | |
libcudart.so.8.0 => not found If somebody can post or upload the three components of a test workunit specification: * <app_version> * <workunit> * <result> all from client_state.xml - make sure you get the right (latest) version of <app_version>, there will be several of them - I can proofread that there are no bugs in the BOINC deployment of the app files. This one could be a problem with the version renaming or copying. | |
ID: 51996 | Rating: 0 | rate: / Reply Quote | |
Hi all, thanks for the reports. No, the boinc installation method is the same (repository meta package) and gcc is installed on both machines (build-essential package). I ran ldd on wrapper_26198_x86_64-pc-linux-gnu and acemd3.e72153abf98cb1fcd0f05fc443818dfc on both machines and the output is identical. Working machine: mark@x20-linux:/var/lib/boinc/projects/www.gpugrid.net$ ldd ./wrapper_26198_x86_64-pc-linux-gnu
linux-vdso.so.1 (0x00007ffc1bfab000)
libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007f7ab23ba000)
libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007f7ab21a2000)
libpthread.so.0 => /lib/x86_64-linux-gnu/libpthread.so.0 (0x00007f7ab1f83000)
libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007f7ab1b92000)
/lib64/ld-linux-x86-64.so.2 (0x00007f7ab2758000)
mark@x20-linux:/var/lib/boinc/projects/www.gpugrid.net$ ldd ./acemd3.e72153abf98cb1fcd0f05fc443818dfc
linux-vdso.so.1 (0x00007ffda9bfe000)
libOpenMM.so => not found
libdl.so.2 => /lib/x86_64-linux-gnu/libdl.so.2 (0x00007ffb4cb37000)
libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007ffb4c7ae000)
libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007ffb4c410000)
libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007ffb4c1f8000)
libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007ffb4be07000)
/lib64/ld-linux-x86-64.so.2 (0x00007ffb4cd3b000) machine with failures: mark@x16-linux:/var/lib/boinc/projects/www.gpugrid.net$ ldd ./wrapper_26198_x86_64-pc-linux-gnu
linux-vdso.so.1 (0x00007ffd96952000)
libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007fd300b09000)
libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007fd3008f1000)
libpthread.so.0 => /lib/x86_64-linux-gnu/libpthread.so.0 (0x00007fd3006d2000)
libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007fd3002e1000)
/lib64/ld-linux-x86-64.so.2 (0x00007fd300ea7000)
mark@x16-linux:/var/lib/boinc/projects/www.gpugrid.net$ ldd ./acemd3.e72153abf98cb1fcd0f05fc443818dfc
linux-vdso.so.1 (0x00007ffef0097000)
libOpenMM.so => not found
libdl.so.2 => /lib/x86_64-linux-gnu/libdl.so.2 (0x00007fabe9b83000)
libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007fabe97fa000)
libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007fabe945c000)
libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007fabe9244000)
libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007fabe8e53000)
/lib64/ld-linux-x86-64.so.2 (0x00007fabe9d87000) | |
ID: 51997 | Rating: 0 | rate: / Reply Quote | |
We are distributing the two files with the app. They are copied (via copy_file) into the slot, and the slot is added to LD_LIBRARY_PATH. It works locally and on many machines; I am inclined to think it's not the problem. The "permission denied" bit seems related to a later stage, possibly an attempt to compile the cuda bytecode into the form necessary for the specific graphic card (done via nvrtc). If anybody is able to capture the "progress.log" file before it's deleted, thanks! T | |
ID: 51998 | Rating: 0 | rate: / Reply Quote | |
I did find a "messy" install of the nvidia driver on the offending machine. There seems to be remnants of a driver installed via download directly from nvidia. I'll clean that up. | |
ID: 51999 | Rating: 0 | rate: / Reply Quote | |
I did find a "messy" install of the nvidia driver on the offending machine. There seems to be remnants of a driver installed via download directly from nvidia. I'll clean that up. From what I know, "apt search" does not look at the packages installed in your system but those "accessible" online. So, the difference may be in the repository configurations. | |
ID: 52000 | Rating: 0 | rate: / Reply Quote | |
Yeah, dpkg -l | grep -i nvidia is the right command. I went ahead and purged everything nvidia and reinstalled the nvidia driver. I didn't install the cuda toolkit though. UPDATE: tasks still failing on this machine. | |
ID: 52001 | Rating: 0 | rate: / Reply Quote | |
Toni
Slotfolder 40 zip: https://filebin.net/jfv8ec4c6q8uszuw/Slot_40.zip?t=tvn13kdj On failed host slot folder are empty. Boinc wipe at crash or application never add files to slotfolder. I could not grab progress.log. Task failed after in 1 sec is impossible to grab and it doesnt store to upload so it wiped out. Getting error this on older os 16.04 with GTX970 driver: 418.56. Same drivers hand out valid task on later system 18.10. So it looks to be on system not driver version. This compile issue still exist on latest application but only effect old system. <core_client_version>7.6.31</core_client_version> <![CDATA[ <message> process exited with code 195 (0xc3, -61) </message> <stderr_txt> 14:22:07 (102554): wrapper (7.7.26016): starting 14:22:07 (102554): wrapper (7.7.26016): starting 14:22:07 (102554): wrapper: running acemd3 (--boinc input --device 0) # Engine failed: Error launching CUDA compiler: 32512 sh: 1: : Permission denied 14:22:08 (102554): acemd3 exited; CPU time 0.132000 14:22:08 (102554): app exit status: 0x1 14:22:08 (102554): called boinc_finish(195) </stderr_txt> ]]> | |
ID: 52002 | Rating: 0 | rate: / Reply Quote | |
Aehm, to clarify: I see the process.log file of successful tasks only. | |
ID: 52003 | Rating: 0 | rate: / Reply Quote | |
If anybody is so inclined, can they try to run the boinc client manually with the --exit_after_finish flag, so the slot directory is preserved on failure? | |
ID: 52004 | Rating: 0 | rate: / Reply Quote | |
There is information in <app_version> but nothing for <workunit> or <result <app_version> <app_name>acemd3</app_name> <version_num>202</version_num> <platform>x86_64-pc-linux-gnu</platform> <avg_ncpus>0.987442</avg_ncpus> <flops>28742507251613.187500</flops> <plan_class>cuda80</plan_class> <api_version>7.7.0</api_version> <file_ref> <file_name>wrapper_26198_x86_64-pc-linux-gnu</file_name> <main_program/> </file_ref> <file_ref> <file_name>acemd3.e72153abf98cb1fcd0f05fc443818dfc</file_name> <open_name>acemd3</open_name> <copy_file/> </file_ref> <file_ref> <file_name>job.xml.1245cc127550a015dcc9b3e1c2c84e13</file_name> <open_name>job.xml</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMMOpenCL.so.6a31fa1ff5ae3a26ea64f2abfb5a66cc</file_name> <open_name>libOpenMMOpenCL.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenCL.so.1.0.0.43d4300566ce59d77e0fa316f8ee5b02</file_name> <open_name>libOpenCL.so.1</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libgomp.so.1.0.0.efdf718669edc7fff00e0c5f7f0b8791</file_name> <open_name>libgomp.so.1.0.0</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMM.so.5406dfd716045d08ad6369e2399a98e2</file_name> <open_name>libOpenMM.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMMCUDA.so.8867021fdc0daf2e39f1b7228ece45af</file_name> <open_name>libOpenMMCUDA.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libcudart.so.8.0.61.af43be839e6366e731accc514633bd1f</file_name> <open_name>libcudart.so.8.0</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libfftw3f_threads.so.3.4.4.dd0c6fcfa550371acf730db2d9d5a270</file_name> <open_name>libfftw3f_threads.so.3</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libgcc_s.so.1.d7f787a9bf6c3633eaebb9015c6d9044</file_name> <open_name>libgcc_s.so.1</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libnvrtc-builtins.so.8.0.61.684f2f1d9f0934bcce91e77b69e17ec7</file_name> <open_name>libnvrtc-builtins.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMMCudaCompiler.so.aaed781fe4caa9d1099312d458a9b902</file_name> <open_name>libOpenMMCudaCompiler.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libfftw3f.so.3.4.4.a4580ddf9efebaad56fab49847a8c899</file_name> <open_name>libfftw3f.so.3</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMMPME.so.3208e45e71567824e8390ab1c79c6a66</file_name> <open_name>libOpenMMPME.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libnvrtc.so.8.0.61.ea3bff3d91151ddf671a0a1491635b57</file_name> <open_name>libnvrtc.so.8.0</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libOpenMMCPU.so.19849b4ff1cf4d33f75d9433b4d5c6bb</file_name> <open_name>libOpenMMCPU.so</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libcufft.so.8.0.61.889be25939bec6f9a2abec790772d28f</file_name> <open_name>libcufft.so.8.0</open_name> <copy_file/> </file_ref> <file_ref> <file_name>libstdc++.so.6.0.25.e344f48acfbd4f5abbf99b2c75cc5e50</file_name> <open_name>libstdc++.so.6</open_name> <copy_file/> </file_ref> <coproc> <type>NVIDIA</type> <count>1.000000</count> </coproc> <gpu_ram>512.000000</gpu_ram> <dont_throttle/> </app_version> | |
ID: 52005 | Rating: 0 | rate: / Reply Quote | |
Thanks. The context was libcudart.so.8.0 => not found Both files will be copied with the correct names into the slot directory, although they will be downloaded under a different (versioned) name. So a static test outside the running BOINC environment will fail to find them, but a dynamic test during running should be OK. I don't think this one will take us much further. | |
ID: 52006 | Rating: 0 | rate: / Reply Quote | |
If anybody is so inclined, can they try to run the boinc client manually with the --exit_after_finish flag, so the slot directory is preserved on failure? I just tried the manual run of the client with the suggested --exit_after_finish parameter but it did not preserve the slot contents. 05-Jun-2019 10:38:59 [GPUGRID] Starting task a3-TONI_TEST9-2-3-RND2847_2 05-Jun-2019 10:39:03 [GPUGRID] [sched_op] Deferring communication for 00:06:31 05-Jun-2019 10:39:03 [GPUGRID] [sched_op] Reason: Unrecoverable error for task a3-TONI_TEST9-2-3-RND2847_2 mv: cannot stat 'slots/8/output.coor': No such file or directory mv: cannot stat 'slots/8/output.vel': No such file or directory mv: cannot stat 'slots/8/output.idx': No such file or directory mv: cannot stat 'slots/8/output.dcd': No such file or directory mv: cannot stat 'slots/8/COLVAR': No such file or directory mv: cannot stat 'slots/8/log.file': No such file or directory mv: cannot stat 'slots/8/HILLS': No such file or directory mv: cannot stat 'slots/8/output.vel.dcd': No such file or directory mv: cannot stat 'slots/8/output.xtc': No such file or directory mv: cannot stat 'slots/8/output.xsc': No such file or directory mv: cannot stat 'slots/8/output.xstfile': No such file or directory 05-Jun-2019 10:39:03 [GPUGRID] Computation for task a3-TONI_TEST9-2-3-RND2847_2 finished 05-Jun-2019 10:39:03 [GPUGRID] Output file a3-TONI_TEST9-2-3-RND2847_2_9 for task a3-TONI_TEST9-2-3-RND2847_2 absent 05-Jun-2019 10:39:05 [GPUGRID] Started upload of a3-TONI_TEST9-2-3-RND2847_2_0 05-Jun-2019 10:39:07 [GPUGRID] Finished upload of a3-TONI_TEST9-2-3-RND2847_2_0 ^C05-Jun-2019 10:39:11 [---] Received signal 2 05-Jun-2019 10:39:11 [---] Exiting keith@Darksider:~/Desktop/BOINC$ | |
ID: 52007 | Rating: 0 | rate: / Reply Quote | |
I thought that all the tasks I had downloaded had failed but I see I have one host that has been successfully processing the acemd3 tasks. | |
ID: 52008 | Rating: 0 | rate: / Reply Quote | |
Anybody successfully run the new acemd3 app on a Turing card yet? I just realized that I still had a gpu_exclude for my Turing card on the host that had been successfully processing tasks. I somehow had skipped over removing the exclusion from that machine while I had done so on all the other hosts with Turing cards. | |
ID: 52009 | Rating: 0 | rate: / Reply Quote | |
I see that there is a new version 2.02, which I just tried on my GTX 1070 (Ubuntu 16.04.6). I just use the Ubuntu repository driver, which is 396.54 (proprietary), without any toolbox that I know of. GPUGRID 2.02 New version of ACEMD (cuda80) a67-TONI_TEST8-2-3-RND3156_0 00:00:03 (-) 0.00 100.000 - 6/10/2019 2:42:16 PM 0.985C + 1NV Computation error 0.00 MB i7-4790-G http://www.gpugrid.net/results.php?hostid=482386 Explain to me (simply) what I should check, and I will do it. | |
ID: 52010 | Rating: 0 | rate: / Reply Quote | |
Anybody successfully run the new acemd3 app on a Turing card yet? I just realized that I still had a gpu_exclude for my Turing card on the host that had been successfully processing tasks. I somehow had skipped over removing the exclusion from that machine while I had done so on all the other hosts with Turing cards. I think the plan is to get a stable acemd3 app running on legacy hardware and then release a beta for turing cards. @jim1348, I get the same error on one of my machines with dual GTX 1080 cards. | |
ID: 52011 | Rating: 0 | rate: / Reply Quote | |
I am also seeing failures due to the acemd binary not finding some libs: [root@node02 www.gpugrid.net]# ldd acemd.919-80.bin linux-vdso.so.1 (0x00007fff6a317000) libcuda.so.1 => /usr/lib/libcuda.so.1 (0x00007f740db2b000) libcudart.so.8.0 => not found libcufft.so.8.0 => not found libdl.so.2 => /usr/lib/libdl.so.2 (0x00007f740db26000) libpthread.so.0 => /usr/lib/libpthread.so.0 (0x00007f740db05000) libstdc++.so.6 => /usr/lib/libstdc++.so.6 (0x00007f740d975000) libm.so.6 => /usr/lib/libm.so.6 (0x00007f740d82d000) libgcc_s.so.1 => /usr/lib/libgcc_s.so.1 (0x00007f740d813000) libc.so.6 => /usr/lib/libc.so.6 (0x00007f740d64e000) librt.so.1 => /usr/lib/librt.so.1 (0x00007f740d644000) libnvidia-fatbinaryloader.so.430.14 => /usr/lib/libnvidia-fatbinaryloader.so.430.14 (0x00007f740d3f6000) /lib64/ld-linux-x86-64.so.2 => /usr/lib64/ld-linux-x86-64.so.2 (0x00007f740eca8000) This is despite the libs being right there in the directory with the binary: [root@node02 www.gpugrid.net]# ls -l libcu* -rwxr-xr-x 1 boinc boinc 394472 May 17 18:34 libcudart.so.8.0 -rwxr-xr-x 1 boinc boinc 426680 Jun 4 18:26 libcudart.so.8.0.61.af43be839e6366e731accc514633bd1f -rwxr-xr-x 1 boinc boinc 146745600 May 17 18:35 libcufft.so.8.0 -rwxr-xr-x 1 boinc boinc 146772424 Jun 4 18:28 libcufft.so.8.0.61.889be25939bec6f9a2abec790772d28f This machine is running Arch linux. Boinc was compiled locally, from the github source. The NVIDIA drivers are from Arch, with no modifications. [root@node02 www.gpugrid.net]# pacman -Ss nvidia | grep installed extra/nvidia 430.14-6 [installed] extra/nvidia-utils 430.14-1 [installed] extra/opencl-nvidia 430.14-1 [installed] This machine is currently successfully crunching GPGPU WUs for Primegrid and Einstein@Home, so its configuration is known good. | |
ID: 52012 | Rating: 0 | rate: / Reply Quote | |
Anybody successfully run the new acemd3 app on a Turing card yet? I just realized that I still had a gpu_exclude for my Turing card on the host that had been successfully processing tasks. I somehow had skipped over removing the exclusion from that machine while I had done so on all the other hosts with Turing cards. OK, that is a very different comprehension that I have for the wrapper app. I thought it was to allow use of the Turing cards. I guess I should put the gpu_exclude back in play for the hosts that failed the tasks. | |
ID: 52013 | Rating: 0 | rate: / Reply Quote | |
Anybody successfully run the new acemd3 app on a Turing card yet? I just realized that I still had a gpu_exclude for my Turing card on the host that had been successfully processing tasks. I somehow had skipped over removing the exclusion from that machine while I had done so on all the other hosts with Turing cards. See this post: http://www.gpugrid.net/forum_thread.php?id=4927&nowrap=true#51934 | |
ID: 52014 | Rating: 0 | rate: / Reply Quote | |
Thanks for the edification. | |
ID: 52015 | Rating: 0 | rate: / Reply Quote | |
I completed one while 5 others had errors. Same result on another PC but all tasks error on a 1080Ti https://www.gpugrid.net/show_host_detail.php?hostid=477247 nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2017 NVIDIA Corporation Built on Fri_Nov__3_21:07:56_CDT_2017 Cuda compilation tools, release 9.1, V9.1.85 | |
ID: 52016 | Rating: 0 | rate: / Reply Quote | |
I think I should take one of the hosts that fail the app and install the cuda toolkit and see if it changes anything. | |
ID: 52017 | Rating: 0 | rate: / Reply Quote | |
It won't hurt. One PC of mine with 1070/1070Ti works and another with 1080Ti doesn't. Both have the same nvcc -V results. | |
ID: 52018 | Rating: 0 | rate: / Reply Quote | |
Misc answers: | |
ID: 52019 | Rating: 0 | rate: / Reply Quote | |
It seems to be working. At least several previously-failing hosts switched to success. | |
ID: 52020 | Rating: 0 | rate: / Reply Quote | |
v2.03 works on both my computers! | |
ID: 52021 | Rating: 0 | rate: / Reply Quote | |
It was a cryptic bug in the order loading shared libraries, or something like that. Otherwise unexplainably system-dependent. | |
ID: 52022 | Rating: 0 | rate: / Reply Quote | |
My host that failed on 2.02 now works with 2.03. | |
ID: 52023 | Rating: 0 | rate: / Reply Quote | |
App 2.04 should support cuda10, if the scheduler collaborates. | |
ID: 52024 | Rating: 0 | rate: / Reply Quote | |
Well done Toni! | |
ID: 52025 | Rating: 0 | rate: / Reply Quote | |
Toni Tests looking good on my machines, even on the 2080Ti! | |
ID: 52026 | Rating: 0 | rate: / Reply Quote | |
Two hosts have received 2.04 and processed on the 2080 along with a 1070 successfully. | |
ID: 52027 | Rating: 0 | rate: / Reply Quote | |
App 2.04 should support cuda10, if the scheduler collaborates. Would it help the performance of a GTX 1070 (Ubuntu 16.04) to upgrade from CUDA 9.2 to CUDA 10 drivers? | |
ID: 52028 | Rating: 0 | rate: / Reply Quote | |
From what we've seen over at Seti with our special app, the CUDA9 application running the 410 series driver is fastest for Pascal cards. Of course if you have a Turing card, you are forced into using the 418 series drivers with CUDA10. | |
ID: 52029 | Rating: 0 | rate: / Reply Quote | |
Yes, it is a bit premature. I will await for the final app. Thanks for the input. | |
ID: 52030 | Rating: 0 | rate: / Reply Quote | |
Yes, it is a bit premature. I will await for the final app. Thanks for the input. Generally for crunching, a newer app is never necessary. The normal reason for updated drivers is to add ever more compatibility with newer games. And the drivers get ever more bloated and slower. Only if new architecture arrives requiring newer drivers that can drive them is it really necessary to update drivers. The low level math functions of the drivers have been pretty much static for years. Of course the silicon gets ever faster which is where the most improvement occurs. On one host with both a 2080 and 1070, the 2080 ran the test WU in 125 seconds versus 230 seconds with the 1070. Both on the 2.04 app. | |
ID: 52032 | Rating: 0 | rate: / Reply Quote | |
On one host with both a 2080 and 1070, the 2080 ran the test WU in 125 seconds versus 230 seconds with the 1070. Both on the 2.04 app. I see 93 seconds on my RTX 2080 Ti (v2.04) (The v2.03 had failed before on my host btw) | |
ID: 52034 | Rating: 0 | rate: / Reply Quote | |
I might be a little bit confused: Any time line when real production task of the 2.04 app will be supplied to Linux hosts? In the next days, weeks, months? I changed my BOINC settings to accommodate and give GPUGRID tasks the preference, however if there are no tasks available in the near future, I would like to change the settings again to better serve other projects. | |
ID: 52053 | Rating: 0 | rate: / Reply Quote | |
We'll be transitioning gradually. Was Toni's last response. I would not expect the project to be as responsive as you request. My suggestion is to move on to your other projects and occasionally check back in here for any news of non-beta apps and newly generated work units. | |
ID: 52054 | Rating: 0 | rate: / Reply Quote | |
To summarize: acemd3 cuda80 and cuda100 are very satisfactory for linux. Very few failures, fast, blazing on rtx 2080. The win app will need work though. | |
ID: 52062 | Rating: 0 | rate: / Reply Quote | |
Toni, did you discover a cutoff limit for driver version you should not go below or you will increase your errors? | |
ID: 52063 | Rating: 0 | rate: / Reply Quote | |
The most recent driver in the linux mint 19.1 repository is 390.116 (cuda 9). | |
ID: 52066 | Rating: 0 | rate: / Reply Quote | |
I was mainly asking if it was necessary to run the latest drivers like the 430 series. You have to run at least the 418 series to be compatible with Turing. | |
ID: 52067 | Rating: 0 | rate: / Reply Quote | |
The most recent driver in the linux mint 19.1 repository is 390.116 (cuda 9).My Linux Mint 19.1 computers have all been using CUDA 10 for a long time and were installed using the Driver Manager: aurum@Rig-27:~$ nvidia-smi Sat Jun 8 14:16:29 2019 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 430.14 Driver Version: 430.14 CUDA Version: 10.2 Side note, I know just enough Linux to be dangerous. Someone once told me to install the CUDA Toolkit to get CUDA 10 and it caused problems with Linux Mint 19. This is anecdotal but I think my computers generally work better without CUDA Toolkit installed. ____________ | |
ID: 52071 | Rating: 0 | rate: / Reply Quote | |
I agree, best to avoid the toolkit unless you need it to develop apps. They recently changed the way the toolkit handles the video drivers. You can get yourself in trouble with the toolkit installing its version of the graphics drivers alongside your runtime version of CUDA with the usual graphics drivers installation. | |
ID: 52072 | Rating: 0 | rate: / Reply Quote | |
To summarize: acemd3 cuda80 and cuda100 are very satisfactory for linux. Very few failures, fast, blazing on rtx 2080. The win app will need work though. When can we expect the linux version of acemd3 released to the masses? | |
ID: 52073 | Rating: 0 | rate: / Reply Quote | |
My guess not until Toni can figure out how to get the acemd3 Windows app working. | |
ID: 52074 | Rating: 0 | rate: / Reply Quote | |
I hope the CUDA 9.2 drivers are good enough. When I tried to update to the latest CUDA 10.x drivers for my GTX 1070 from the Nvidia repository (Ubuntu 16.04.6), it blanked out the desktop. BOINC still worked, but I had to reload the OS to get the desktop back. I suppose something in the X.Org server (if that is what they use) does not like the latest drivers, though the Linux experts will need to fix it. | |
ID: 52075 | Rating: 0 | rate: / Reply Quote | |
I hope the CUDA 9.2 drivers are good enough. When I tried to update to the latest CUDA 10.x drivers for my GTX 1070 from the Nvidia repository (Ubuntu 16.04.6), it blanked out the desktop. BOINC still worked, but I had to reload the OS to get the desktop back. I suppose something in the X.Org server (if that is what they use) does not like the latest drivers, though the Linux experts will need to fix it. I have nVidia proprietary drivers ver 430.14 with cuda ver 10.2 installed for three GTX-1060's working fine on my two FX-8350 Fedora 30 systems. Don't know where Ubuntu 16.x falls in the latest distro release line but F30 is the latest release (last month release date) and all is well. | |
ID: 52076 | Rating: 0 | rate: / Reply Quote | |
I still think using the graphics-drivers ppa is the easiest solution for obtaining and installing the drivers. As long as you have a debian distro newer than 14, you are good to go with the drivers right up to the current 430 series. | |
ID: 52077 | Rating: 0 | rate: / Reply Quote | |
I have a task that appears to be misbehaving. | |
ID: 52078 | Rating: 0 | rate: / Reply Quote | |
Try to suspend and restart it. Or kill it. | |
ID: 52079 | Rating: 0 | rate: / Reply Quote | |
Try to suspend and restart it. Or kill it. Suspending and restarting it set it back many hours, but only let it have many more hours to show the same problem. Aborting it worked, though. | |
ID: 52082 | Rating: 0 | rate: / Reply Quote | |
Any plans of sending out more beta tasks for Linux? I'm eager to try it on my 2070 and 1660 Ti. | |
ID: 52085 | Rating: 0 | rate: / Reply Quote | |
This thread has been very quiet since the last posting by Toni. I should hope that means he is quietly working hard on polishing off the new apps for both Linux and Windows. And soon that new work with new apps will become available. | |
ID: 52086 | Rating: 0 | rate: / Reply Quote | |
Yes. The problem is now the windows app and its dependencies. We also need to setup some hardware to test it. | |
ID: 52090 | Rating: 0 | rate: / Reply Quote | |
to make a bootable win10 disk i use wintousb software, with win10 iso image. | |
ID: 52091 | Rating: 0 | rate: / Reply Quote | |
Thanks for the tips! | |
ID: 52092 | Rating: 0 | rate: / Reply Quote | |
Erik was referring to Win 10 installation from USB. Supposedly Win 10 can also boot from a stick - however, this requires certified sticks. I don't have any further experience with that. | |
ID: 52097 | Rating: 0 | rate: / Reply Quote | |
Below is a link to a step by step guide with screenshots and links to the appropriate software to make the USB. | |
ID: 52100 | Rating: 0 | rate: / Reply Quote | |
I would consider purchasing a 250 GB SSD drive and setting up a Win/linux dual boot option using grub on one of your linux workstations. You'd still have a buy a license for Win10 pro. Building a "cheap" dedicated Win10 system would be even better. I would think Win10 on a USB stick would a be painful experience (slow). | |
ID: 52104 | Rating: 0 | rate: / Reply Quote | |
Erik was referring to Win 10 installation from USB. Supposedly Win 10 can also boot from a stick - however, this requires certified sticks. I don't have any further experience with that. nope, i was referring about how to make bootable win10 usb or portable disk. i was not talking about bootable usb to install win10, like you are thinking. my suggestion was: already installed win10 on bootable usb or ssd (or whatever) using wintousb or ultraiso | |
ID: 52105 | Rating: 0 | rate: / Reply Quote | |
I would think Win10 on a USB stick would a be painful experience (slow). Agreed. consider purchasing a 250 GB SSD drive and setting up a Win/linux dual boot option +1 Due to the Win10 licensing model, the USB stick would not be portable between host PCs. It would go to setup mode every time it is plugged into a different host PC, and another Win 10 license would be needed. | |
ID: 52106 | Rating: 0 | rate: / Reply Quote | |
it is portabke between all host machines. my experience with enterprise and education version wwin10 is portable between intel and amd platforms. only....when you switch between platforms, you need to activate win10. but be honest....who cares about activating when you are running it for test purposes and temporarlyI would think Win10 on a USB stick would a be painful experience (slow). | |
ID: 52111 | Rating: 0 | rate: / Reply Quote | |
Erik: thanks, I didn't even know this was possible. I would think Win10 on a USB stick would a be painful experience (slow). That's what the USB stick certification from MS is about: only allowing drives which are not painfully slow. MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 52112 | Rating: 0 | rate: / Reply Quote | |
it is portabke between all host machines. Ahh... Nice to know. It seems Microsoft don't insist on licensing each host individually, but rely on the PC being covered by Software Assurance. This is for Windows 8.1, I assume is the same for Win10. https://docs.microsoft.com/en-us/previous-versions/windows/it-pro/windows-8.1-and-8/jj592680(v=ws.11) See heading How is Windows To Go licensed? | |
ID: 52114 | Rating: 0 | rate: / Reply Quote | |
I am not sure if this was an anomaly or not but I did get a "usb windows 10" stick to boot and be activated. | |
ID: 52131 | Rating: 0 | rate: / Reply Quote | |
Any word on the status of acemd3? Last I saw there was a hold up due to a Windows (10?) issue. | |
ID: 52161 | Rating: 0 | rate: / Reply Quote | |
Haven't heard a peep either. Still waiting on Windows app I guess. | |
ID: 52162 | Rating: 0 | rate: / Reply Quote | |
Toni is trying out a new acemd3 application. 2.04 works as well as the previous 2.03 version. Still having issues getting the Windows app to work consistently. | |
ID: 52198 | Rating: 0 | rate: / Reply Quote | |
We are still dreaming of Linux app :з | |
ID: 52249 | Rating: 0 | rate: / Reply Quote | |
We are still dreaming of Linux app :зMy RTX 2080Ti's eagerly waiting for the acemd3 Linux app. My ranking in SETI@home has risen by 43,140 in the meantime. (51,148 -> 8,008) | |
ID: 52261 | Rating: 0 | rate: / Reply Quote | |
We have an application working to new application but forced to wait on win. Not fair when our old application expired. | |
ID: 52270 | Rating: 0 | rate: / Reply Quote | |
Hi, | |
ID: 52614 | Rating: 0 | rate: / Reply Quote | |
Hi, I would assume so. | |
ID: 52616 | Rating: 0 | rate: / Reply Quote | |
I setup for some acemd3 if any WU drop then I'll see what happens | |
ID: 52618 | Rating: 0 | rate: / Reply Quote | |
I setup for some acemd3 if any WU drop then I'll see what happens The wrapper app is able to handle CC capabilities of up to 7.5 for Turing cards. So your Titan V should be good to go Problem is getting some of the test tasks. I have been hammering the project since the announcement of the new acemd3 apps and I have yet to receive one. I hope that Toni releases some more work so I can get back to crunching for this project. I had no issues with the test versions back in July. Want some more work to get the latest apps. | |
ID: 52619 | Rating: 0 | rate: / Reply Quote | |
I'm really dissatisfied with the new acemd3 tasks for the following reasons:
| |
ID: 52817 | Rating: 0 | rate: / Reply Quote | |
i have build 4 rigs (with 24 gpu cards, gtx 1070, gtx 1080, gtx 1080 ti and rtx cards, all mixed together) special for gpugrid. but after summer vacation i have impression that this project is nearly dead. because after summer i dont get any WU to crunch (2 rigs with 12* gtx 1070 only) while before summer no problems at all. | |
ID: 52818 | Rating: 0 | rate: / Reply Quote | |
I'm really dissatisfied with the new acemd3 tasks for the following reasons:That's true under Windows Vista+ because of WDDM. Linux and Windows XP have significant performance benefit from SWAN_SYNC. This should be corrected. That's because this is beta work, and work from the short queue. | |
ID: 52821 | Rating: 0 | rate: / Reply Quote | |
This should be corrected. The wrapper app can handle that - see https://boinc.berkeley.edu/trac/wiki/WrapperApp#Thejobdescriptionfile: the final parameter in the job description file should be 2 (I think). If you're running a test task, you should be able to see and change that value. If changing it doesn't correct the problem, we'll have to dig deeper. | |
ID: 52823 | Rating: 0 | rate: / Reply Quote | |
i am leaving gpugrid, going to F@H (folding @ home) with my hardware. bye bye gpugrid You'll be back when you get tired of F@H's buggy software. ____________ | |
ID: 52824 | Rating: 0 | rate: / Reply Quote | |
That's true under Windows Vista+ because of WDDM. Linux and Windows XP have significant performance benefit from SWAN_SYNC. Well, when I played around with SWAN_SYNC about a year ago on Linux, there definitely was a small performance benefit, but it was so minimal that I decided it's not worth sacrificing half a CPU core for it for me. That's because this is beta work, and work from the short queue. Ah, so the tasks for acemd3 are just tasks that have been moved over from the "Short runs" application? Makes sense. Is there any particular reason why the "short runs" give much less credit than the "long runs" for the same runtime? For the "TEST" work units, I figured that they give less credit because they are just tests. Is it possible to opt-out of test work units? The wrapper app can handle that - see https://boinc.berkeley.edu/trac/wiki/WrapperApp#Thejobdescriptionfile: the final parameter in the job description file should be 2 (I think). This would solve the problem, but I'm pretty sure that BOINC verifies the files and re-downloads them if they have been altered... | |
ID: 52825 | Rating: 0 | rate: / Reply Quote | |
The wrapper app can handle that - see https://boinc.berkeley.edu/trac/wiki/WrapperApp#Thejobdescriptionfile: the final parameter in the job description file should be 2 (I think). You could still inspect the file and confirm whether GPUGrid have deployed it properly. A local edit wouldn't be permanent, for the reasons you state, but an edit to a task that has downloaded but not yet started might last for long enough for you to observe and report back. | |
ID: 52826 | Rating: 0 | rate: / Reply Quote | |
If the task isn't started yet, the files aren't in the "slots" directory yet. But suspending GPU, editing the file of the already started task and resuming GPU worked! On Linux, setting priority to 5 achieves the desired result: The acemd3 process always has the exact priority defined in cc_config.xml. Now the question is whether this would actually increase the priority unwantedly on Windows... | |
ID: 52827 | Rating: 0 | rate: / Reply Quote | |
What figure did you find there, before you edited it? | |
ID: 52828 | Rating: 0 | rate: / Reply Quote | |
It was unset. I added the "priority" xml tag myself. | |
ID: 52830 | Rating: 0 | rate: / Reply Quote | |
It was unset. I added the "priority" xml tag myself. OK, memo to project devs. I'd recommend value 2 for general use, rather than 5. | |
ID: 52831 | Rating: 0 | rate: / Reply Quote | |
Well, when I played around with SWAN_SYNC about a year ago on Linux, there definitely was a small performance benefit, but it was so minimal that I decided it's not worth sacrificing half a CPU core for it for me.It depends also on the GPU. High-end GPUs gain more (up to 30% on a GTX 1080Ti). To optimize the performance of the GPUGrid app, you should not over-commit the CPU feeding the GPU(s). That is one CPU task per CPU core. (on hyperthreaded CPUs you should reduce the number of CPU tasks down to 50%). You can achieve best GPUGrid performance (on high-end GPUs) if only one CPU task is running parallel the GPUGrid app (or none). Is there any particular reason why the "short runs" give much less credit than the "long runs" for the same runtime?Because they intended to be shorter than the "long run" workunits. The actual run time got mixed since then, some "long" workunits take about the same time to process as a "short" takes. For the "TEST" work units, I figured that they give less credit because they are just tests. Is it possible to opt-out of test work units?Yes. You could set your venues in your GPUGrid preferences not to receive "beta tasks" or you can deselect the entire ACEMD3 queue. But there are not so many beta and short tasks to even bother with this. | |
ID: 52840 | Rating: 0 | rate: / Reply Quote | |
I crunched a couple of the new tasks (New version of ACEMD v2.10 (cuda100))on my GTX1660Ti which is the new turing type and they processed and validated. Good that I can use this machine on GPUGrid now. | |
ID: 53074 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : New app update (acemd3)