11) Message boards : News : First part of server upgrade complete (Message 22698)
Posted 4524 days ago by Profile Bikermatt
I'm not getting short or long tasks right now.

Wed 14 Dec 2011 07:31:47 PM PST GPUGRID Message from server: No tasks sent
Wed 14 Dec 2011 07:31:47 PM PST GPUGRID Message from server: No tasks are available for ACEMD2: GPU molecular dynamics
Wed 14 Dec 2011 07:31:47 PM PST GPUGRID Message from server: No tasks are available for Long runs (8-12 hours on fastest card)

server says over 3K short are available....

ACEMD2: GPU molecular dynamics 3,511 1,408 0.81 (0.04 - 10.22) 743

Edit: right as I posted one machine got a long task, another still will not get long or short tasks..
12) Message boards : Server and website : Task will not upload (Message 21710)
Posted 4671 days ago by Profile Bikermatt
It worked. Thanks!
13) Message boards : Server and website : Task will not upload (Message 21705)
Posted 4672 days ago by Profile Bikermatt
This task has been trying to upload since the most recent down time. I have restarted my system and others are working fine.

Tue 19 Jul 2011 07:28:20 PM PDT GPUGRID [error] Error reported by file upload server: can't open file
Tue 19 Jul 2011 07:28:20 PM PDT GPUGRID Temporarily failed upload of s0r119-TONI_SH2MS3-6-100-RND9601_1_4: transient upload error


14) Message boards : Server and website : Too many workunits (Message 20945)
Posted 4769 days ago by Profile Bikermatt
Yes, I set this host's buffer to two days and the host picked up 5 tasks, for a total of 8 with only 2 GPUs. I aborted 3 of the tasks and set my buffer back to .01 days.


15) Message boards : Wish list : User reset for host error count (Message 20206)
Posted 4852 days ago by Profile Bikermatt
Right now I have a host that can not get tasks because it was having errors on a lot of tasks. I think one of the video cards may be bad or I may be having a driver issue.
Either way, I will not get tasks from GPU grid on this host for a while so it makes tracking down the problem very hard. I switched out a video card, but by the time the host starts getting tasks again I may not be around.

If I didn’t fix the issue and the errors continue it could be a really long time before I get tasks again making fixing the problem even harder. I’m glad GPU grid shuts down my hosts because it saves bandwidth and alerts me that there is a problem.

What would be nice though is if there were somewhere I could go to manually reset a host’s error count, it would allow troubleshooting and get hosts back online sooner once the user sees that there is a problem.

Is there any way this is possible?
16) Message boards : Graphics cards (GPUs) : Linux and GTX295 (Message 19728)
Posted 4902 days ago by Profile Bikermatt
From what I can see, all of your failures have been on device 1. You might what to take a closer look at that card. The GIANNI_DHFR tasks that are giving you errors are using a new algorithm and run a lot faster than than all of the other tasks on GPU grid.

At the same time I have noticed one of my recently overclocked 470s now errors on the GIANNI_DHFR tasks quite often were it never did before. The card still runs all of the other tasks just fine So I'm thinking the GIANNI_DHFR tasks are probably more demanding also.
17) Message boards : Graphics cards (GPUs) : ACEMD2 6.12 cuda and 6.13 cuda31 for windows and linux (Message 19455)
Posted 4920 days ago by Profile Bikermatt
The 6.13 app is running slower on my GTX 460 in Win7 compared to the 6.11 app.

The GTX 460 is running good in Linux on the 6.13 app, so far I am seeing around 21ms per step for the IBUCH tasks.

p2-IBUCH_15_PQpYEEIPI_101019-14-40-RND7762_2

# Time per step (avg over 1250000 steps): 33.258 ms
# Approximate elapsed time for entire WU: 41573.039 s

application version ACEMD2: GPU molecular dynamics v6.13 (cuda31)

p25-IBUCH_3_PQpYEEIPI_101019-14-40-RND3646_1

# Time per step (avg over 275000 steps): 31.505 ms
# Approximate elapsed time for entire WU: 39381.489 s

application version ACEMD2: GPU molecular dynamics v6.11 (cuda31)
18) Message boards : Graphics cards (GPUs) : ACEMD2 6.12 cuda and 6.13 cuda31 for windows and linux (Message 19440)
Posted 4922 days ago by Profile Bikermatt
GTX 470 driver 260.19.12

IBUCH_*_pYEEI tasks:

6.06 app ~ 11.2ms per step
6.13 app ~ 11.8ms per step

The one input_*-TONI task that I have ran was .7ms per step slower also.

On a positive note, I do have a GTX 460 running in Linux now, I can't imagine it could perform worse than it did with the 6.11 app in Win7.
19) Message boards : Graphics cards (GPUs) : ACEMD2 6.12 cuda and 6.13 cuda31 for windows and linux (Message 19431)
Posted 4922 days ago by Profile Bikermatt
In Linux app 6.12 driver 260.19.12 GT 240 at stock clocks:

KASHIF_HIVPR tasks ~42ms per step

IBUCH_*_pYEEI tasks: ~38ms per step

My first 6.13 app just finished in Linux on a GTX 470 driver 260.19.12 also.

It was an IBUCH and it ran about 800 sec longer than the 6.06 app, so not good for this one but we will see after a few different tasks have run.


20) Message boards : Graphics cards (GPUs) : ACEMD2 6.12 cuda and 6.13 cuda31 for windows and linux (Message 19426)
Posted 4922 days ago by Profile Bikermatt
6.12 task were running good on my GT240 now I have three in a row that have died.

<core_client_version>6.10.56</core_client_version>
<![CDATA[
<message>
process exited with code 127 (0x7f, -129)
</message>
<stderr_txt>

</stderr_txt>
]]>

This is the same error that my GTX 470s are getting now, and they were getting the same error on the beta work units.

Both systems are Ubuntu 10.4 with driver 260.19.12

The project has been reset on both systems, any suggestions?


Previous 10 | Next 10
//