Advanced search

Message boards : News : New app is out for testing

Author Message
Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist
Send message
Joined: 14 Mar 07
Posts: 1957
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 28498 - Posted: 13 Feb 2013 | 21:44:34 UTC

We have finished beta testing and we are now submitting workunits into a new queue for short runs.
If all works, we are going to update also the long queue.

Only cuda4.2 for the new app of course. Soon we will disable cuda3.1 as the application is way too old.

gdf

Richard Haselgrove
Send message
Joined: 11 Jul 09
Posts: 1576
Credit: 5,599,986,851
RAC: 8,794,097
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28499 - Posted: 13 Feb 2013 | 22:09:30 UTC

I've got one of these waiting to run, and I noticed it's up to replication _4 already:

http://www.gpugrid.net/workunit.php?wuid=4173049

3 of the previous runs ended with error -9

Anything special you'd like me to watch out for when it runs?

Dagorath
Send message
Joined: 16 Mar 11
Posts: 509
Credit: 179,005,236
RAC: 0
Level
Ile
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28503 - Posted: 13 Feb 2013 | 23:31:16 UTC - in response to Message 28499.

Put your safety glasses on and watch for smoke?
____________
BOINC <<--- credit whores, pedants, alien hunters

Serious Stuff
Send message
Joined: 20 Jan 10
Posts: 4
Credit: 2,569,014
RAC: 0
Level
Ala
Scientific publications
watwatwatwatwatwatwatwatwat
Message 28504 - Posted: 14 Feb 2013 | 1:26:27 UTC - in response to Message 28498.

Does this mean that those of us who have only been able to run the cuda 3.1 code are no longer wanted?

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28505 - Posted: 14 Feb 2013 | 6:57:06 UTC

Hm im suprised that cuda31 will finally disabled after switching it extra to short units queue. My 285gtx can normally do 6 wus per day :(
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist
Send message
Joined: 14 Mar 07
Posts: 1957
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 28506 - Posted: 14 Feb 2013 | 8:26:09 UTC - in response to Message 28505.

It will always be possible to run with 280s but on new drivers.
Simply the new application cannot be compiled with cuda3.1.

gdf

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28508 - Posted: 14 Feb 2013 | 9:27:24 UTC
Last modified: 14 Feb 2013 | 9:30:45 UTC

Possible but for the half performance of now, i dont invest >200w/h on 3 short wus per day ;) buuuut perhaps the new app runs better, so i will see and test some wus when 31 queue is empty. I will report then ;)

Ps: is it a typeerror to see now cuda32 on the site? Or is this cuda31 or something other?
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Richard Haselgrove
Send message
Joined: 11 Jul 09
Posts: 1576
Credit: 5,599,986,851
RAC: 8,794,097
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28509 - Posted: 14 Feb 2013 | 10:45:09 UTC - in response to Message 28503.

Put your safety glasses on and watch for smoke?

Well, I went to bed and pulled the duvet over my head, which amounts to much the same thing.

Results for host 43404

As you can see, the _4 task completed successfully, as did the subsequent _7 - that was the was last opportunity to get any science done, according to the "max # of error/total/success tasks 7, 10, 6" policy. And now I've got another _4.

That's a horribly high error rate - are you sure this app was ready for prime time?

While we're here, could we have some thoughts about the naming of the various application types, please? It's very misleading to have two separate (but identically-named) filters for short runs - especially when the the second one (appid=18) seems to be described as "CUDA 3.2" on the task selection preference page, but jobs from that queue were allocated as cuda42 to my host.

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28510 - Posted: 14 Feb 2013 | 11:08:43 UTC - in response to Message 28509.

Ps: is it a typeerror to see now cuda32 on the site? Or is this cuda31 or something other?
Yes it should be 3.1, but saying as it's being deprecated I wouldn't worry about it now.

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28511 - Posted: 14 Feb 2013 | 11:14:15 UTC - in response to Message 28510.
Last modified: 14 Feb 2013 | 11:14:48 UTC

Just watched a tasks complete and two subsequently fail after 2seconds.

trypsin_lig_375_run1-NOELIA_RL3_equ-0-1-RND1921_1 4141973 13 Feb 2013 | 9:40:31 UTC 13 Feb 2013 | 10:58:54 UTC Completed and validated 2,033.93 1,484.83 1,500.00 ACEMD beta version v6.48 (cuda42)

trypsin_lig_905_run3-NOELIA_RL3_equ-0-1-RND5342_2 4144209 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.07 0.06 --- ACEMD beta version v6.48 (cuda42)

trypsin_lig_905_run2-NOELIA_RL3_equ-0-1-RND6964_2 4144208 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)

Stderr output

<core_client_version>7.0.44</core_client_version>
<![CDATA[
<message>
- exit code 98 (0x62)
</message>
<stderr_txt>
ERROR: file mdioload.cpp line 207: Error reading parmtop file
called boinc_finish

</stderr_txt>
]]>

Both tasks that failed had already done so 2 times and have not been resent:

6459826 30790 14 Feb 2013 | 8:51:55 UTC 14 Feb 2013 | 9:18:53 UTC Error while computing 3.05 0.14 --- ACEMD beta version v6.48 (cuda42)
6503647 126506 14 Feb 2013 | 10:24:54 UTC 14 Feb 2013 | 10:30:32 UTC Error while computing 2.06 0.08 --- ACEMD beta version v6.48 (cuda42)
6503815 139265 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)
6503960 --- --- --- Unsent --- --- ---
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Bedrich Hajek
Send message
Joined: 28 Mar 09
Posts: 467
Credit: 8,187,371,966
RAC: 10,563,071
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28514 - Posted: 14 Feb 2013 | 12:16:55 UTC - in response to Message 28511.

Just watched a tasks complete and two subsequently fail after 2seconds.

trypsin_lig_375_run1-NOELIA_RL3_equ-0-1-RND1921_1 4141973 13 Feb 2013 | 9:40:31 UTC 13 Feb 2013 | 10:58:54 UTC Completed and validated 2,033.93 1,484.83 1,500.00 ACEMD beta version v6.48 (cuda42)

trypsin_lig_905_run3-NOELIA_RL3_equ-0-1-RND5342_2 4144209 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.07 0.06 --- ACEMD beta version v6.48 (cuda42)

trypsin_lig_905_run2-NOELIA_RL3_equ-0-1-RND6964_2 4144208 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)

Stderr output

<core_client_version>7.0.44</core_client_version>
<![CDATA[
<message>
- exit code 98 (0x62)
</message>
<stderr_txt>
ERROR: file mdioload.cpp line 207: Error reading parmtop file
called boinc_finish

</stderr_txt>
]]>

Both tasks that failed had already done so 2 times and have not been resent:

6459826 30790 14 Feb 2013 | 8:51:55 UTC 14 Feb 2013 | 9:18:53 UTC Error while computing 3.05 0.14 --- ACEMD beta version v6.48 (cuda42)
6503647 126506 14 Feb 2013 | 10:24:54 UTC 14 Feb 2013 | 10:30:32 UTC Error while computing 2.06 0.08 --- ACEMD beta version v6.48 (cuda42)
6503815 139265 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)
6503960 --- --- --- Unsent --- --- ---



I had a bunch of failures as well:

http://www.gpugrid.net/workunit.php?wuid=4144270

http://www.gpugrid.net/workunit.php?wuid=4144240

http://www.gpugrid.net/workunit.php?wuid=4144211

http://www.gpugrid.net/workunit.php?wuid=4144208

http://www.gpugrid.net/workunit.php?wuid=4144196

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28515 - Posted: 14 Feb 2013 | 14:10:06 UTC - in response to Message 28514.

These Betas are all failing on my systems, so I've had to suspend any more Beta testing for a while (otherwise I'll stop getting tasks):

trypsin_lig_941_run4-NOELIA_RL3_equ-0-1-RND4515_3 4144364 139265 14 Feb 2013 | 13:17:12 UTC 14 Feb 2013 | 13:19:09 UTC Error while computing 2.07 0.05 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_940_run3-NOELIA_RL3_equ-0-1-RND0852_1 4144359 139265 14 Feb 2013 | 13:17:12 UTC 14 Feb 2013 | 13:19:09 UTC Error while computing 2.07 0.05 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_941_run3-NOELIA_RL3_equ-0-1-RND2477_2 4144363 139859 14 Feb 2013 | 12:10:32 UTC 14 Feb 2013 | 12:16:50 UTC Error while computing 2.35 0.08 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_911_run2-NOELIA_RL3_equ-0-1-RND2760_2 4144232 139265 14 Feb 2013 | 11:45:48 UTC 14 Feb 2013 | 11:47:38 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_929_run2-NOELIA_RL3_equ-0-1-RND8942_1 4144310 139859 14 Feb 2013 | 12:22:28 UTC 14 Feb 2013 | 12:28:57 UTC Error while computing 2.26 0.08 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_933_run4-NOELIA_RL3_equ-0-1-RND6668_1 4144329 139859 14 Feb 2013 | 11:59:09 UTC 14 Feb 2013 | 12:04:48 UTC Error while computing 2.29 0.06 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_912_run3-NOELIA_RL3_equ-0-1-RND2352_2 4144238 139859 14 Feb 2013 | 12:16:50 UTC 14 Feb 2013 | 12:22:28 UTC Error while computing 2.24 0.08 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_900_run3-NOELIA_RL3_equ-0-1-RND4793_2 4144189 139265 14 Feb 2013 | 11:45:48 UTC 14 Feb 2013 | 11:47:38 UTC Error while computing 2.06 0.05 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_916_run3-NOELIA_RL3_equ-0-1-RND4035_2 4144255 139859 14 Feb 2013 | 11:46:58 UTC 14 Feb 2013 | 11:52:44 UTC Error while computing 2.21 0.09 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_900_run2-NOELIA_RL3_equ-0-1-RND3255_2 4144188 139859 14 Feb 2013 | 11:41:13 UTC 14 Feb 2013 | 11:46:58 UTC Error while computing 2.20 0.05 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_905_run3-NOELIA_RL3_equ-0-1-RND5342_2 4144209 139265 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.07 0.06 --- ACEMD beta version v6.48 (cuda42)
trypsin_lig_905_run2-NOELIA_RL3_equ-0-1-RND6964_2 4144208 139265 14 Feb 2013 | 11:03:01 UTC 14 Feb 2013 | 11:03:51 UTC Error while computing 2.11 0.05 --- ACEMD beta version v6.48 (cuda42)

I would suggest that anyone also seeing numerous Errors, stop running the Beta's for a while. Stick to the Long &/or Short tasks and after you complete a few try the odd Beta again.
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Richard Haselgrove
Send message
Joined: 11 Jul 09
Posts: 1576
Credit: 5,599,986,851
RAC: 8,794,097
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28516 - Posted: 14 Feb 2013 | 15:13:00 UTC - in response to Message 28515.

Tried a few as confirmation, with the same result - 12 errors in a row.

Beta tasks for host 132158

But it must be a data error - you can see the host has over 100 valid tasks, all done last weekend after the call went out to clear the queue so that proper application testing could resume.

At least these tasks weren't of the crashing/BSODing kind.

Profile Stoneageman
Avatar
Send message
Joined: 25 May 09
Posts: 224
Credit: 34,057,224,498
RAC: 231
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28519 - Posted: 14 Feb 2013 | 17:30:02 UTC

Thought I'd dip a toe back into the Beta testing pool, but I'm getting 'No beta tasks available'. Is it windows only?

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28520 - Posted: 14 Feb 2013 | 17:42:30 UTC - in response to Message 28519.

I think it is Windows only.
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Toni
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist
Send message
Joined: 9 Dec 08
Posts: 1006
Credit: 5,068,599
RAC: 0
Level
Ser
Scientific publications
watwatwatwat
Message 28522 - Posted: 14 Feb 2013 | 20:08:24 UTC - in response to Message 28520.
Last modified: 14 Feb 2013 | 20:09:31 UTC

Hi, a subset of the betas had indeed a problem that makes them fail immediately. We devised a way to selectively remove single unsent tasks and cancelled them, so many should have disappeared from the queue; those already downloaded will disappear gradually.

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28524 - Posted: 15 Feb 2013 | 1:11:08 UTC - in response to Message 28522.

By 'disappear gradually' I presume you mean they will fail, get resent, fail, get resent, fail and then be cancelled. But for the stubborn scheduler, the 2sec runtime wouldn't be such an issue.

Anyway, I've been running a few again and they are not failing. However the other issues persist. Of note is the dependence on high CPU Kernel time. At 85% CPU usage I was seeing 10% GPU usage, and on another system with only 50% CPU usage (but high Kernel usage) I only saw 2% GPU utilization. Another app was hogging the Kernel and memory, and GPU Utilization went up to 50% when I suspended it.
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28526 - Posted: 15 Feb 2013 | 12:16:40 UTC - in response to Message 28524.

trypsin_lig_901_run1-NOELIA_RL3_equ-0-1-RND1273_7

errors Too many errors (may have bug)

All the same 2" errors,
http://www.gpugrid.net/workunit.php?wuid=4144191
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28541 - Posted: 16 Feb 2013 | 10:32:30 UTC

Since today there is only 1 user left who connected the last 24hours to the short cuda31 queue (Serverstats). Im proud to tell, im this lonely guy ;) So i need at least 3 more (@24h crunching) days to clear this queue up (~4 hours per WU). Only as little estimate when the adminstaff can deactived it, and the problems with the queue selection on some computers should go away then ;)
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28543 - Posted: 16 Feb 2013 | 18:12:56 UTC
Last modified: 16 Feb 2013 | 18:21:17 UTC

hmm ok gpugrid dont sends me anymore tasks from cuda31 queue. strange. who should compute them now? O.o
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28546 - Posted: 16 Feb 2013 | 19:56:38 UTC

Sry for spaming around, only want to help finishing the queue :(

I only want to tell i getting tasks again. Seems i hit any limit before..
____________
DSKAG Austria Research Team: http://www.research.dskag.at



TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28547 - Posted: 16 Feb 2013 | 22:41:46 UTC - in response to Message 28541.

Since today there is only 1 user left who connected the last 24hours to the short cuda31 queue (Serverstats). Im proud to tell, im this lonely guy ;) So i need at least 3 more (@24h crunching) days to clear this queue up (~4 hours per WU). Only as little estimate when the adminstaff can deactived it, and the problems with the queue selection on some computers should go away then ;)


Sorry to spoil your party but I am getting only cuda 3.1 short runs on one of my system while this pc will accept all sorts jobs.
I think the server page is not corect with this, as yesterday was to see as well, its always 2 left, however my pc did already 4 today and 1 will be ready in the next hours (could be just Saturday in UTC).
____________
Greetings from TJ

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28548 - Posted: 16 Feb 2013 | 23:18:35 UTC

It counts down on the " in work" unit when i finished something when i remeber right. Perhaps it shows 2 units free to send but all "in work" units are ready to resend and it fills the ready to send queue only until 2 or something. But im happy when someone help me, so we can finish it up faster :) gogogooo
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile AdamYusko
Send message
Joined: 29 Jun 12
Posts: 26
Credit: 21,540,800
RAC: 0
Level
Pro
Scientific publications
watwatwatwatwatwat
Message 28550 - Posted: 17 Feb 2013 | 3:34:01 UTC

I am a bit confused and I hope it gets resolved soon, I have my lowest power GPU set to only crunch Short Run tasks, but somehow it keeps on being given the Long Run Tasks, even though it has no chance of finishing them even within the 48 hour Bonus time. I hope the settings just didn't set in once I realized Short run 4.2 were available.
____________

werdwerdus
Send message
Joined: 15 Apr 10
Posts: 123
Credit: 1,004,473,861
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28551 - Posted: 17 Feb 2013 | 6:14:27 UTC - in response to Message 28550.

I am also getting only long runs on my computers that are set to accept only short runs. I have them set to receive long runs only if there are no short runs available, but there seems to be plenty available.
____________
XtremeSystems.org - #1 Team in GPUGrid

Dagorath
Send message
Joined: 16 Mar 11
Posts: 509
Credit: 179,005,236
RAC: 0
Level
Ile
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28552 - Posted: 17 Feb 2013 | 9:27:32 UTC - in response to Message 28550.
Last modified: 17 Feb 2013 | 9:28:12 UTC

I am a bit confused and I hope it gets resolved soon, I have my lowest power GPU set to only crunch Short Run tasks, but somehow it keeps on being given the Long Run Tasks, even though it has no chance of finishing them even within the 48 hour Bonus time. I hope the settings just didn't set in once I realized Short run 4.2 were available.


There are three things you must do to prevent getting the long runs. You have to double-check, read carefully and make sure you have done all of these things.

1) uncheck long runs

2) uncheck the box that says "If no work for selected applications is available, accept work from other applications?"

3) click the update button

I am willing to bet you did not do step number 2) or possibly not 3) and I am willing to bet you will post at least 2 more times that it does not work properly before you RTFP and realize you neglected to do 2). The settings work properly for everybody else therefore if they do not work for you then you are not setting them properly.
____________
BOINC <<--- credit whores, pedants, alien hunters

Dagorath
Send message
Joined: 16 Mar 11
Posts: 509
Credit: 179,005,236
RAC: 0
Level
Ile
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28553 - Posted: 17 Feb 2013 | 9:37:17 UTC - in response to Message 28551.

I am also getting only long runs on my computers that are set to accept only short runs. I have them set to receive long runs only if there are no short runs available, but there seems to be plenty available.


What makes you think there are plenty short runs available, the server status page? That page lies. Never trust the numbers you see there, never, not at this project or any other project. If you're getting long runs then it's because when your computer requests work there are no short runs. Notice that leaves open the possibility that sometime AFTER your host receives the long runs someone/something puts tasks in the short run queue.


____________
BOINC <<--- credit whores, pedants, alien hunters

ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28554 - Posted: 17 Feb 2013 | 11:30:20 UTC
Last modified: 17 Feb 2013 | 11:32:09 UTC

The server status page doesn't lie, but since it's usually updated every 10 minutes it never shows the current status either (i.e. not giving you the information you're looking for).

If you set up your machines as Dagorath suggested you should have a backup project setup, as you'll likely not get enough work from GPU-Grid to keep your GPU busy.

BTW: GPU-Grid generates WUs based on the results of a previous WU. That's why short return times are important. And that's probably why the amount of "WUs ready to send" always increases by 1 as soon as someone returns a result.

MrS
____________
Scanning for our furry friends since Jan 2002

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28557 - Posted: 17 Feb 2013 | 12:08:13 UTC - in response to Message 28554.

It's normally reasonably reliable (unlike some projects), but it might be telling a few porkies at the minute; some tasks were pulled out manually. Perhaps the server isn't aware of this?
http://www.gpugrid.net/forum_thread.php?id=3285&nowrap=true#28522

I tried to get short tasks at one stage (yesterday or the day before) when it said there were hundreds of tasks, but I was unable to for hours.
http://www.gpugrid.net/forum_thread.php?id=3284
Don't know what the situation is right now though and I don't think there is any point second-guessing what's in the queue - lots of failures, beta testing different task types, 3.1 due to be deprecated... It's likely to keep changing.

Certainly with so many quick failures, task return and re-issue could be rapid.

Profile ritterm
Avatar
Send message
Joined: 31 Jul 09
Posts: 88
Credit: 244,413,897
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28559 - Posted: 17 Feb 2013 | 13:07:37 UTC - in response to Message 28557.

I tried to get short tasks at one stage (yesterday or the day before) when it said there were hundreds of tasks, but I was unable to for hours...

What is the difference between the two "Short runs (2-3 hours on fastest card)" listed on the server status page?
____________

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28560 - Posted: 17 Feb 2013 | 13:13:08 UTC
Last modified: 17 Feb 2013 | 13:14:49 UTC

The first one is cuda31 support, "specially" for 2xx cards with drivers <3xx.xx. It will deactivaded soon. The second is cuda42 support for 4xx or better cards with driver >=3xx.xx

You can compute with eatch card in every queue, but i recommend so for the most performanceoutput.
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile ritterm
Avatar
Send message
Joined: 31 Jul 09
Posts: 88
Credit: 244,413,897
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28561 - Posted: 17 Feb 2013 | 13:46:40 UTC - in response to Message 28560.
Last modified: 17 Feb 2013 | 13:46:55 UTC

The second is cuda42 support for 4xx or better cards with driver >=3xx.xx

Thanks for the feedback, dskag. That's what I thought. :-)

The problem for me is that I've got two hosts set to request tasks for that app only (in preferences, run test apps? No; only short runs CUDA42 checked; accept other work? No), yet I'm not getting any of those tasks. :-( One host has a 550Ti w/306.97 and the other is a 260 w/310.90.
____________

Profile microchip
Avatar
Send message
Joined: 4 Sep 11
Posts: 110
Credit: 326,102,587
RAC: 0
Level
Asp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28562 - Posted: 17 Feb 2013 | 14:26:43 UTC - in response to Message 28561.

The second is cuda42 support for 4xx or better cards with driver >=3xx.xx

Thanks for the feedback, dskag. That's what I thought. :-)

The problem for me is that I've got two hosts set to request tasks for that app only (in preferences, run test apps? No; only short runs CUDA42 checked; accept other work? No), yet I'm not getting any of those tasks. :-( One host has a 550Ti w/306.97 and the other is a 260 w/310.90.


Same here.. I have both short queues checked yet don't get any tasks. Beta and long queues are disabled
____________

Team Belgium

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28563 - Posted: 17 Feb 2013 | 14:26:58 UTC - in response to Message 28561.
Last modified: 17 Feb 2013 | 15:32:20 UTC

That's a topic already being discussed in the No tasks available for ACEMD short runs cuda 3.2 thread.
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Profile ritterm
Avatar
Send message
Joined: 31 Jul 09
Posts: 88
Credit: 244,413,897
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28566 - Posted: 17 Feb 2013 | 16:05:21 UTC - in response to Message 28563.

That's a topic already being discussed in the No tasks available for ACEMD short runs cuda 3.2 thread.

Thanks for the feedback... Sorry to have missed that, but there's been so much talk lately about similar things in different threads that it's hard for me to keep track! I will set my prefs to accept both short queues and see what happens.
____________

Dagorath
Send message
Joined: 16 Mar 11
Posts: 509
Credit: 179,005,236
RAC: 0
Level
Ile
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28567 - Posted: 17 Feb 2013 | 17:11:22 UTC - in response to Message 28566.
Last modified: 17 Feb 2013 | 17:12:30 UTC

That's a topic already being discussed in the No tasks available for ACEMD short runs cuda 3.2 thread.

Thanks for the feedback... Sorry to have missed that, but there's been so much talk lately about similar things in different threads that it's hard for me to keep track! I will set my prefs to accept both short queues and see what happens.


From reports in this very thread it is evident that sometimes neither of the two short queues has tasks. If your host requests a task when both queues are empty then it won't get one. And remember, even if the server says there are tasks in one of the queues it isn't necessarily so.
____________
BOINC <<--- credit whores, pedants, alien hunters

Profile ritterm
Avatar
Send message
Joined: 31 Jul 09
Posts: 88
Credit: 244,413,897
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28572 - Posted: 17 Feb 2013 | 19:57:44 UTC - in response to Message 28566.

...I will set my prefs to accept both short queues and see what happens.

And I got one...a short run CUDA42, that is. :-)

____________

Profile AdamYusko
Send message
Joined: 29 Jun 12
Posts: 26
Credit: 21,540,800
RAC: 0
Level
Pro
Scientific publications
watwatwatwatwatwat
Message 28583 - Posted: 17 Feb 2013 | 22:32:03 UTC

Dagorath,

I understand the settings, and I do have them set how I want them. I was just making a remark as around the time I got those tasks I checked the server status and it appeared that there were plenty of short run tasks ( but yes I realize it is not a perfect exactly up to date page, just a ball park). I do not mind crunching long runs on the card, so I do leave the box checked that says it can grab basically any GPU grid task it finds if my initial request does not come up with a result.
____________

Dagorath
Send message
Joined: 16 Mar 11
Posts: 509
Credit: 179,005,236
RAC: 0
Level
Ile
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 28586 - Posted: 18 Feb 2013 | 3:02:52 UTC - in response to Message 28583.

Adam,

It's obvious I misunderstood your other message and I'm glad you've got it worked out now. Happy crunching!

____________
BOINC <<--- credit whores, pedants, alien hunters

Post to thread

Message boards : News : New app is out for testing

//