How long are the Work Units going to be?


Advanced search

Message boards : Number crunching : How long are the Work Units going to be?

Sort
Author Message
Profile Conan
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 219
ID: 100
Credit: 4,256,493
RAC: 0
Message 194 - Posted 14 Sep 2006 23:28:09 UTC

My first batch of 100 WU's have all been given a Boinc time of 8.23 minutes and the 17 I have done all took about this time (Successful but with an error).
Another 12 have downloaded with a mixture of 8 minutes and 59 hour Work Units.
That is a huge difference in time, even if Boinc has trouble allocating times can it be this much of a difference?
What will the average time become when things settle down?

Profile Andre Kerstens
Forum moderator
Project tester
Volunteer tester
Avatar

Joined: Sep 11 06
Posts: 749
ID: 1
Credit: 15,199
RAC: 0
Message 195 - Posted 15 Sep 2006 0:22:04 UTC - in response to Message ID 194 .

I've tested the new WUs this afternoon and they estimate and run as 1.5 hours on a P4 3.2 GHz. We will deploy this evening.

AK

My first batch of 100 WU's have all been given a Boinc time of 8.23 minutes and the 17 I have done all took about this time (Successful but with an error).
Another 12 have downloaded with a mixture of 8 minutes and 59 hour Work Units.
That is a huge difference in time, even if Boinc has trouble allocating times can it be this much of a difference?
What will the average time become when things settle down?

Angus
Volunteer tester

Joined: Sep 13 06
Posts: 17
ID: 32
Credit: 15,111
RAC: 0
Message 226 - Posted 15 Sep 2006 15:23:09 UTC - in response to Message ID 195 .

I've tested the new WUs this afternoon and they estimate and run as 1.5 hours on a P4 3.2 GHz. We will deploy this evening.

AK

My first batch of 100 WU's have all been given a Boinc time of 8.23 minutes and the 17 I have done all took about this time (Successful but with an error).
Another 12 have downloaded with a mixture of 8 minutes and 59 hour Work Units.
That is a huge difference in time, even if Boinc has trouble allocating times can it be this much of a difference?
What will the average time become when things settle down?



I think your estimates are way too short. I just completed 2 of the new 5.02 WUs and they both ran over 4 hours on an XP2600.
LoneStar
Volunteer tester

Joined: Sep 13 06
Posts: 4
ID: 27
Credit: 13,448
RAC: 0
Message 237 - Posted 15 Sep 2006 21:19:56 UTC
Last modified: 15 Sep 2006 21:28:39 UTC

I got about 8 hours on a P4/1.6GHz, which should have gotten around 3 hours, eh?

Unit in progress is at 44% for 03:23... 1tng_mod001_33_299565_2, if that helps.
BOINC.exe time is back down to normal though, not high like it was with the first workunits.
-D

EDIT: Well, it *is* a P4 Mobile, but that hasn't ever seemed to make a difference, I don't think even the cache is any smaller than a normal P4, and it seems to benchmark in line with standard chips in other tests, soo...

EDIT again: Cool, I just restarted BOINC and now the estimated times for my remaining workunits has recalculated up to 08:13. Never realized it would do that.

Profile Krunchin-Keith [USA]
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 41
ID: 4
Credit: 1,539,093
RAC: 0
Message 239 - Posted 15 Sep 2006 22:01:16 UTC - in response to Message ID 195 .

AK wrote:
I've tested the new WUs this afternoon and they estimate and run as 1.5 hours on a P4 3.2 GHz. We will deploy this evening.


Is yours a single CPU ?

I started on 5.02 earlier today

No downlaod problems.

On Windoze at 3.06GHz HT running 1 CPU charmm 5.02 took 4 hrs 9 mins. That was down from 5 hrs 30 mins for charmm 5.01 on this host.

It just occured to me, that actually that is only using 50% of my CPU. So if I run both CPU's I can crank out 2 in about the same time, so effective wu time is about 2 hours. Higher but not as far off your estimate as being more than double.

I just cranked my 3.06GHz HT and 3.8GHz HT up to 2 CPUs to see the results. I had been not running boinc on them up until Sept 1st, put out too much heat into my room which requires extra A/C for me to be comfortable, they can take it but I can't. I was not going to run both CPUs until Oct 1st, when it cools off more, but I'll test it for the week-end.

4 wu download now, no problems.
Profile Andre Kerstens
Forum moderator
Project tester
Volunteer tester
Avatar

Joined: Sep 11 06
Posts: 749
ID: 1
Credit: 15,199
RAC: 0
Message 240 - Posted 15 Sep 2006 23:47:57 UTC - in response to Message ID 239 .

Hi Keith,

Yes, both of the machine I got the estimate from are single cpu. And even if they were dual cpus, boinc would schedule only job on one cpu (2 running at a time).

Would other users mind posting some of their running times? (windows, Linux and mac)

Andre

AK wrote:
I've tested the new WUs this afternoon and they estimate and run as 1.5 hours on a P4 3.2 GHz. We will deploy this evening.


Is yours a single CPU ?

I started on 5.02 earlier today

No downlaod problems.

On Windoze at 3.06GHz HT running 1 CPU charmm 5.02 took 4 hrs 9 mins. That was down from 5 hrs 30 mins for charmm 5.01 on this host.

It just occured to me, that actually that is only using 50% of my CPU. So if I run both CPU's I can crank out 2 in about the same time, so effective wu time is about 2 hours. Higher but not as far off your estimate as being more than double.

I just cranked my 3.06GHz HT and 3.8GHz HT up to 2 CPUs to see the results. I had been not running boinc on them up until Sept 1st, put out too much heat into my room which requires extra A/C for me to be comfortable, they can take it but I can't. I was not going to run both CPUs until Oct 1st, when it cools off more, but I'll test it for the week-end.

4 wu download now, no problems.

gamer007
Volunteer tester

Joined: Sep 13 06
Posts: 8
ID: 61
Credit: 13,988
RAC: 0
Message 242 - Posted 16 Sep 2006 1:04:18 UTC - in response to Message ID 240 .

Hi Keith,

Yes, both of the machine I got the estimate from are single cpu. And even if they were dual cpus, boinc would schedule only job on one cpu (2 running at a time).

Would other users mind posting some of their running times? (windows, Linux and mac)

Andre

AK wrote:
I've tested the new WUs this afternoon and they estimate and run as 1.5 hours on a P4 3.2 GHz. We will deploy this evening.


Is yours a single CPU ?

I started on 5.02 earlier today

No downlaod problems.

On Windoze at 3.06GHz HT running 1 CPU charmm 5.02 took 4 hrs 9 mins. That was down from 5 hrs 30 mins for charmm 5.01 on this host.

It just occured to me, that actually that is only using 50% of my CPU. So if I run both CPU's I can crank out 2 in about the same time, so effective wu time is about 2 hours. Higher but not as far off your estimate as being more than double.

I just cranked my 3.06GHz HT and 3.8GHz HT up to 2 CPUs to see the results. I had been not running boinc on them up until Sept 1st, put out too much heat into my room which requires extra A/C for me to be comfortable, they can take it but I can't. I was not going to run both CPUs until Oct 1st, when it cools off more, but I'll test it for the week-end.

4 wu download now, no problems.



My 3 results: http://docking.utep.edu/result.php?resultid=8591
http://docking.utep.edu/result.php?resultid=8594
http://docking.utep.edu/result.php?resultid=8910

For my 3.0GHz P4, looks like it averages between 17,000 - 18,000 seconds.
Honza
Volunteer tester

Joined: Sep 13 06
Posts: 25
ID: 72
Credit: 5,064
RAC: 0
Message 251 - Posted 16 Sep 2006 6:08:50 UTC

For dual-cored machine when running both cores, we can usually see 85-95% of speed on each. There is no need (nor possibility) of changing computation estimation on dual/multi cores or CPUs system.

Already provided ~40 hours of overclocked Conroe machine, I have duration_correction_factor of >4
It should be clear to anybody experienced on BOINC that running times are defaulty overestimated by factor of ~4

(Initially, my estimation was ~28 which showed almost 3 hours. Later results are ~ 2 hours of computation time).

Hmmm, got 10 results in pending credit/validation...

Profile Krunchin-Keith [USA]
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 41
ID: 4
Credit: 1,539,093
RAC: 0
Message 264 - Posted 16 Sep 2006 13:28:08 UTC - in response to Message ID 240 .

Andre wrote:
Yes, both of the machine I got the estimate from are single cpu. And even if they were dual cpus, boinc would schedule only job on one cpu (2 running at a time).


Yes, that is true

But what I meant is I can run 2 results at the same time, both completing in the same 4.5 hours approx. I turn out 2 in that time. while a single CPU turns out 1 at a time. If a single CPU took 2.25 hours each it too would turn out 2 in the same time, just they were processed 1 after the other and not simultaneously. So effectively a HT CPU taking 4.5 hours for two results has an equivilemnt of 2.25 single CPU hours per result. If only running 1 thread then a HT CPU is really only running at 50% of its capacity, not 50% speed.

Anyway last nights sucessful results for charmm 5.02.

Windoze on 3.8GHz HT completed 4 at average 4hr 26mins each (4:25 to 4:29) fairly even
Windoze on 3.06GHz HT completed 4 at average 4hr 37min each (4:09 to 5:06)
Honza
Volunteer tester

Joined: Sep 13 06
Posts: 25
ID: 72
Credit: 5,064
RAC: 0
Message 275 - Posted 16 Sep 2006 19:44:23 UTC - in response to Message ID 264 .

If only running 1 thread then a HT CPU is really only running at 50% of its capacity, not 50% speed.
I can't agree there.
HT is not a dual-core, nor dual-CPU...actually far from it.
If you are running without HT (or running only single thread/application), you are at worst loosing 25% of total performance.
Older Xeon are known to do poor on HT and there are cases when running both threads on HT actually makes whole system running slower and overall performance goes down (due to high latencies to CPU cache and memory, all running via FSB etc.); for example on highly optimalized code like Einstein akosf's optimalization was.

You may find some interesting vaules in my BOINC table .
(I have been running Intel HT machine but abandomed them 1+ years ago when I went dual-cores and again upgraed recently to Conroes).
Profile Rebirther
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 63
ID: 52
Credit: 69,033
RAC: 0
Message 277 - Posted 16 Sep 2006 20:17:20 UTC - in response to Message ID 275 .

If only running 1 thread then a HT CPU is really only running at 50% of its capacity, not 50% speed.
I can't agree there.
HT is not a dual-core, nor dual-CPU...actually far from it.
If you are running without HT (or running only single thread/application), you are at worst loosing 25% of total performance.
Older Xeon are known to do poor on HT and there are cases when running both threads on HT actually makes whole system running slower and overall performance goes down (due to high latencies to CPU cache and memory, all running via FSB etc.); for example on highly optimalized code like Einstein akosf's optimalization was.

You may find some interesting vaules in my BOINC table .
(I have been running Intel HT machine but abandomed them 1+ years ago when I went dual-cores and again upgraed recently to Conroes).


Haha Honza, I also have a table running ;) Can you send me the conroe at home :p Next time I will have one. PS: Good table so far, better overview than mine :)
Profile Frank Boerner
Volunteer tester

Joined: Sep 13 06
Posts: 18
ID: 101
Credit: 744,548
RAC: 0
Message 295 - Posted 17 Sep 2006 8:29:49 UTC

I have dopenloaded 116 WUs on my Linux machine (AMD 3000+). The cruniching time is about 4 minutes. All WUs stay on pending.

Angus
Volunteer tester

Joined: Sep 13 06
Posts: 17
ID: 32
Credit: 15,111
RAC: 0
Message 298 - Posted 17 Sep 2006 15:02:42 UTC - in response to Message ID 295 .
Last modified: 17 Sep 2006 15:03:16 UTC

I have dopenloaded 116 WUs on my Linux machine (AMD 3000+). The cruniching time is about 4 minutes. All WUs stay on pending.


Your results are not "finishing" - they are failing.

<core_client_version>5.5.0</core_client_version>
<stderr_txt>
Calling BOINC init.
Starting charmm run...
ERROR - Charmm exited with code 1.
Calling BOINC finish.

</stderr_txt>
Profile Conan
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 219
ID: 100
Credit: 4,256,493
RAC: 0
Message 311 - Posted 17 Sep 2006 22:08:35 UTC - in response to Message ID 298 .

I have dopenloaded 116 WUs on my Linux machine (AMD 3000+). The cruniching time is about 4 minutes. All WUs stay on pending.


Your results are not "finishing" - they are failing.

<core_client_version>5.5.0</core_client_version>
<stderr_txt>
Calling BOINC init.
Starting charmm run...
ERROR - Charmm exited with code 1.
Calling BOINC finish.

</stderr_txt>


I am also getting the same error after doing 50 odd WU's, all done in less than 300 seconds. I believe I may still doing the 5.01 WU's. I dumped over 100 of these when Andre asked us to but I then downloaded 5.01 WU's again.
At this rate I will get rid of then fairly quickly and can start on the 5.02 WU's.
I am running Linux on a AMD Opteron with dual 848 cpus.
Profile Andre Kerstens
Forum moderator
Project tester
Volunteer tester
Avatar

Joined: Sep 11 06
Posts: 749
ID: 1
Credit: 15,199
RAC: 0
Message 314 - Posted 17 Sep 2006 23:24:56 UTC - in response to Message ID 311 .

OK;maybe this is not completely clear yet. The windows fix was an app fix and therefore the windows app version went to 5.2. The linux and mac version of the app are still at 5.01. because the error -131 fix was only a fix to the input file of the app (no version change). We are currently working on the 'charmm exited with error 1' problem, which turns out to be much harder, because not all linux machines have this problem, only some.

Thanks!
Andre

I have dopenloaded 116 WUs on my Linux machine (AMD 3000+). The cruniching time is about 4 minutes. All WUs stay on pending.


Your results are not "finishing" - they are failing.

<core_client_version>5.5.0</core_client_version>
<stderr_txt>
Calling BOINC init.
Starting charmm run...
ERROR - Charmm exited with code 1.
Calling BOINC finish.

</stderr_txt>


I am also getting the same error after doing 50 odd WU's, all done in less than 300 seconds. I believe I may still doing the 5.01 WU's. I dumped over 100 of these when Andre asked us to but I then downloaded 5.01 WU's again.
At this rate I will get rid of then fairly quickly and can start on the 5.02 WU's.
I am running Linux on a AMD Opteron with dual 848 cpus.

Profile clownius
Volunteer tester
Avatar

Joined: Nov 14 06
Posts: 61
ID: 280
Credit: 2,677
RAC: 0
Message 1396 - Posted 14 Nov 2006 10:40:16 UTC
Last modified: 14 Nov 2006 10:48:20 UTC

Just out of interest on what i should attach to this project.

I just threw a 2Ghz Celery system at it as a test and was wondering if i should even bother on my 700Mhz system?

Profile Conan
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 219
ID: 100
Credit: 4,256,493
RAC: 0
Message 1399 - Posted 14 Nov 2006 12:10:18 UTC

> @ clownius,

I have been running Docking on an AMD Opteron 275 and now a Opteron 285 using Linux. Over the last 2 days I have added an AMD 4800+ and an Intel P4 (Windows).
The Opteron 285 runs at 2.6 GHz and takes about 2 hours 30 minutes (Linux).
The Opteron 275 runs at 2.2 GHz and takes about 2 hours 51 minutes (Linux).
The 4800+ runs at 2.4 GHz and takes about 5 hours 50 minutes (Windows XP).
The P4 is a 2.53 GHz @ 2.75 GHz and takes about 8 hours 51 minutes (Windows XP).

This may give you a guide.
Scanning finished jobs found the following, all run Windows
A 2.0 GHz Pentium M processor doing it in around 6 hours 45 minutes.
P4 2.80 GHz taking about 7 hours 20 minutes.
Another P4 2.80 GHz took 11 hours 27 minutes (maybe a problem?).
P4 3.00 GHz took 10 hours 30 minutes.
P4 3.06 GHz took 17 hours 47 minutes. (maybe a problem?)
Pentium D 2.80 GHz took 6 hours 35 minutes.
Found a Intel model 8 stepping 3 at 651 MHz it took 25 hours 15 minutes.

Hope this helps, the last one is relavent to your question.

____________

Nightbird
Volunteer tester

Joined: Oct 2 06
Posts: 35
ID: 129
Credit: 11,804
RAC: 0
Message 1408 - Posted 14 Nov 2006 20:07:40 UTC - in response to Message ID 1399 .

> @ clownius,

I have been running Docking on an AMD Opteron 275 and now a Opteron 285 using Linux. Over the last 2 days I have added an AMD 4800+ and an Intel P4 (Windows).
The Opteron 285 runs at 2.6 GHz and takes about 2 hours 30 minutes (Linux).
The Opteron 275 runs at 2.2 GHz and takes about 2 hours 51 minutes (Linux).
The 4800+ runs at 2.4 GHz and takes about 5 hours 50 minutes (Windows XP).
The P4 is a 2.53 GHz @ 2.75 GHz and takes about 8 hours 51 minutes (Windows XP).

This may give you a guide.
Scanning finished jobs found the following, all run Windows
A 2.0 GHz Pentium M processor doing it in around 6 hours 45 minutes.
P4 2.80 GHz taking about 7 hours 20 minutes.
Another P4 2.80 GHz took 11 hours 27 minutes (maybe a problem?).
P4 3.00 GHz took 10 hours 30 minutes.
P4 3.06 GHz took 17 hours 47 minutes. (maybe a problem?)
Pentium D 2.80 GHz took 6 hours 35 minutes.
Found a Intel model 8 stepping 3 at 651 MHz it took 25 hours 15 minutes.

Hope this helps, the last one is relavent to your question.

I will add
Barton 3200+ (2.2 Ghz) - 6h28min (Win2k)
Profile suguruhirahara
Forum moderator
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 282
ID: 15
Credit: 56,614
RAC: 0
Message 1411 - Posted 15 Nov 2006 0:58:40 UTC

My linux machine completed 15% or so, though 7 hours are spent... It seems too slow. What's wrong?:(
____________

I'm a volunteer participant; my views are not necessarily those of Docking@Home or its participating institutions.

Profile John Hunt
Volunteer tester
Avatar

Joined: Nov 14 06
Posts: 40
ID: 270
Credit: 114,129
RAC: 0
Message 1418 - Posted 15 Nov 2006 11:49:37 UTC


My first completed WU !
http://docking.utep.edu/result.php?resultid=47638
Final CPU time is in line with Conans observed figure for a Windows Pentium 4 3GHz, mine was a bit longer possibly because I run multiple projects on this single machine.
Also, as noted elsewhere, because I am using 'old' BOINC 5.2.13 (in my opinion, if it ain't broke, don't fix it!) the 'Progress' always showed 0% and the 'time to completion' remained static from start to finish.



____________

Profile David Ball
Forum moderator
Volunteer tester
Avatar

Joined: Sep 18 06
Posts: 274
ID: 115
Credit: 1,634,401
RAC: 0
Message 1419 - Posted 15 Nov 2006 12:36:20 UTC - in response to Message ID 1411 .

My linux machine completed 15% or so, though 7 hours are spent... It seems too slow. What's wrong?:(


I looked at your Linux machine and noticed a couple of things.

1. It only has 512 MB of ram and it could be that something else running on the machine has caused it to reach the point where it's doing a lot of swapping to disk. From the Linux command line, what do you get if you type the command "free" ? Also, on many machines with built in graphics and even some with graphics cards, some of the ram memory is used by the display. Have you recently gone to a larger screen resolution, which would take more memory? Bringing the machine up to 1 GB of ram would probably do wonders.

2. It's a Pentium D, which is dual core, but it says it has 1 CPU. Linux should report dual core as 2 CPU's. The introduction of dual core chips has really confused the terminology. They're really 2 CPU's on one chip, with a little shared circuitry. Most software (like Microsoft windows) licenses software by how many physical chips are plugged into CPU sockets, therefore it's called 1 CPU but 2 cores. Even though Windows says it's 1 CPU for licensing purposes, the operating system has to treat it as 2 CPU's. You'll even see separate graphs for each CPU in the Windows performance monitor. Linux also has to treat it as 2 CPU's and that means that you need the multiple CPU (or multiple core) Linux kernel, which has extra code in it that is not included in the single CPU (Single core) Linux Kernel. Both the single CPU and multiple CPU Linux kernels should have come with your distribution, including some variations on both, such as the "hugemem" kernels, if you have a LOT of memory. You don't need the "hugemem" variation, just the "smp" variation. How many CPU's does the command line program "top" say your system has? Maybe it's just BOINC that is reporting your system as single CPU. I tried to find info on your distribution, but I can't speak or read Japanese and most of the Vine Linux site was only in Japanese.

Hope this helps,

-- David

Profile suguruhirahara
Forum moderator
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 282
ID: 15
Credit: 56,614
RAC: 0
Message 1421 - Posted 15 Nov 2006 16:50:59 UTC - in response to Message ID 1419 .
Last modified: 15 Nov 2006 17:04:02 UTC

Thanks david:)

1. It only has 512 MB of ram and it could be that something else running on the machine has caused it to reach the point where it's doing a lot of swapping to disk. From the Linux command line, what do you get if you type the command "free" ? Also, on many machines with built in graphics and even some with graphics cards, some of the ram memory is used by the display. Have you recently gone to a larger screen resolution, which would take more memory? Bringing the machine up to 1 GB of ram would probably do wonders.

It says that 220/512MB is used, and no swap is used. So this issue isn't due to lack of RAM.

2. It's a Pentium D, which is dual core, but it says it has 1 CPU. Linux should report dual core as 2 CPU's. The introduction of dual core chips has really confused the terminology. They're really 2 CPU's on one chip, with a little shared circuitry. Most software (like Microsoft windows) licenses software by how many physical chips are plugged into CPU sockets, therefore it's called 1 CPU but 2 cores. Even though Windows says it's 1 CPU for licensing purposes, the operating system has to treat it as 2 CPU's. You'll even see separate graphs for each CPU in the Windows performance monitor. Linux also has to treat it as 2 CPU's and that means that you need the multiple CPU (or multiple core) Linux kernel, which has extra code in it that is not included in the single CPU (Single core) Linux Kernel. Both the single CPU and multiple CPU Linux kernels should have come with your distribution, including some variations on both, such as the "hugemem" kernels, if you have a LOT of memory. You don't need the "hugemem" variation, just the "smp" variation. How many CPU's does the command line program "top" say your system has? Maybe it's just BOINC that is reporting your system as single CPU. I tried to find info on your distribution, but I can't speak or read Japanese and most of the Vine Linux site was only in Japanese.

Thanks for explanation in detail:)
The linux in fact runs on virtual machine VMWare and crunches a workunit, and also windows, its host OS, crunches another workunit at the same time. Hence two cores on 1 CPU is actively used. Also I'd love to test Windows besides Linux:)

As the Japanese linux distro is improved with a view of stability, the upgrade option is relatively stricted and conservative. There are yet unavailable to get smp kernel 2.6 for this version of vine linux 4.0. If Kentsfield weren't so expensive, I would buy one and seek for other distro and install kernel-smp!


YET workunits on the linux are being computed and progressed well now, though I don't know why. Hmm... it seems whether workunits are crunched well or not depends on the virtual machine's condition:( I let it crunch workunits as many as possible until it will go bad/mad;-)
____________

I'm a volunteer participant; my views are not necessarily those of Docking@Home or its participating institutions.
Nightbird
Volunteer tester

Joined: Oct 2 06
Posts: 35
ID: 129
Credit: 11,804
RAC: 0
Message 1424 - Posted 15 Nov 2006 20:35:40 UTC - in response to Message ID 1418 .


------
Also, as noted elsewhere, because I am using 'old' BOINC 5.2.13 (in my opinion, if it ain't broke, don't fix it!) the 'Progress' always showed 0% and the 'time to completion' remained static from start to finish.

At least, i'm not the only one now with the 'progress bar' =0 and the 'time to completion' static.
Profile David Ball
Forum moderator
Volunteer tester
Avatar

Joined: Sep 18 06
Posts: 274
ID: 115
Credit: 1,634,401
RAC: 0
Message 1427 - Posted 15 Nov 2006 23:19:24 UTC - in response to Message ID 1421 .


Thanks for explanation in detail:)
The Linux in fact runs on virtual machine VMWare and crunches a work unit, and also windows, its host OS, crunches another work unit at the same time. Hence two cores on 1 CPU is actively used. Also I'd love to test Windows besides Linux:)


Oh, that makes a lot of difference. I've never run VMWare, but I think most virtual environments are limited to 1 CPU for the client OS, no matter how many the host computer has.

Keeping in mind that I've never run VMWare, it might have been swapping. If the host OS (Windows) decided to swap some of the memory allocated to the VMWare virtual machine, the client OS (Linux) might not know about it. As far as I know, the Client OS thinks it has 512 MB of ram and some swap space on disk. I don't think the client OS would know if part of it's ram was really on disk in the host OS swap space.

On some of the virtual machine managers that run on Linux and only host Linux virtual machines, they have both a guaranteed and a burstable memory limit on the client Linux VPS (Virtual Private Servers running a Linux client OS). From what I can tell, the virtual machine manager allows the sum of the "Guaranteed" ram for the clients to equal the host Linux machines ram + swap space. In other words, a host machine with 2 GB of ram and 4 GB of swap space can host several VPS with a total for all the virtual Linux machines "Guaranteed" ram being about 6 GB.

Things get really confusing when Virtual Machines are involved. I have run a VPS, but it's been months since I had access to it so I've forgotten exactly what top and free reported on the client OS. Since it was Linux hosted on Linux (that's all the virtual environment supported), it was probably different from Linux hosted on Windows. You might have been able to choose to run a different client version of Linux than the host Linux machine but the VPS I had was running the same for both host and client. It was more of a partitioned environment rather than a true virtual machine, but you could reboot a VPS without having any effect on the other VPSs running on the same host. I don't know how it changed if the client Linux was a different version but I think that was possible (not sure on this point).

Virtual machines do some weird stuff!!!

Also, Windows has a LOT of OS processes running in the background, which results in a lot of context switches that flush a lot of the CPU cache memory. Is there any chance that the host machine was running a virus scan or something like that when the work unit was running so slow.

One of these days, I hope to get around to trying VMWare. It sounds like you've done a good job of getting it to work. How hard was it to install VMWare and setup Linux as a client OS?

BTW, I have 2 machines that should be roughly equivalent running Docking @ Home. The Windows XP machine ( Socket 478 Celeron 2.3 GHz with 128 KB l2 cache and 1.5 GB DDR Ram ) takes about 10 hours and 20 minutes of CPU time to run a Docking work unit. The Linux machine ( Socket A Sempron 2500+ with 256KB cache and 1 GB DDR ram on Fedora core 3 with a 2.6 kernel ) takes about 4 hours and 8 minutes of CPU time to run a Docking work unit. That's a Socket A Sempron so it uses a FSB (Front Side Buss) like the old Duron and Athlon chips instead of having a memory controller like the modern Semprons. For some reason, Climate Prediction takes almost exactly the same time on both machines (3.9 seconds per timestep). I guess it depends on what kind of code you're running. I'm almost certain that Linux does lazy writes to disk (puts it in a buffer and flushes the buffer about every 5 seconds or earlier if it needs the memory). I think Windows defaults to doing the disk write immediately and then returning to the program doing the write. With all the debug info being written to disk in the current Docking application, I suspect that is why it takes so much longer on windows. The work units use cumulative CPU time though, not cumulative real time so the extra CPU time for the docking work units on windows would probably be caused by:

1. Windows counts some or all of the time used in the disk write as CPU time used?
2. By not doing a Lazy write, windows does a context switch for each disk write and that (plus running the actual disk write code) causes the cache to get flushed.
3. Both of the above.

Note that Climate Prediction, which takes about the same time on both machines, only does a disk write every few minutes.

I hope this makes sense. These are confusing subjects, even for an old x86 assembler language programmer like myself. I try to keep up with the various cpu architectures when I have the time and I read a lot of the linux kernel mailing lists, but these days I don't have as much time for programming and most of what I do program is in c/c++. I still have the compiler produce an assembly language listing for me when I'm trying to optimize something though. It's amazing how a subtle difference can speed up code, especially with out-of-order execution and cache line alignment, etc. The architecture of modern CPUs is amazing!!!

Regards,

-- David
Profile Atomic Booty
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 59
ID: 23
Credit: 30,370
RAC: 0
Message 1429 - Posted 15 Nov 2006 23:54:47 UTC

David, I appreciate for your detailed explanations of things. You have a natural knack for being able to break complicated subjects down into simpler terms that even I can understand. There are a few other folks around here that are also good at troubleshooting/explaining technical issues and happy to help out fellow crunchers; and to all of you, I'd like to say, "Thanks for taking the time."

Atomic

Profile suguruhirahara
Forum moderator
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 282
ID: 15
Credit: 56,614
RAC: 0
Message 1439 - Posted 16 Nov 2006 11:26:17 UTC - in response to Message ID 1427 .
Last modified: 16 Nov 2006 11:29:07 UTC

Agree with miss booty:) Thanks for tough work!

Keeping in mind that I've never run VMWare, it might have been swapping. If the host OS (Windows) decided to swap some of the memory allocated to the VMWare virtual machine, the client OS (Linux) might not know about it. As far as I know, the Client OS thinks it has 512 MB of ram and some swap space on disk. I don't think the client OS would know if part of it's ram was really on disk in the host OS swap space.

I'm confusing on this point when I see Windows task manager. It says that vmware-vmx is just using 3.4MB, while charmm on windows is using 6.5MB. Yet the manager says that physical memory is used 80%. In addition, System monitor showed that 0 bytes is used, whereas 138/512MB RAM is used. Also my host OS windows is running without stress, so I think swap isn't used so heavily that charmm is run on swap. but it exists lag between the time spent and that really used to crunch. I'm puzzled in many pointa.

One of these days, I hope to get around to trying VMWare. It sounds like you've done a good job of getting it to work. How hard was it to install Vmware and setup Linux as a client OS?

It's easy for you, and you can get install any distro you want, with some config files and ISO image of the distribution such as fedora, debian. But you'll need many RAM if you'd like to install SUSE, which I tried to install but the host OS has freezed. There are plenty of install guide on the net, so it's not hard for you to get manuals of "vmware player"

I guess it depends on what kind of code you're running. I'm almost certain that Linux does lazy writes to disk (puts it in a buffer and flushes the buffer about every 5 seconds or earlier if it needs the memory). I think Windows defaults to doing the disk write immediately and then returning to the program doing the write. With all the debug info being written to disk in the current Docking application, I suspect that is why it takes so much longer on windows.

I'm not sure on this point, but charmm for linux is ver 5.2 and disk writing process is probably different from each charmm.

I hope this makes sense. These are confusing subjects, even for an old x86 assembler language programmer like myself. I try to keep up with the various cpu architectures when I have the time and I read a lot of the linux kernel mailing lists, but these days I don't have as much time for programming and most of what I do program is in c/c++. I still have the compiler produce an assembly language listing for me when I'm trying to optimize something though. It's amazing how a subtle difference can speed up code, especially with out-of-order execution and cache line alignment, etc. The architecture of modern CPUs is amazing!!!

As you may know by readin a thread on docking@home science, the science app named charmm is written in FORTRAN, not c/c++, and according to Dr Andre, the system has been improved for many years around 20 years. I'm not quite sure on this point, but I feel there isn't large room for them to make a quite radical change on the application.

____________

I'm a volunteer participant; my views are not necessarily those of Docking@Home or its participating institutions.
Profile KWSN - A Shrubbery
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 24
ID: 12
Credit: 25,912
RAC: 0
Message 1452 - Posted 16 Nov 2006 21:37:45 UTC - in response to Message ID 1396 .

Just out of interest on what i should attach to this project.

I just threw a 2Ghz Celery system at it as a test and was wondering if i should even bother on my 700Mhz system?



Might be a little late for this discussion but a 700mhz system is a PIII class?

If so, there have been some issues with validating results. The CPU will handle the work just fine but you may get no credit for it. Last I heard it's still being worked on. The project staff could give you more detailed information.
____________
KWSN - A Shrubbery
http://www.kwsnforum.com
Nicolas
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 66
ID: 17
Credit: 29,510
RAC: 0
Message 1454 - Posted 16 Nov 2006 21:51:50 UTC - in response to Message ID 1427 .


Thanks for explanation in detail:)
The Linux in fact runs on virtual machine VMWare and crunches a work unit, and also windows, its host OS, crunches another work unit at the same time. Hence two cores on 1 CPU is actively used. Also I'd love to test Windows besides Linux:)


Oh, that makes a lot of difference. I've never run VMWare, but I think most virtual environments are limited to 1 CPU for the client OS, no matter how many the host computer has.

VMware can work with up to two CPUs, provided that the physical hardware has two or more CPUs (dual core, two CPUs, or even single P4 with Hyperthreading).
Profile suguruhirahara
Forum moderator
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 282
ID: 15
Credit: 56,614
RAC: 0
Message 1460 - Posted 17 Nov 2006 1:20:55 UTC - in response to Message ID 1454 .


Thanks for explanation in detail:)
The Linux in fact runs on virtual machine VMWare and crunches a work unit, and also windows, its host OS, crunches another work unit at the same time. Hence two cores on 1 CPU is actively used. Also I'd love to test Windows besides Linux:)


Oh, that makes a lot of difference. I've never run VMWare, but I think most virtual environments are limited to 1 CPU for the client OS, no matter how many the host computer has.

VMware can work with up to two CPUs, provided that the physical hardware has two or more CPUs (dual core, two CPUs, or even single P4 with Hyperthreading).

Yes, but it needs to purchase VMWare ESX Server to run smp kernel. If one use VMWare player, the number of available CPU is limited to 1.

____________

I'm a volunteer participant; my views are not necessarily those of Docking@Home or its participating institutions.
Nicolas
Volunteer tester
Avatar

Joined: Sep 13 06
Posts: 66
ID: 17
Credit: 29,510
RAC: 0
Message 1482 - Posted 18 Nov 2006 23:54:21 UTC - in response to Message ID 1460 .


Thanks for explanation in detail:)
The Linux in fact runs on virtual machine VMWare and crunches a work unit, and also windows, its host OS, crunches another work unit at the same time. Hence two cores on 1 CPU is actively used. Also I'd love to test Windows besides Linux:)


Oh, that makes a lot of difference. I've never run VMWare, but I think most virtual environments are limited to 1 CPU for the client OS, no matter how many the host computer has.

VMware can work with up to two CPUs, provided that the physical hardware has two or more CPUs (dual core, two CPUs, or even single P4 with Hyperthreading).

Yes, but it needs to purchase VMWare ESX Server to run smp kernel. If one use VMWare player, the number of available CPU is limited to 1.

Oh well... I use a not-exactly-purchased Vmware Workstation.. shh! >:)

Message boards : Number crunching : How long are the Work Units going to be?

Database Error
: The MySQL server is running with the --read-only option so it cannot execute this statement
array(3) {
  [0]=>
  array(7) {
    ["file"]=>
    string(47) "/boinc/projects/docking/html_v2/inc/db_conn.inc"
    ["line"]=>
    int(97)
    ["function"]=>
    string(8) "do_query"
    ["class"]=>
    string(6) "DbConn"
    ["object"]=>
    object(DbConn)#35 (2) {
      ["db_conn"]=>
      resource(150) of type (mysql link persistent)
      ["db_name"]=>
      string(7) "docking"
    }
    ["type"]=>
    string(2) "->"
    ["args"]=>
    array(1) {
      [0]=>
      &string(50) "update DBNAME.thread set views=views+1 where id=31"
    }
  }
  [1]=>
  array(7) {
    ["file"]=>
    string(48) "/boinc/projects/docking/html_v2/inc/forum_db.inc"
    ["line"]=>
    int(60)
    ["function"]=>
    string(6) "update"
    ["class"]=>
    string(6) "DbConn"
    ["object"]=>
    object(DbConn)#35 (2) {
      ["db_conn"]=>
      resource(150) of type (mysql link persistent)
      ["db_name"]=>
      string(7) "docking"
    }
    ["type"]=>
    string(2) "->"
    ["args"]=>
    array(3) {
      [0]=>
      object(BoincThread)#3 (16) {
        ["id"]=>
        string(2) "31"
        ["forum"]=>
        string(1) "2"
        ["owner"]=>
        string(3) "100"
        ["status"]=>
        string(1) "0"
        ["title"]=>
        string(40) "How long are the Work Units going to be?"
        ["timestamp"]=>
        string(10) "1163894061"
        ["views"]=>
        string(4) "1889"
        ["replies"]=>
        string(2) "29"
        ["activity"]=>
        string(20) "9.8572744439152e-128"
        ["sufferers"]=>
        string(1) "0"
        ["score"]=>
        string(1) "0"
        ["votes"]=>
        string(1) "0"
        ["create_time"]=>
        string(10) "1158276489"
        ["hidden"]=>
        string(1) "0"
        ["sticky"]=>
        string(1) "0"
        ["locked"]=>
        string(1) "0"
      }
      [1]=>
      &string(6) "thread"
      [2]=>
      &string(13) "views=views+1"
    }
  }
  [2]=>
  array(7) {
    ["file"]=>
    string(63) "/boinc/projects/docking/html_v2/user/community/forum/thread.php"
    ["line"]=>
    int(184)
    ["function"]=>
    string(6) "update"
    ["class"]=>
    string(11) "BoincThread"
    ["object"]=>
    object(BoincThread)#3 (16) {
      ["id"]=>
      string(2) "31"
      ["forum"]=>
      string(1) "2"
      ["owner"]=>
      string(3) "100"
      ["status"]=>
      string(1) "0"
      ["title"]=>
      string(40) "How long are the Work Units going to be?"
      ["timestamp"]=>
      string(10) "1163894061"
      ["views"]=>
      string(4) "1889"
      ["replies"]=>
      string(2) "29"
      ["activity"]=>
      string(20) "9.8572744439152e-128"
      ["sufferers"]=>
      string(1) "0"
      ["score"]=>
      string(1) "0"
      ["votes"]=>
      string(1) "0"
      ["create_time"]=>
      string(10) "1158276489"
      ["hidden"]=>
      string(1) "0"
      ["sticky"]=>
      string(1) "0"
      ["locked"]=>
      string(1) "0"
    }
    ["type"]=>
    string(2) "->"
    ["args"]=>
    array(1) {
      [0]=>
      &string(13) "views=views+1"
    }
  }
}
query: update docking.thread set views=views+1 where id=31