Multi-processor problems

Kian-Tat Lim ktl at wag240.caltech.edu
Sat Jan 13 05:52:20 AEST 1990


	On our 4D/240GTXB running 3.1F, we have observed similar
effects.  For a small, nearly-completely parallelizable program, here
are some representative timings (the compute-bound job was running in
the background while these tests were run):

1 compute-bound job + 1 thread:  14.5u 0.1s 0:14 98%
1 compute-bound job + 2 threads:  7.0u 0.1s 0:07 98%
1 compute-bound job + 3 threads:  4.9u 0.1s 0:05 98%
1 compute-bound job + 4 threads:  6.5u 0.1s 0:07 87%
1 compute-bound job + 5 threads: 17.0u 0.1s 0:26 65%
1 thread alone: 14.5u 0.1s 0:14 98%
2 threads alone: 7.0u 0.1s 0:07 96%
4 threads alone: 3.8u 0.1s 0:04 96%

A computation-less version of the same program that just did m_forks
was timed as follows (without the compute-bound background job):

1 thread:   0.3u 0.1s 0:00 73%
2 threads:  0.3u 0.0s 0:00 86%
3 threads:  0.3u 0.1s 0:00 82%
4 threads:  0.3u 0.1s 0:00 81%
5 threads: 18.9u 0.1s 0:28 66%

Multiple parallel jobs ("time cmd & time cmd")

2 x 2 threads: 7.4u 0.1s 0:07 98%
2 x 4 threads: 34.0u 0.1s 1:08 49%

Increasing the computation per thread by 25 times gave:

4 threads: 109.0u 1.1s 1:50 99%
2 x 4 threads: 150.8u 7.4s 5:11 50%

	There appear to be severe scheduling problems with more than
four simultaneous threads of execution.  We're in the process of
confirming these numbers and sending a letter to SGI; if someone has a
fix, we'd love to hear it.

--
Kian-Tat Lim (ktl at wagvax.caltech.edu, KTL @ CITCHEM.BITNET, GEnie: K.LIM1)



More information about the Comp.sys.sgi mailing list