Download
Download
net/publication/2330092
CITATIONS READS
9 318
2 authors:
All content following this page was uploaded by Prabhas Chongstitvatana on 15 November 2015.
6 nodes (Asyn) 16
Speedup
10 nodes (Syn)
70 10 nodes (Asyn)
13
10
60
7
50
4
40 1
0 10 20 30 40 50 60 70 80 90 100 1 2 3 4 5 6 7 8 9 10
Communication Computation
In the second implementation, we attempt to fur-
100%
ther improve the speedup of the parallel algorithm by 90%
the asynchronous migration. When the fastest node 80%
phase, the other nodes must wait. The waiting time will 20%
800
tations of Genetic Programming successfully exploit the
600
computing resource of a dedicated cluster of PC work-
stations. Superlinear speedup of GP can be acquired by
400
improving a coarse-grained model for parallelization as
200
less computational work needs to be done. Furthermore,
0 the timing analyses indicate the scalability of the paral-
2 (Syn) 2 (Asyn) 4 (Syn) 4 (Asyn) 6 (Syn) 6 (Asyn) 10 (Syn) 10 (Asyn)
Number of Processors
lel approaches, as the size of the problem increases, the
speedup will be improved.
Figure 5: Absolute time spent in communication References
[1] Chongstitvatana P (1998), Improving robustness of
robot programs generated by genetic programming
of the computation time of each node. As the number for dynamic environments. Proc. of IEEE Asia Pa-
of nodes is increased, the computation time per node is cific Conference on Circuits and Systems, p.523–526
decreased. Hence, the barrier time is reduced.
In contrast, the barrier time in the asynchronous im- [2] Koza JR, Andre D (1995), Parallel genetic program-
plementation increases as the number of processors in- ming on a network of transputers. Proc. of the Work-
creases. This is due to the fact that the time spent in the shop on Genetic Programming: From Theory to
second barrier (waiting for the next broadcast) increases Real-World Applications, University of Rochester,
with the number of nodes. However, the asynchronous National Resource Laboratory for the Study of Brain
implementation eliminates the first barrier therefore it and Behavior, Technical Report 95-2, p.111-120
reduces the total time in the barriers compared to the [3] Dracopoulos DC, Kent S (1996), Bulk synchronous
synchronous implementation. parallelisation of genetic programming. Proc. of the
The absolute time spent in a broadcast increases con- Third International Workshop on Applied Parallel
siderably – greater than linear. From the inspection in Computing in Industrial Problems and Optimization
the trace information by using a visualization tool, we (PARA ’96), Springer Verlag, Berlin
found that the transmission of the broadcast functions
in the implementation of MPI that we use may be exe- [4] Punch B (1998), How effective are multiple poplu-
cuted more than once, especially for a large number of lations in genetic programming. Proc. of the Third
processors. Annual Conference in Genetic Programming, pp.308-
313
After obtaining some timing analyses, the results re-
veal the cause of the problem. The performance degra- [5] Lin S-C, Punch WF, Goodman ED (1994), Coarse-
dation in 10 processors is caused by the excessive com- grain parallel genetic algorithms: Categorization and
munication time due to the broadcast function. Al- new approach. Proc. of the Sixth IEEE SPDP, pp.28-
though the asynchronous migration reduces the barrier 37
time effectively compared to the synchronous migration,
the increase in the communication time in 10 processors [6] Tongchim S, Chongstitvatana P (1999), Speedup Im-
obliterates this advantage. In case of the small number provement on Automatic Robot Programming by
of processors (2,4,6), the gain from the asynchronous mi- Parallel Genetic Programming. Proc. of 1999 IEEE
gration is considerable as the evolution proceeds at the International Symposium on Intelligent Signal Pro-
speed of the fastest node. cessing and Communication Systems (ISPACS’99),
As the size of the work increases (i.e., the number Phuket, Thailand
of training environments increases), the serial and par- [7] Cantú-Paz E (1999), Designing efficient and accurate
allel computation time will be increased when the time parallel genetic algorithms. PhD thesis, University of
spent in the communication is constant. If the ratio Illinois at Urbana-Champaign
of the computation/communication can be kept large
(large work load), then one can expect that the parallel
performance will be improved.