1. Trang chủ
  2. » Kỹ Thuật - Công Nghệ

So sánh hiệu năng các phần mềm cài đặt giao thức MPICH, LAM/MPI và PVM trên cụm máy tính Linux qua mạng Fast Ethernet. doc

8 665 0

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 8
Dung lượng 5,13 MB

Nội dung

Among message-passing protocols such as MPI, PVM, ESP, MPI MessagePassing Interface and PVM Parallel Virtual Machine are adopted as most popular protocols for distributed memory computin

Trang 1

A COMPARATIVE STUDY ON PERFORMANCE OF MPICH, LAM/MPI

NGUYEN HAl CHAU

Abstract Cluster computing provides a distributed memory model to users and therefore requires mes-sage-passingprotocols to exchange data Among message-passing protocols (such as MPI, PVM, ESP), MPI (MessagePassing Interface) and PVM (Parallel Virtual Machine) are adopted as most popular protocols for distributed memory computing model In this paper, we give a practical comparative study on the perfor-mance of MPICH 1.2.1, LAM/MPI6.3.2 and PVM 3.4.2 implementations of the MPI and PVM protocols,

on a Linux cluster over our Fast Ethernet network We alsocompare some parallel applications' performance runging over the three environments

T6m tJ{t Cum may tinh cung dLp cho ngiro-isd' dung m9t moi tru-o-ngtinh toan theo kie'u b9 nh& ph an tan, do do c'an co cac giaothirc chuye'n thong di~pMtrao do'i dir li~u Trong s5 cac giaothirc chuye'n thong di~p (vi du MPI, PVM, ESP), MPI va PVM 111,cac giao thtrc dtro'c su: dung nhieu nhfit Trong bai nay, chung toi dira raSlr so sanh hieu nang ctia cac phan mern cai d~t cac giao tlnrc MPI va PVM: MPICH 1.2.1, LAM/MPI 6.3.2 va PVM 3.4.2 tren cum may tinh Linux dtro'c Ht n5i qua mang Fast Ethernet

1 INTRODUCTION

In recent years, cluster computing has been growing quickly because of low cost of fast network hardware equipments and workstations Many universities, institutes and research groups started

to use low cost clusters to meet their demands of parallel processing instead of expensive super-computers or mainframes [1,4] Linux clusters has been increasingly using today due to their free distribution and open source policy Cluster computing provides a distributed memory model to users/programmer and therefore requires message-passing protocols for exchanging data Among message passing protocols such as MPI [6]' PVM [15]' BSP [13] MPI (Message Passing Interface) and PVM (Parallel Virtual Machine) are most widely adopted for cluster computing Two implemen-tations of MPI, MPICH [7] and LAM/MPI [5],are most widely used MPICH comes from Argonne National Laboratory and LAM/MPI is maintained by the University of Notre Dame PVM's imple-mentation Oak Ridge National Laboratory (ORNL) is also popular The software can be ported to many different platforms and acted as cluster middleware, over which parallel compilers for parallel languages such as HPF, HPC++ can be implemented

Due to greet requirements of large parallel applications, network traffic in computer cluster is increasing heavily Therefore performance of cluster middleware is one of important factors that affect performance parallel applications running on clusters Since PVM, LAM and MPICH all us TCP/IP to exchange messages among nodes of a cluster, it is useful to investigate PVM, LAM and MPICH performance together with TCP /IP performanc to assist one make a right choice his/her cluster configuration

In this paper, we will practically evaluate performance of MPICH 1.2.1, LAM/MPI 6.3.2 and PVM 3.4.2 on Linux Cluster if Institute of Physics, Hanoi, Vietnam in terms of latency and peak throughput To conduct performance tests, we use NetPIPE, a network protocol independent per-formance evaluation tool [12], developed by Ames Laboratory/Scalable Computing Lab, USA We also compare performance of some parallel applications running over the three cluster middleware packages The remaining parts of this paper are organized as follows: In Section 2, we give a brief description of computer cluster architecture and some cluster middleware In Section 3 we describe our testing environment Results evaluation will be given in Section 4 In the last section, we provide conclusions and future works

Trang 2

NGUYEN HAl CHAU

- Computers

- Operating systems such as Linux, FreeBSD

- High speed network connections and swith hes such as Ethernet, Fast Ethernet, Gigabit

Ether-net, Myrinet

ben hmarkin tolls such as ADAPTOR, XMPI, XPVM, XMTV, LinPACK

Se quential applications I I Parallel programming environment

Cluster middleware

NIC NIC

as Comm S/W

as Comm S/W

NIC

as Comm S/W

High-speed network

Fig. 1 Cluster architecture

Since parallel and distributed applications consume much of cluster, especially network

2.2 Cluster middleware

Trang 3

Hewlett-Packard and others supported it In addition, there are competing implementations of MPI

The MPI Chameleon (MPICH) began in 1993 It was developed at Argonne National Laboratory

The Local Area Multi-computer (LAM or LAM/MPI) was launched at Ohio Supercomputing

The following are features summary of the three message passing packages [11]

Table 1 LAM/MPI, MPICH and PVM features

3 Testing Environment

Our testing environment for performance comparsion consists of 6 Intel Pentium III at 600 MHz

swith (24 ports) by a RealTek 10/100 auto sensing NIC and a category 5 cable The computers are also connected back-to-back forTCP/IP verus MVIA [8] additional performance testing In this test,

installed RedHat Linux 6.2 with 2.2.14 kernel

Trang 4

NGUYEN HAl CHAU

for acknowledgement for receiver and con.iinue transmitting other packets The receiver will answer

an acknowledgement to sender when it receives appropriate packets LAM supports two modes of

communicating The first one is C2C (client-to-client) a d the other isLAMD (LAM daemon) C2Q allows processes to exchange data directly without notifying LAM daemon process in contrast with

,00 f ~ - ~- - - -r - · r~ - - '- - -· ·" '- - "

M t V _ I OO O -

MT 300 • •.

_ MTU 1S0Q

o~L o.- ~~ ~ • 1 • ~ _ J - ,

Al o ~ ! lle i n by t e

80

6

~

s

i

('; ' 0

20

1001 - ' - '

-I

- e - - _ _ , r _ ~ • _ _ - o · ' \

I

r:

I

MTU ,., 100 -

r > , .,.i - _ _

l j

8

' 0

' 0

T HR O U G H T Q R APH

; ir r <!

performance ofLAM/MPI and MPICH, we implemented the two packages with their default pa

Trang 5

LAM/MPl's performance is better in C2C mode as shown in Fig 6.

We also did the above tests with an Intel 10/100 hub (8 ports) and found that the letency of LAM/MPI, MPICH and PVM was increased by 15-16% and their peak throughput wa reduced by 10% in compare with the test conducted with the switch

80

7

60

50

30

20

IAMC2C 3 2KB - ~

lAMC2C MK B - _

LAMC2C 1 28KB lAMC2C 2S6K8

L - • ~ - l ' L_" ' - -' ~_ """ _~L - ' I - _ L - '

-o

1 10 10 0 1000 1 00 0 100000 le ~ O ~ 1 e.07

r u oc ksrze i n b y ' "

70

50

1

1

-I AMC2C 32K 8 _ I

lAMC2C 6 K8 - -

-LAMC2C 128KB

o _ , J.L- ' L L ~ L_ " , ~~~2~3~~~f -' L

20

1£ ' - 0 0 on o - 0001 0 01 0

T i m e

Fig s. LAM/MPI performance in C2C mode versus short message sise

70

60

5

30

2

L A MO 3 K9

LA MO 1 8KB

~ - ' LJ - ~ L ~_ ~ - , ~ LA _ M~p_ 2 SGK B """"" '- - '

o

1 1 0000 t boooo 1( ' HOf> IN07

B t oc k si ze in b t e

80 (- - • • ~ • , .-~~- - • - ~ - -.- ~ " t -.-.- - 70

S I G N Al ! JRE GRAPH

60

50

30

2

10

LAMD 32KS _

L A M O 6 4KB '"

LAMO 128KB LAMO 256KB

oL-_ -'-~' -_L~_ .' _ L ' _

T i me

Fig 4. LAM/MPI performance in LAMD mode versus sh rt me sage s r ze

10

10

Trang 6

NGUYEN HAl CHAU

T a le 2. Performance compariso of LAM/MPICH and PVM

80 r - -.• · ' r • r '" "-" ' ~ "r ' - 0r r ,

• :~h i

~ ' r4 ~_.tli,t ~

70

6

J

Ll

i=

30

2

i

10

l A M l MP I

-M P IC H

P M

10 10 0 ' 000 10 0 00 100 0 le • 06 l e • 07

ruoc k sr z in b yt e

i i

-.-7

so

~.,t<'"

: '

I i!

· i ·

u

i

~

!

i

i

50

' 0

30

20 ·

S I GN A TUR E GR A

1

L A M / ~ I

MP I C ••• •

Time

Table 9 Simple applications comparison

Trang 7

Table 4. A molecular dynamics simulation's performance

Table 5 Overall comparison on performance of LAM/MPI, MPICH and PVM

better than that in LAMD mode

fhl • - "'

1

-60

-30

2

'0

o

1

nnI

10

, :\

lAMC2C 64KB

, ,,:c : ~ , _ · : l _ L _ - J _L ' - • L o ~ o - _ ~.l~~ , ~ _ • ~

Btr-cksbe in byte

70

'

-60

'0

30

-T i me

Fig 6 LAMD and LAMC2C modes in comparison

5 CONCLUSIONS

Trang 8

cluster is a difficult task because of the present of many software packages for cluster computing, this study may help forpeople who want to design and implement a Linux cluster for parallel computation

in providing decision As a result of performance testing, we has been running LAM/MPI 6.3.2 for Linux PC cluster in Institute of Physics, Hanoi, Vietnam because of its low latency and highest peak thro ghput in comparison with MPICH and PVM In addition, by practice aspect, we found that LAM/MPI launches ends and clears up its parallel applications more quickly than MPICH does Our future works can be expressed as follows The cluster of Institute of Physics will be used for scientific comp ting such as particle physics, high-energy physics and molecular dynamics simula-tion Thus benchmarking the cluster with NPB [9] (NAS Parallel Benchmark) and LinPACK [16] is important Due to great demands of parallel ap lications, there are many efforts to improve TCP /IP performance However the TCP /IP improvement is in moderate progress as its delay while data passing layers of protocol stacks and seems not to meet large parallel applications' requirements VIA (Virtual Interface Architecture) is developed recently for speed up communication ability in cluster and obtained promising results by bypassing protocol tacks to reduce data transfer delay We will conduct performance comparison of parallel applications in LAM/MPI, MPICH and MVICH, an implementation of MPI over MVIA

Acknowledgements The author wishes to thank Prof Ho Tu Bao (JAIST), Dr Ha Quang Thuy (Vietnam National University, Hanoi) and Dr Nguyen Trong Dung (JAIST) for their supports and advices

REFERENCES

[1] A Apon, R Buyya, H Jin, J Mache, Cluster Computing in the Classroom: Topics, Guidelines, and Experiences, http://www.csse.monash.edu.au/~rajkumar /papers/CC-Edu.pdf

[2] ADAPTOR - GMD's High Performance Fortran Compilation System,

http:www.gmd.de/SCAI/lab / adaptor/adaptor home.htrnl

[3] Foster, J Geisler, W Gropp, N Karonis, E Lusk, G Thiruvathukal, S Tuecke, Wide-area implementation of the Message Passing Interface, Parallel Computing 24 (1998) 1734-1749 [4] K.A Hawick, D A Grove, P D Coddington, M.A Buntine, Commodity Cluster Computing for Computational Chemistry, DHPC Technical Report DHPC-073, University of Adelaide, Jan 2000

[5] LAM/MPI Parallel Computing, http:/www.mpi.nd.edu/larn

[6] MPI Forum, http://www/mpi-forum.org/docs/docs.html

[7] MPICH - A Portable MPI Implementation, http://www-unix.mcs.anl.gov/mpi/mpich

[8] M-VIA: A High Performance Modular VIA for Linux, http://www /nersc.gov /research/FTG /via [9] NAS Parallel benchmark, http://www.nas.nasa.gov /NPB

[10] P Crernonesi, E Rosti, G Serazzi, E Smirni, Performance evaluation of parallel systems,

Parall e l C omputing 25 (1999) 1677-1698

[11] P.H Carns, W.B Logon III, S P McMillan, R B Ross, An Evaluation of Message Passing Implementations on Beowulf Workstations,

http://parlweb.parl.clemson.edu/ <-spmcmily aero99 / eval.htm

[12] Quinn O Snell, Armin R Mikler, and John L Gustafson, NetPIPE: A Network Protocol Independent Performance Evaluator, http://www /scl/ ameslab.gov /netpipe/paper /full.html

[13] S.R Donaldson, J M D Hill, D.B Skillicorn, BSP clusters: High performance, reliable and

very low cost, Parallel Computing 26 (2000) 199-242.

[14] The Beowulf project, http://www.beowulf.org

[15] The PVM project, http://www.epm.ornl/gov /pvm

[16] Top 500 clusters, http://www/top500clusters.org

Received March 19, 2001 Institute of Physics, NCST of Vietnam

Ngày đăng: 25/03/2014, 20:22

TỪ KHÓA LIÊN QUAN

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN

w