Bug 311551
Summary: | openmpi programs over tcp layer with more than one 1bX configured fails to start | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 5 | Reporter: | Gurhan Ozen <gozen> |
Component: | openmpi | Assignee: | Doug Ledford <dledford> |
Status: | CLOSED WORKSFORME | QA Contact: | |
Severity: | medium | Docs Contact: | |
Priority: | medium | ||
Version: | 5.1 | CC: | jburke |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | All | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2007-10-01 15:53:57 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Gurhan Ozen
2007-09-28 18:58:37 UTC
I suspect that this is a local configuration issue. With two ib interfaces up I'm still perfectly able to run an mpi job over tcp. [root@pe840 ~]# ifconfig eth0 Link encap:Ethernet HWaddr 00:15:C5:F6:00:FE inet addr:192.168.33.125 Bcast:192.168.35.255 Mask:255.255.252.0 inet6 addr: 2002:a00:0:1:215:c5ff:fef6:fe/64 Scope:Global inet6 addr: fe80::215:c5ff:fef6:fe/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:1398620 errors:0 dropped:0 overruns:0 frame:0 TX packets:196291 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:263800471 (251.5 MiB) TX bytes:23703692 (22.6 MiB) Interrupt:169 ib0 Link encap:InfiniBand HWaddr 00:00:04:04:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00 inet addr:10.250.2.254 Bcast:10.250.2.255 Mask:255.255.255.0 inet6 addr: fe80::205:ad00:3:491/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:2044 Metric:1 RX packets:3550868 errors:0 dropped:0 overruns:0 frame:0 TX packets:3962304 errors:0 dropped:9 overruns:0 carrier:0 collisions:0 txqueuelen:128 RX bytes:6146383878 (5.7 GiB) TX bytes:6303006983 (5.8 GiB) ib1 Link encap:InfiniBand HWaddr 00:00:04:05:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00 inet addr:10.250.3.254 Bcast:10.250.3.255 Mask:255.255.255.0 inet6 addr: fe80::205:ad00:3:492/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:2044 Metric:1 RX packets:3366691 errors:0 dropped:0 overruns:0 frame:0 TX packets:3763472 errors:0 dropped:9 overruns:0 carrier:0 collisions:0 txqueuelen:128 RX bytes:6130123368 (5.7 GiB) TX bytes:6165782998 (5.7 GiB) lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:688 errors:0 dropped:0 overruns:0 frame:0 TX packets:688 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:82139 (80.2 KiB) TX bytes:82139 (80.2 KiB) [root@pe840 ~]# mpirun -np 2 -host ib0test1,ib0test2 -mca btl tcp,self /usr/bin/mpitests-com ################################################################################ com Point-to-Point MPI Bandwidth and Latency Benchmark Version 1.4.0 Run at 10/01/07 11:48:34, with rank 0 on ibtest1.test.redhat.com ################################################################################ Test Processes Op Size (bytes) Ops BW (MB) ----------------------------------------------------------------- Unidirectional 2 32 10 2.556 Unidirectional 2 64 10 5.627 Unidirectional 2 128 10 11.558 Unidirectional 2 256 10 22.526 Unidirectional 2 512 10 52.022 Unidirectional 2 1024 10 94.917 Unidirectional 2 2048 10 111.286 Unidirectional 2 4096 10 161.933 Unidirectional 2 8192 10 203.260 Unidirectional 2 16384 10 242.796 Unidirectional 2 32768 10 339.351 Unidirectional 2 65536 10 257.030 Unidirectional 2 131072 10 303.131 Unidirectional 2 262144 10 349.909 Unidirectional 2 524288 10 357.937 Unidirectional 2 1048576 10 357.985 Unidirectional 2 2097152 10 366.657 Unidirectional 2 4194304 10 379.115 Unidirectional 2 8388608 10 381.135 Test Processes Op Size (bytes) Ops BW (MB) ----------------------------------------------------------------- Bidirectional 2 32 10 1.985 Bidirectional 2 64 10 3.360 Bidirectional 2 128 10 7.645 Bidirectional 2 256 10 14.463 Bidirectional 2 512 10 29.998 Bidirectional 2 1024 10 50.770 Bidirectional 2 2048 10 64.068 Bidirectional 2 4096 10 85.793 Bidirectional 2 8192 10 107.259 Bidirectional 2 16384 10 118.359 Bidirectional 2 32768 10 122.310 Bidirectional 2 65536 10 100.895 Bidirectional 2 131072 10 122.261 Bidirectional 2 262144 10 131.490 Bidirectional 2 524288 10 133.315 Bidirectional 2 1048576 10 130.882 Bidirectional 2 2097152 10 132.207 Bidirectional 2 4194304 10 134.658 Bidirectional 2 8388608 10 143.842 Test Processes Op Size (bytes) Ops Latency (us) ----------------------------------------------------------------- Latency 2 0 10 49.591 Max Unidirectional Bandwidth : 381.13 for message size of 8388608 bytes Max Bidirectional Bandwidth : 143.84 for message size of 8388608 bytes ################################################################################ Test Parameters --------------- Process pair allocation : block MB size for BW calculation : 1000000 Barrier not included in measurement. Bandwidth calculated as sum of process bandwidths. MPI_Wtick returns 0.000001000 MPI_Wtime overhead 0.000000318 ################################################################################ [root@pe840 ~]# |