From: ¬Ð˾´ (xingjinglu_at_gmail_dot_com)
Date: Tue Oct 28 2008 - 15:42:02 PST
Paul,
Thank you for your reply, I have tried your advice, the problem is still
there.
For 1) I add the GASNET_MPI_SPAWNER=ssh it the upc.sh, but it did notint to
the problem; else if I use "vapi-conduit"it will cause other problem, like
lack the library for vapi.
For 2) , I changed the MPIRUN_CMD=/home/autopar/mpich/mpirun -N %M %P
..... in GASNET_MPI.pl under the directory of /upc/dbg/bin and /upc
/opt/bin ,but it doesn't matter.
And I have tried install the upc with the network of vapi, but there are
still the same problem.
Now, I guess the problem maybe casued by the mpich.
When I install upc:
#/bin/sh
export PATH=/home/autopar/mpich2/bin:$PATH
export LD_LIBRARY_PATH=/home/autopar/mpich2/lib:$LD_LIBRARY_PATH
bash
Configure CC=gcc CXX=g++ MPI_CC=mpicc -enable-mpi
--prefix=/home/autopar/upc
Then make and make install
Wish your reply.
----------------------------------------------------------------------------
------------
Eric,
1) Lines 3-6 of the warnings tell you you should be looking to use the
native "vapi-conduit" for communiction not MPI. I am not sure why it is
not getting used by default, but passing "-network vapi" to upcc should
ensure it gets used. You may also want/need to set
GASNET_VAPI_SPAWNER=ssh in your environment, because the rest of your
problem seems linked to mpirun (which is the default way to launch
processes w/ vapi-conduit).
2) The first warning tells you we can't control process layout with your
mpirun. You should consider leaving off the '-N 4' when you run to rid
yourself of the warning. The problem of getting the process run on the
right nodes should be solved one of two ways:
2a) Switch to vapi-conduit and set GASNET_VAPI_SPAWNER=ssh in your
environment as mentioned above
or
2b) Figure out what environment variables or command line flags are
needed to get a "hello world" MPI application run the way you want. Then
set MPIRUN_CMD as needed. Once you know how to run MPI apps the way you
want, we can help with the MPIRUN_CMD details.
Let us know if you need more help.
-Paul
luxingjing wrote:
>
> Hi,
>
> After I intstalled upc2.6.0 version and the network is mpich1.2.7 version;
>
> I compile and run the program like below:
>
> Upcc -T=32 hello.c ¨Co hello
>
> Upcrun ¨CN 4 hello
>
> But in fact, all the threads are layed out on one node only. My
> environment set shows below:
>
> export
>
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/autopar/mpich1.2.7/lib:/home/autopar/
upc_mpi1/dbg/lib:/home/autopar/upc_mpi1/opt/lib
>
> export PATH=/home/autopar/mpich1.2.7/bin:/home/autopar/upc_mpi1/bin:$PATH
>
> export UPC_NODES="node17 node18 node19 node20"
>
> And the the warnings are:
>
> WARNING: Don't know how to control process->node layout with your mpirun
>
> WARNING: PROCESS LAYOUT MIGHT NOT MATCH YOUR REQUEST
>
> WARNING: Using GASNet's mpi-conduit, which exists for portability
> convenience.
>
> WARNING: Support was detected for native GASNet conduits: vapi
>
> WARNING: You should *really* use the high-performance native GASNet
> conduit
>
> WARNING: if communication performance is at all important in this
> program run.
>
> UPCR: UPC thread 31 of 32 on gnode20 (process 31 of 32, pid=19952)
>
> UPCR: UPC thread 2 of 32 on gnode20 (process 2 of 32, pid=19718)
>
> UPCR: UPC thread 7 of 32 on gnode20 (process 7 of 32, pid=19758)
>
> UPCR: UPC thread 15 of 32 on gnode20 (process 15 of 32, pid=19822)
>
> UPCR: UPC thread 27 of 32 on gnode20 (process 27 of 32, pid=19920)
>
> UPCR: UPC thread 29 of 32 on gnode20 (process 29 of 32, pid=19936)
>
> UPCR: UPC thread 25 of 32 on gnode20 (process 25 of 32, pid=19904)
>
> UPCR: UPC thread 13 of 32 on gnode20 (process 13 of 32, pid=19806)
>
> UPCR: UPC thread 24 of 32 on gnode20 (process 24 of 32, pid=19894)
>
> UPCR: UPC thread 22 of 32 on gnode20 (process 22 of 32, pid=19878)
>
> UPCR: UPC thread 23 of 32 on gnode20 (process 23 of 32, pid=19886)
>
> UPCR: UPC thread 21 of 32 on gnode20 (process 21 of 32, pid=19870)
>
> UPCR: UPC thread 9 of 32 on gnode20 (process 9 of 32, pid=19774)
>
> UPCR: UPC thread 19 of 32 on gnode20 (process 19 of 32, pid=19854)
>
> UPCR: UPC thread 11 of 32 on gnode20 (process 11 of 32, pid=19790)
>
> UPCR: UPC thread 8 of 32 on gnode20 (process 8 of 32, pid=19766)
>
> UPCR: UPC thread 20 of 32 on gnode20 (process 20 of 32, pid=19862)
>
> UPCR: UPC thread 26 of 32 on gnode20 (process 26 of 32, pid=19912)
>
> UPCR: UPC thread 18 of 32 on gnode20 (process 18 of 32, pid=19846)
>
> UPCR: UPC thread 28 of 32 on gnode20 (process 28 of 32, pid=19928)
>
> UPCR: UPC thread 0 of 32 on gnode20 (process 0 of 32, pid=19706)
>
> UPCR: UPC thread 12 of 32 on gnode20 (process 12 of 32, pid=19798)
>
> UPCR: UPC thread 1 of 32 on gnode20 (process 1 of 32, pid=19710)
>
> UPCR: UPC thread 17 of 32 on gnode20 (process 17 of 32, pid=19838)
>
> UPCR: UPC thread 30 of 32 on gnode20 (process 30 of 32, pid=19944)
>
> UPCR: UPC thread 6 of 32 on gnode20 (process 6 of 32, pid=19750)
>
> UPCR: UPC thread 5 of 32 on gnode20 (process 5 of 32, pid=19742)
>
> UPCR: UPC thread 14 of 32 on gnode20 (process 14 of 32, pid=19814)
>
> UPCR: UPC thread 10 of 32 on gnode20 (process 10 of 32, pid=19782)
>
> UPCR: UPC thread 4 of 32 on gnode20 (process 4 of 32, pid=19734)
>
> UPCR: UPC thread 16 of 32 on gnode20 (process 16 of 32, pid=19830)
>
> UPCR: UPC thread 3 of 32 on gnode20 (process 3 of 32, pid=19726)
>
> Hello World from Thread 0 (of 32 THREADS)
>
> Hello World from Thread 30 (of 32 THREADS)
>
> Hello World from Thread 2 (of 32 THREADS)
>
> Hello World from Thread 25 (of 32 THREADS)
>
> Hello World from Thread 23 (of 32 THREADS)
>
> Hello World from Thread 1 (of 32 THREADS)
>
> Hello World from Thread 24 (of 32 THREADS)
>
> Hello World from Thread 10 (of 32 THREADS)
>
> Hello World from Thread 4 (of 32 THREADS)
>
> Hello World from Thread 14 (of 32 THREADS)
>
> Hello World from Thread 26 (of 32 THREADS)
>
> Hello World from Thread 9 (of 32 THREADS)
>
> Hello World from Thread 31 (of 32 THREADS)
>
> Hello World from Thread 5 (of 32 THREADS)
>
> Hello World from Thread 7 (of 32 THREADS)
>
> Hello World from Thread 8 (of 32 THREADS)
>
> Hello World from Thread 15 (of 32 THREADS)
>
> Hello World from Thread 20 (of 32 THREADS)
>
> Hello World from Thread 18 (of 32 THREADS)
>
> Hello World from Thread 17 (of 32 THREADS)
>
> Hello World from Thread 27 (of 32 THREADS)
>
> Hello World from Thread 16 (of 32 THREADS)
>
> Hello World from Thread 6 (of 32 THREADS)
>
> Hello World from Thread 22 (of 32 THREADS)
>
> Hello World from Thread 21 (of 32 THREADS)
>
> Hello World from Thread 12 (of 32 THREADS)
>
> Hello World from Thread 11 (of 32 THREADS)
>
> Hello World from Thread 13 (of 32 THREADS)
>
> Hello World from Thread 3 (of 32 THREADS)
>
> Hello World from Thread 28 (of 32 THREADS)
>
> Hello World from Thread 29 (of 32 THREADS)
>
> Hello World from Thread 19 (of 32 THREADS)
>
> Wish your help!
>
> Thank you!
>
> Yours Eric.
>
--
Paul H. Hargrove PHHargrove_at_lbl_dot_gov
Future Technologies Group
HPC Research Department Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory Fax: +1-510-486-6900