Re: hi, I encouter a problem

From: Paul H. Hargrove (PHHargrove_at_lbl_dot_gov)
Date: Sun Oct 26 2008 - 17:51:52 PST

  • Next message: ¬�˾�: "(no subject)"
    Eric,
    
    
    1) Lines 3-6 of the warnings tell you you should be looking to use the
    native "vapi-conduit" for communiction not MPI. I am not sure why it is
    not getting used by default, but passing "-network vapi" to upcc should
    ensure it gets used. You may also want/need to set
    GASNET_VAPI_SPAWNER=ssh in your environment, because the rest of your
    problem seems linked to mpirun (which is the default way to launch
    processes w/ vapi-conduit).
    
    2) The first warning tells you we can't control process layout with your
    mpirun. You should consider leaving off the '-N 4' when you run to rid
    yourself of the warning. The problem of getting the process run on the
    right nodes should be solved one of two ways:
    2a) Switch to vapi-conduit and set GASNET_VAPI_SPAWNER=ssh in your
    environment as mentioned above
    or
    2b) Figure out what environment variables or command line flags are
    needed to get a "hello world" MPI application run the way you want. Then
    set MPIRUN_CMD as needed. Once you know how to run MPI apps the way you
    want, we can help with the MPIRUN_CMD details.
    
    Let us know if you need more help.
    
    -Paul
    
    
    luxingjing wrote:
    >
    > Hi,
    >
    > After I intstalled upc2.6.0 version and the network is mpich1.2.7 version;
    >
    > I compile and run the program like below:
    >
    > Upcc -T=32 hello.c �Co hello
    >
    > Upcrun �CN 4 hello
    >
    > But in fact, all the threads are layed out on one node only. My
    > environment set shows below:
    >
    > export
    > LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/autopar/mpich1.2.7/lib:/home/autopar/upc_mpi1/dbg/lib:/home/autopar/upc_mpi1/opt/lib
    >
    > export PATH=/home/autopar/mpich1.2.7/bin:/home/autopar/upc_mpi1/bin:$PATH
    >
    > export UPC_NODES="node17 node18 node19 node20"
    >
    > And the the warnings are:
    >
    > WARNING: Don't know how to control process->node layout with your mpirun
    >
    > WARNING: PROCESS LAYOUT MIGHT NOT MATCH YOUR REQUEST
    >
    > WARNING: Using GASNet's mpi-conduit, which exists for portability
    > convenience.
    >
    > WARNING: Support was detected for native GASNet conduits: vapi
    >
    > WARNING: You should *really* use the high-performance native GASNet
    > conduit
    >
    > WARNING: if communication performance is at all important in this
    > program run.
    >
    > UPCR: UPC thread 31 of 32 on gnode20 (process 31 of 32, pid=19952)
    >
    > UPCR: UPC thread 2 of 32 on gnode20 (process 2 of 32, pid=19718)
    >
    > UPCR: UPC thread 7 of 32 on gnode20 (process 7 of 32, pid=19758)
    >
    > UPCR: UPC thread 15 of 32 on gnode20 (process 15 of 32, pid=19822)
    >
    > UPCR: UPC thread 27 of 32 on gnode20 (process 27 of 32, pid=19920)
    >
    > UPCR: UPC thread 29 of 32 on gnode20 (process 29 of 32, pid=19936)
    >
    > UPCR: UPC thread 25 of 32 on gnode20 (process 25 of 32, pid=19904)
    >
    > UPCR: UPC thread 13 of 32 on gnode20 (process 13 of 32, pid=19806)
    >
    > UPCR: UPC thread 24 of 32 on gnode20 (process 24 of 32, pid=19894)
    >
    > UPCR: UPC thread 22 of 32 on gnode20 (process 22 of 32, pid=19878)
    >
    > UPCR: UPC thread 23 of 32 on gnode20 (process 23 of 32, pid=19886)
    >
    > UPCR: UPC thread 21 of 32 on gnode20 (process 21 of 32, pid=19870)
    >
    > UPCR: UPC thread 9 of 32 on gnode20 (process 9 of 32, pid=19774)
    >
    > UPCR: UPC thread 19 of 32 on gnode20 (process 19 of 32, pid=19854)
    >
    > UPCR: UPC thread 11 of 32 on gnode20 (process 11 of 32, pid=19790)
    >
    > UPCR: UPC thread 8 of 32 on gnode20 (process 8 of 32, pid=19766)
    >
    > UPCR: UPC thread 20 of 32 on gnode20 (process 20 of 32, pid=19862)
    >
    > UPCR: UPC thread 26 of 32 on gnode20 (process 26 of 32, pid=19912)
    >
    > UPCR: UPC thread 18 of 32 on gnode20 (process 18 of 32, pid=19846)
    >
    > UPCR: UPC thread 28 of 32 on gnode20 (process 28 of 32, pid=19928)
    >
    > UPCR: UPC thread 0 of 32 on gnode20 (process 0 of 32, pid=19706)
    >
    > UPCR: UPC thread 12 of 32 on gnode20 (process 12 of 32, pid=19798)
    >
    > UPCR: UPC thread 1 of 32 on gnode20 (process 1 of 32, pid=19710)
    >
    > UPCR: UPC thread 17 of 32 on gnode20 (process 17 of 32, pid=19838)
    >
    > UPCR: UPC thread 30 of 32 on gnode20 (process 30 of 32, pid=19944)
    >
    > UPCR: UPC thread 6 of 32 on gnode20 (process 6 of 32, pid=19750)
    >
    > UPCR: UPC thread 5 of 32 on gnode20 (process 5 of 32, pid=19742)
    >
    > UPCR: UPC thread 14 of 32 on gnode20 (process 14 of 32, pid=19814)
    >
    > UPCR: UPC thread 10 of 32 on gnode20 (process 10 of 32, pid=19782)
    >
    > UPCR: UPC thread 4 of 32 on gnode20 (process 4 of 32, pid=19734)
    >
    > UPCR: UPC thread 16 of 32 on gnode20 (process 16 of 32, pid=19830)
    >
    > UPCR: UPC thread 3 of 32 on gnode20 (process 3 of 32, pid=19726)
    >
    > Hello World from Thread 0 (of 32 THREADS)
    >
    > Hello World from Thread 30 (of 32 THREADS)
    >
    > Hello World from Thread 2 (of 32 THREADS)
    >
    > Hello World from Thread 25 (of 32 THREADS)
    >
    > Hello World from Thread 23 (of 32 THREADS)
    >
    > Hello World from Thread 1 (of 32 THREADS)
    >
    > Hello World from Thread 24 (of 32 THREADS)
    >
    > Hello World from Thread 10 (of 32 THREADS)
    >
    > Hello World from Thread 4 (of 32 THREADS)
    >
    > Hello World from Thread 14 (of 32 THREADS)
    >
    > Hello World from Thread 26 (of 32 THREADS)
    >
    > Hello World from Thread 9 (of 32 THREADS)
    >
    > Hello World from Thread 31 (of 32 THREADS)
    >
    > Hello World from Thread 5 (of 32 THREADS)
    >
    > Hello World from Thread 7 (of 32 THREADS)
    >
    > Hello World from Thread 8 (of 32 THREADS)
    >
    > Hello World from Thread 15 (of 32 THREADS)
    >
    > Hello World from Thread 20 (of 32 THREADS)
    >
    > Hello World from Thread 18 (of 32 THREADS)
    >
    > Hello World from Thread 17 (of 32 THREADS)
    >
    > Hello World from Thread 27 (of 32 THREADS)
    >
    > Hello World from Thread 16 (of 32 THREADS)
    >
    > Hello World from Thread 6 (of 32 THREADS)
    >
    > Hello World from Thread 22 (of 32 THREADS)
    >
    > Hello World from Thread 21 (of 32 THREADS)
    >
    > Hello World from Thread 12 (of 32 THREADS)
    >
    > Hello World from Thread 11 (of 32 THREADS)
    >
    > Hello World from Thread 13 (of 32 THREADS)
    >
    > Hello World from Thread 3 (of 32 THREADS)
    >
    > Hello World from Thread 28 (of 32 THREADS)
    >
    > Hello World from Thread 29 (of 32 THREADS)
    >
    > Hello World from Thread 19 (of 32 THREADS)
    >
    > Wish your help!
    >
    > Thank you!
    >
    > Yours Eric.
    >
    
    
    -- 
    Paul H. Hargrove                          PHHargrove_at_lbl_dot_gov
    Future Technologies Group                 
    HPC Research Department                   Tel: +1-510-495-2352
    Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900
    

  • Next message: ¬�˾�: "(no subject)"