From: Ben Allan (baallan_at_ca_dot_sandia_dot_gov)
Date: Thu May 27 2004 - 10:43:32 PDT
Thanks for the prompt reply. We'll give that a try and let you know. Would you prefer we keep upc-users on copy, or should we just mail jcduell_at_lbl_dot_gov thanks, ben At 10:26 AM 5/27/2004, jcduell_at_lbl_dot_gov wrote: >On Wed, May 26, 2004 at 04:16:23PM +0000, Ben Allan wrote: > >> setenv MPI_CC /usr/bin/gcc >> setenv MPI_CFLAGS "-I/usr/local/lampi/include" >> setenv MPI_LIBS /usr/local/lampi/lib/libmpi.a >> >> we get >> "checking for working MPI configuration... no" > >Hi Ben, > >Could you do a couple quick things for me? > >1) Download the 'stable' build from our website, and see if it builds: > > http://mantis.lbl.gov/upc/berkeley_upc-stable.tar.gz > > You're better off with our stable build than the 1.1.0 release > anyway--at this point it's more reliable for a wider range of UPC > code than our official release (which we're going to update soon). > >2) Email us the 'config.log' file in the 'gasnet' subdirectory of your > build tree (NOT the one in the top level directory). It will contain > any compiler/linker errors from the failed attempt to use your > MPI_CC, etc., to compile a trivial MPI program (just MPI_Init and > MPI_Finalize). If you're feeling helpful, you could even take a > quick peek at the error message (just search for 'working MPI' and > the error message(s) should follow soon after), and tell me if it > looks like anything obvious, given your MPI compiler and system. > >I don't know if we've built against LAMPI before, so this is probably >just a minor glitch somewhere. > >Cheers, > >-- >Jason Duell Future Technologies Group ><jcduell_at_lbl_dot_gov> Computational Research Division >Tel: +1-510-495-2354 Lawrence Berkeley National Laboratory