User Tools

Site Tools


arcca:raven

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
arcca:raven [2012/12/04 11:52] – [LALSuite] stephen.fairhurst@LIGO.ORGarcca:raven [2012/12/20 13:11] (current) – [Grid tools] stephen.fairhurst@LIGO.ORG
Line 26: Line 26:
   * pkg-config: This is available by default on Raven (as /usr/bin/pkg-config).  The installed version is 0.23, which seems to suffice   * pkg-config: This is available by default on Raven (as /usr/bin/pkg-config).  The installed version is 0.23, which seems to suffice
   * gsl: This is available as a module.  So run <code bash> module load gsl</code> to pick this up.  Current version is 1.15, which is more recent than version 1.12 listed on the DASWG page.   * gsl: This is available as a module.  So run <code bash> module load gsl</code> to pick this up.  Current version is 1.15, which is more recent than version 1.12 listed on the DASWG page.
-  * git: This is required to check out the LALSuite software stack.  It **should** be available as a module, but isn't working yet To work around the issueI've installed version 1.8.1 in /home/spxsf2/opt/lscsoft/non-lsc.+  * git: This is required to check out the LALSuite software stack.  It is available in /usr/bin (version 1.7.1)a newer version (1.7.9) is available with <code bash> module load git</code>
   * fftw: This **should** be available as a module, but isn't working yet!  I have installed it at /home/spxsf2/opt/lscsoft/ **Note**: the instructions call for fftw to be built twice -- float and double.  Not sure if this is still relevant and/or can be set up on Raven by default.   * fftw: This **should** be available as a module, but isn't working yet!  I have installed it at /home/spxsf2/opt/lscsoft/ **Note**: the instructions call for fftw to be built twice -- float and double.  Not sure if this is still relevant and/or can be set up on Raven by default.
   * libframe/metaio/glue: This is LIGO specific software.  It is not available as a module, but in the longer term, we should see if we can get it installed.  For now, I downloaded the "stable" version from [[http://www.lsc-group.phys.uwm.edu/daswg/download/software/source/ | here]] and have installed (following the instructions  [[https://www.lsc-group.phys.uwm.edu/daswg/docs/howto/lscsoft-install.html | here]]) at /home/spxsf2/opt/lscsoft/.   * libframe/metaio/glue: This is LIGO specific software.  It is not available as a module, but in the longer term, we should see if we can get it installed.  For now, I downloaded the "stable" version from [[http://www.lsc-group.phys.uwm.edu/daswg/download/software/source/ | here]] and have installed (following the instructions  [[https://www.lsc-group.phys.uwm.edu/daswg/docs/howto/lscsoft-install.html | here]]) at /home/spxsf2/opt/lscsoft/.
-  * User environment script: This is to set up your environment, same instructions as for libframe and friends.  It seems more complicated than strictly necessary.  It winds up making the file lscsoft-user-env.sh.  I think that the majority of this is unnecessary.  For now, I'll just source it, but if we do try installing our software as modules, then I think we should look at what's really needed here.+  * User environment script: This is to set up your environment, same instructions as for libframe and friends.  It seems more complicated than strictly necessary.  It winds up making the file lscsoft-user-env.sh.  If you want to pick up the software I have installed, you need to run <code bash>. /home/spxsf2/opt/lscsoft/lscsoft-user-env.sh</code>  I think that the majority of this is unnecessary.  For now, I'll just source it, but if we do try installing our software as modules, then I think we should look at what's really needed here.
    
 ===== LALSuite ===== ===== LALSuite =====
Line 43: Line 43:
   * This seems to work, other than "../../../lalinspiral/src/LALInspiralSBankOverlap.c:22:19: fatal error: fftw3.h: No such file or directory compilation terminated." which I got around by removing this file from the make.   * This seems to work, other than "../../../lalinspiral/src/LALInspiralSBankOverlap.c:22:19: fatal error: fftw3.h: No such file or directory compilation terminated." which I got around by removing this file from the make.
   * To pick up this install, you need to <code bash>. /home/spxsf2/opt/lalsuite/master/etc/lscsoftsrc</code>   * To pick up this install, you need to <code bash>. /home/spxsf2/opt/lalsuite/master/etc/lscsoftsrc</code>
 +
 +I have successfully run a piece of LAL code (lalapps_tmpltbank to be precise), and it seems to have worked!
 +
 +===== Grid tools =====
 +
 +Instructions to install the LIGO Data Grid Client from source are taken from [[https://www.lsc-group.phys.uwm.edu/lscdatagrid/doc/installclient-source.html | here]].
 +
 +<code>
 +wget http://www.globus.org/ftppub/gt5/5.2/5.2.0/installers/src/gt5.2.0-all-source-installer.tar.gz
 +tar xf gt5.2.0-all-source-installer.tar.gz 
 +mkdir gt5.2.0-all
 +export GLOBUS_LOCATION=~/gt5.2.0-all/
 +export PATH=/bin:/usr/bin; 
 +export FLAVOUR=gcc64dbg
 +
 +cd gt5.2.0-all-source-installer
 +./configure --prefix=$GLOBUS_LOCATION --with-flavor=$FLAVOUR
 +make gsi-openssh
 +make postinstall
 +. $GLOBUS_LOCATION/etc/globus-user-env.sh
 +</code>
 +
 +The VDT Certificate Bundle can be installed using the instructions from [[https://www.lsc-group.phys.uwm.edu/lscdatagrid/doc/installclient-source.html | the same page]]. (Note the link to the archive is out of date). I was also found necessary to update the certificates.
 +
 +<code>
 +wget http://software.grid.iu.edu/pacman/cadist/1.32/osg-certificates-1.32.tar.gz
 +tar xf osg-certificates-1.32.tar.gz -C $GLOBUS_LOCATION/share
 +globus-update-certificate-dir
 +</code>
 +
 +Now copy your Grid certificates into the ''.globus'' folder in your home directory and make sure the permissions are correct.
 +
 +<code>
 +chmod 600 ~/.globus/usercert.pem 
 +chmod 400 ~/.globus/userkey.pem 
 +</code>
 +
 +To source the install, you need:
 +<code>
 +.  ~spxph/gt5.2.0-all/etc/globus-user-env.sh
 +</code>
  
 ===== Data ===== ===== Data =====
  
-Do we have any data and where is it?+  * We have data at <code bash>/scratch/LIGO/LDR/</code> This can be found using, for example 
 +<code bash>ligo_data_find --observatory L --url-type file --gps-start-time 832326736 --gps-end-time 832328926 --output L-L1_RDS_C03_L2_CACHE-832326736-2190.lcf --lal-cache  --type L1_RDS_C03_L2 --match localhost --server=ldr-arcca.phys.uwm.edu</code> 
 +  * I **believe** that /scratch on Raven is a different file-server from /scratch on Merlin/GEO.  Also, I think that LDR is running on GEO.  So, the data_find query will likely return the correct answer, but this is only going to be true as long as we don't move the data around or ask for any new data.  We need to get LDR running on a machine that can see the new cluster storage. 
 +  * We don't have a clear statement of the data that is available on Raven.  It would be nice to know what we had on Merlin that's now on Raven.  The only page I could find is [[cardiff:computing:available_data | this one]] for coma.  I expect Merlin/Raven have about the same data, but it would be good to know for sure. 
 + 
 +===== Workflows ===== 
 + 
 +  * Raven is set up to run under PBSpro.  Our workflows are written to output condor dags.   
 +  * We had been running by submitting requests to PBS to reserve nodes for condor which then reported back to GEO.  This setup should work on the new cluster, provided we have a machine running condor that can talk to the nodes. 
 +  * In the medium term, we might set things up differently so that condor talks to the PBS submission machine and gets the right jobs submitted in the PBS queue.  This removes the need for the machine running condor to talk to the nodes. 
 +  * It would be nice to set up something to run a few jobs as proof of principle.  Is there any easy way to do this? 
 + 
 +====== BAM on Raven ====== 
 +  * Load these modules: 
 +<code bash> 
 +module load intel/intel 
 +module load bullxmpi/bullxmpi-1.1.17.1 
 +</code> 
 +  * Use a standard MyConfig for bam -- you don't need to point it to specific mpi libraries; just use mpicc as the compiler. 
 +  * (Re)compile bam: 
 +<code bash>make clean && make -j 8 bam</code> 
 + 
 +  * Sample pbs script (adapted from the Merlin version with some changes): 
 +<code bash> 
 +#!/bin/bash 
 + 
 +#PBS -q workq 
 +#PBS -l select=8:ncpus=16:mpiprocs=16 
 +#PBS -l place=scatter:excl 
 +#PBS -l walltime=1:00:00 
 + 
 +#PBS -N R6_PN_64_128 
 +#PBS -o R6_PN_64.out 
 +#PBS -e R6_PN_64.err 
 + 
 +#PROJECT=PR37 
 + 
 +pardir=/home/spxmp/MachineConfig/ARCCA/ 
 +parfile=R6_PN_64.par 
 +bamexe=/home/spxmp/MachineConfig/bam/exe/bam 
 + 
 +cd /scratch/spxmp 
 +cp $pardir/$parfile . 
 + 
 +mpirun -np 128 $bamexe ./$parfile 
 +</code> 
 + 
 +  * This pbs-script copies the parameter file over to the lustre filesystem in /scratch/$USER. This is where you should run, i.e. *not* in your home directory. 
 +  * Don't forget to mirror the data with rsync to minion, since files will get deleted after some time. 
 +  * At the moment we are using queue 'workq'. There is also a dedicated 'ligo' queue and we should maybe use that.
  
arcca/raven.1354621975.txt.gz · Last modified: 2012/12/04 11:52 by stephen.fairhurst@LIGO.ORG