- General information
- Data readiness
- Grid and Cloud
- Articles and publications
- Cloud computing
- Data Management
- Documentation
- Getting site information from VORS
- Globus 1.1.x
- Globus Toolkit Error FAQ
- Intro to FermiGrid site for STAR users
- Introduction to voms proxies for grid cert users
- Job Managers
- Modifying Virtual Machine Images and Deploying Them
- Rudiments of grid map files on gatekeepers
- Running Magellan Cloud at NERSC, Run 11
- SRM instructions for bulk file transfer to PDSF
- Scalability Issue Troubleshooting at EC
- Specification for a Grid efficiency framework
- Starting up a Globus Virtual Workspace with STAR’s image.
- Troubleshooting gsiftp at STAR-BNL
- Using the GridCat Python client at BNL
- Grid Infrastructure
- Grid Production
- Monitoring
- MySQL project activities
- Infrastructure
- Machine Learning
- Offline Software
- Production
- S&C internal group meetings
- Test tree
Intro to FermiGrid site for STAR users
Updated on Wed, 2006-12-27 15:16. Originally created by wbetts on 2006-12-18 19:16.
Under:
Open question: What is the preferred file transfer mechanism?
In GridCat they list an SRM server at srm://fndca1.fnal.gov:8443/ but I have not made any attempt to use it.
The FNAL_FERMIGRID site policy and some documentation can be found here:
http://fermigrid.fnal.gov/policy.html
You must use VOMS proxies (rather than grid certificate proxies) to run at this site. A brief intro to voms proxies is here: Introduction to voms proxies for grid cert users
All users with STAR VOMS proxies are mapped to a single user account ("star").
Technical note: (Quoting from an email that Steve Timm sent to Levente) "Fermigrid1.fnal.gov is not a simple jobmanager-condor. It is emulating the jobmanager-condor protocol and then forwarding the jobs on to whichever clusters have got free slots, 4 condor clusters and actually one pbs cluster behind it too." For instance, I noticed jobs submitted to this gatekeeper winding up at the USCMS-FNAL-WC1-CE site in MonAlisa. (What are the other sites?)
You can use SUMS to submit jobs to this site (though this feature is still in beta testing) following this example:
star-submit-beta -p dynopol/FNAL_FERMIGRID jobDesription.xml
where jobDescription.xml is the filename of your job's xml file.
Site gatekeeper info:
Hostname: fermigrid1.fnal.gov
condor queue is available (fermigrid1.fnal.gov/jobmanager-condor)
If no jobmanager is specified, the job runs on the gatekeeper itself (jobmanager-fork, I’d assume)
[stargrid02] ~/> globus-job-run fermigrid1.fnal.gov
/bin/cat /etc/redhat-release
Scientific Linux Fermi LTS release 4.2 (Wilson)
Fermi worker node info:
[stargrid02] ~/> globus-job-run fermigrid1.fnal.gov/jobmanager-condor /bin/cat /etc/redhat-release
Scientific Linux SL release 4.2 (Beryllium)
[stargrid02] ~/> globus-job-run fermigrid1.fnal.gov/jobmanager-condor /usr/bin/gcc -v
Using built-in specs.
Configured with: ../configure --prefix=/usr --mandir=/usr/share/man --infodir=/usr/share/info --enable-shared --enable-threads=posix --disable-checking --with-system-zlib --enable-__cxa_atexit --disable-libunwind-exceptions --enable-java-awt=gtk --host=i386-redhat-linux
Thread model: posix
gcc version 3.4.4 20050721 (Red Hat 3.4.4-2)
There doesn't seem to be a GNU fortran compiler such as g77 on the worker nodes.
Open question: What is the preferred file transfer mechanism?
In GridCat they list an SRM server at srm://fndca1.fnal.gov:8443/ but I have not made any attempt to use it.
»
- Printer-friendly version
- Login or register to post comments