Tags:
view all tags
---+ Local CRAB This page is intended to describe how to use CRAB in T3-Higgs %TOC% ---++ Environment : I'm hoping that users are familiar with the [[https://twiki.cern.ch/twiki/bin/view/CMSPublic/WorkBookCRABTutorial][usual CRAB recipe]]. For the specific environments here they are : <verbatim> source /cvmfs/cms.cern.ch/cmsset_default.sh # CMSSW cmsrel RELEASE ; cd RELEASE/src ; cmsenv # Just a guideline on how to get cmsenv to run *the order is important* source /cvmfs/oasis.opensciencegrid.org/osg-software/osg-wn-client/3.2/current/el5-x86_64/setup.sh # Grid tools source /share/apps/crab/crab.sh </verbatim> ---++ Configuration Everything is usually whatever you want, except these : <verbatim> [USER] return_data = 1 copy_data = 0 [CRAB] scheduler = condor </verbatim> This will ensure that we will find the right CMSSW.sh that will help us achieve faster the goal of setting the stage-out to Hadoop. ---++ Steps After your environment is set and the crab.cfg is configured, time to create the jobs with crab -create. This will give you a task directory : <verbatim> working directory /home/samir/CMSSW_5_3_11/src/crab_0_140705_005804/ </verbatim> Now is the part where we set the stageout location, you need to use the taskdir to edit CMSSW.sh : /home/samir/CMSSW_5_3_11/src/crab_0_140705_005804/job/CMSSW.sh Towards the end of the file, search for "file_list". Just below a line like this : <verbatim> file_list="$SOFTWARE_DIR/outfile_$OutUniqueID.root" </verbatim> Is where we will place the copy command : <verbatim> cp $RUNTIME_AREA/outfile_* /mnt/hadoop/store/user/$USER/crabtest/ </verbatim> *NOTE*: "outfile" was in my case where I specify it in crab.cfg. Make sure that you pick the right filename for the copy command. You will have a hint some lines above in the same script. <verbatim> output_file = outfile.root </verbatim> That should be all. Now you can submit your jobs and if the input data is in the right place (/mnt/hadoop) your jobs should run and copy the output files to the directory you specify ---++ Running jobs with Xrootd input Is out of the scope of this document to teach how to use Xrootd, but if you already know, here is how you successfully run a job with Xrootd input, in our T3 : * The main point is that we need to get rid of the canonical proxy location and use one in our home directories, which is shared with the worker nodes, check where is your proxy : <verbatim> -bash-3.2$ voms-proxy-info subject : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira/CN=proxy issuer : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira identity : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira type : proxy strength : 1024 bits path : /tmp/x509up_u2611 timeleft : 191:59:58 </verbatim> If you see it there, *delete* it. It will only stay on your way. * Create a proxy in your home directory, I have chosen to do it in my CMSSW project area : <verbatim> -bash-3.2$ voms-proxy-init -voms cms -valid 192:00 -out $PWD/samir.proxy export X509_USER_PROXY=$PWD/samir.proxy # You also need to tell the system about the unusual location -bash-3.2$ pwd /home/samir/CMSSW_5_3_11/src </verbatim> * Make sure that all is fine by running voms-proxy-info : <verbatim> -bash-3.2$ voms-proxy-info subject : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira/CN=proxy issuer : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira identity : /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=samir/CN=695732/CN=Samir Cury Siqueira type : proxy strength : 1024 bits path : /home/samir/CMSSW_5_3_11/src/samir.proxy timeleft : 191:58:21 </verbatim> * Good. As in our system the environment of job submission is the job environment, your home directory is visible from the nodes so the proxy will be found and used. In my case the output was happily sitting in the output dir, as specified previously. ---++ Running jobs with SRM stage-out (most recommended) If you follow the steps above for Xrootd Input, you get for free the ability of copying your output to any grid site. Including of course, T2_US_Caltech. The method is pretty standard, as you would with any CRAB task, in my file for example : <verbatim> [USER] return_data = 0 copy_data = 1 storage_element = T2_US_Caltech data_location_override = T2_US_Caltech user_remote_dir = CrabXrootd2 </verbatim> So I get under /store/user/samir/CrabXrootd2 the output. That's all you need to worry about -- and you won't need to patch CSSW.sh as described in the first local stage-out method. -- Main.samir - 2014-07-05
Edit
|
Attach
|
Watch
|
P
rint version
|
H
istory
:
r5
<
r4
<
r3
<
r2
<
r1
|
B
acklinks
|
V
iew topic
|
Raw edit
|
More topic actions...
Topic revision: r2 - 2014-07-06
-
samir
Home
Site map
Main web
Sandbox web
TWiki web
Main Web
Users
Groups
Index
Search
Changes
Notifications
RSS Feed
Statistics
Preferences
View
Raw View
Print version
Find backlinks
History
More topic actions
Edit
Raw edit
Attach file or image
Edit topic preference settings
Set new parent
More topic actions
Account
Log In
Edit
Attach
Copyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki?
Send feedback