-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
use input_args.json file in CMSRunAnalysis.py via --jobId arg #8869
Comments
harder than it looked. Can't have a single |
my branch https://github.com/belforte/CRABServer/tree/use-jobid-arg-for-CMSRunAnalysis-8869 One possibility is to insert the input_args.json in the PreJob picking the correct json file. But in order to keep CMSRunAnalysis scripts simple and for mind sanity it is better that file in the WN have constant names. Back to ... add new jobs to same input_args.json as they are created by DagmanCreator |
making automatic splitting work is still tricky because of lots of old code which is not broken but likely useless. E.g. there are some arguments to CMSRunAnalysis like |
first task with automatic splitting which submitted the processing DAG and completed successfully |
rats... creation of tail dag gets stuck insite PreDag at
i.e. inside Splitter |
Big mistery, I even tried to kill the predag and the dag bootstrap, but now I find that log is completed, 3 tails had been submittedf and completed finely
|
new automatic splitting test with branch https://github.com/belforte/CRABServer/tree/use-jobid-arg-for-CMSRunAnalysis-8869 |
task ran finely, even if again spending almost 40 min on Splitting
For peace of mind I am submitting same task to prod scheduler |
the task which ran in productions still spend an amazing amount of time in Splitter
anyhow speeding up AutomaticSplittting is not a goal, nor likely possible nor useful. So I conclude that new code is working as well as old one and will merge #8883 |
Get rid of the awful long arg list in Job.submit (ref. dmwm/CRABClient#5288 (comment))
Arguments = "-a $(CRAB_Archive) --sourceURL=$(CRAB_ISB) --jobNumber=$(CRAB_Id) --cmsswVersion=$(CRAB_JobSW) --scramArch=$(CRAB_JobArch) '--inputFile=$(inputFiles)' '--runAndLumis=$(runAndLumiMask)' --lheInputFiles=$(lheInputFiles) --firstEvent=$(firstEvent) --firstLumi=$(firstLumi) --lastEvent=$(lastEvent) --firstRun=$(firstRun) --seeding=$(seeding) --scriptExe=$(scriptExe) --eventsPerLumi=$(eventsPerLumi) --maxRuntime=$(maxRuntime) '--scriptArgs=$(scriptArgs)' -o $(CRAB_AdditionalOutputFiles)"
and make it possible to use instead simply
--jobId=$(CRAB_Id)
Now every task has
input_args.json
in its spool dir, so add it to list oftransfer_input_files
inJob.submit
and add the new arg to CMSRunAnalysis.pyThis will also allow to simplify the
RunJobs.dag
file where all those env.vars. are definedVARS Job1 count="1" runAndLumiMask="job_lumis_1.json" lheInputFiles="False" firstEvent="None" firstLumi="None" lastEvent="None" firstRun="None" maxRuntime="-60" eventsPerLumi="None" seeding="AutomaticSeeding" inputFiles="job_input_file_list_1.txt" scriptExe="None" scriptArgs="[]" +CRAB_localOutputFiles="\"kk.root=kk_1.root\"" +CRAB_DataBlock="\"/GenericTTbar/HC-CMSSW_9_2_6_91X_mcRun1_realistic_v2-v2/AODSIM#3517e1b6-76e3-11e7-a0c8-02163e00d7b3\"" +CRAB_Destination="\"davs://eoscms.cern.ch:443/eos/cms/store/user/belforte/GenericTTbar/crab_20241216_165508/241218_101115/0000/log/cmsRun_1.log.tar.gz, davs://eoscms.cern.ch:443/eos/cms/store/user/belforte/GenericTTbar/crab_20241216_165508/241218_101115/0000/kk_1.root\""
Since everything is in
input_args.json
!jobId
arg so can test along current code--jobId
to--jobNumber
The text was updated successfully, but these errors were encountered: