- genevb's home page
- Posts
- 2025
- 2024
- 2023
- 2022
- September (1)
- 2021
- 2020
- 2019
- 2018
- 2017
- December (1)
- October (3)
- September (1)
- August (1)
- July (2)
- June (2)
- April (2)
- March (2)
- February (1)
- 2016
- November (2)
- September (1)
- August (2)
- July (1)
- June (2)
- May (2)
- April (1)
- March (5)
- February (2)
- January (1)
- 2015
- December (1)
- October (1)
- September (2)
- June (1)
- May (2)
- April (2)
- March (3)
- February (1)
- January (3)
- 2014
- 2013
- 2012
- 2011
- January (3)
- 2010
- February (4)
- 2009
- 2008
- 2005
- October (1)
- My blog
- Post new blog entry
- All blogs
Estimate for Run 24 pp200_production_LowLuminosity production
Updated on Thu, 2025-06-12 17:32. Originally created by genevb on 2025-06-12 11:37.
Nightly test of SL7, 32-bit, optimized is ~2 CPU-seconds per event. First event is ~70-80 CPU-seconds, which becomes negligible at O(1%) for files with a few thousand events.
Total events=~3.3B and total files=~50k:
So files have on average ~67k events in them, but many files have something like 110k events in them!
Thus, for 32-bit optimized, we need about ~7e9 CPU-seconds = ~80k CPU-days. That's ~80 days if given 1k CPUs.
I tried a 32-bit vs. 64-bit comparison by running the nightly test job interactively at the same time on a single node, and got a ratio of 1.3:1 for duration. That's close enough to use 4:3 as the ratio, which leads to...
64-bit optimized would be about ~60 days (~2 months) given 1k CPUs.
The other question is how long the longest jobs might take. At 2 CPU-seconds per event, a file with 1.1e5 events = 2.2e5 CPU-seconds = ~2.5 CPU-days, and 64-bit optimized should be on the order of ~2 CPU-days for the longest jobs even if they're a little slower. Both of these are just fine.
-Gene
Total events=~3.3B and total files=~50k:
> get_file_list.pl -keys 'sum(events)' -cond year=2024,trgsetupname=pp200_production_LowLuminosity,filetype=online_daq,storage=hpss,sname2=st_physics 3338594169 > get_file_list.pl -keys 'sum(events)' -cond year=2024,trgsetupname=pp200_production_LowLuminosity,filetype=online_daq,storage=hpss,sname2=st_physics_adc 15632831 > get_file_list.pl -keys 'count(fdid)' -cond year=2024,trgsetupname=pp200_production_LowLuminosity,filetype=online_daq,storage=hpss,sname2=st_physics 49831(Ignoring the ADC files, which are essentially negligible...)
So files have on average ~67k events in them, but many files have something like 110k events in them!
Thus, for 32-bit optimized, we need about ~7e9 CPU-seconds = ~80k CPU-days. That's ~80 days if given 1k CPUs.
I tried a 32-bit vs. 64-bit comparison by running the nightly test job interactively at the same time on a single node, and got a ratio of 1.3:1 for duration. That's close enough to use 4:3 as the ratio, which leads to...
64-bit optimized would be about ~60 days (~2 months) given 1k CPUs.
The other question is how long the longest jobs might take. At 2 CPU-seconds per event, a file with 1.1e5 events = 2.2e5 CPU-seconds = ~2.5 CPU-days, and 64-bit optimized should be on the order of ~2 CPU-days for the longest jobs even if they're a little slower. Both of these are just fine.
-Gene
»
- genevb's blog
- Login or register to post comments