Status%20of%20PRS/Muon%20Activities - PowerPoint PPT Presentation

About This Presentation
Title:

Status%20of%20PRS/Muon%20Activities

Description:

Complete HLT selection for high-lumi scenario. HLT results on B physics ... probably partly intrinsic to ORCA/COBRA, and partly due to inexperienced ... – PowerPoint PPT presentation

Number of Views:96
Avg rating:3.0/5.0
Slides: 9
Provided by: darina6
Learn more at: http://www.phys.ufl.edu
Category:

less

Transcript and Presenter's Notes

Title: Status%20of%20PRS/Muon%20Activities


1
Status of PRS/Muon Activities
  • D. AcostaUniversity of Florida
  • Work toward the June HLT milestones
  • US analysis environment

2
HLT Milestones
  • The June HLT milestones are
  • Complete HLT selection for high-lumi scenario
  • HLT results on B physics
  • CPU analysis for high lumi selection
  • Repeat on line selection for low-lumi
  • Must have results in DAQ TDR by September!
  • We dont have these results yet, but the current
    status and L1 results were reported at the June
    CMS Week
  • HLT Muon code had severe crashes, infinite
    loops, and memory leaks that prohibited
    collecting any statistics on our HLT algorithms
  • After monumental debugging effort, crashes
    traced to incorrect use of ReferenceCounted
    objects
  • User must never delete, even if performed new!
  • In L1 results, rate spike appeared at ?1.6
  • New holes in geometry?
  • Problems were fixed for ORCA_6_2_0 release

3
L1 CSC Rate Spike
?
?
  • Contributes 1 kHz to L1 rate
  • Spike occurs in ? and ?!
  • Region of crack between barrel/endcap
  • Traced to ambiguity in pT assignment for low pT
    muons (or punch-through)
  • Fixed in CSC Track-Finder (but not sure why this
    is a problem only now)

4
Not out of the woods
  • Tony reports that the Muon RootTreeMaker has a
    massive memory leak (200500kB/event)
  • Analysis stopped at CERN (batch nodes were
    dying)
  • But muon HLT code alone was shown to have a leak
    of only 16 kB/event when released
  • So is it because events have more occupancy with
    pile-up, or is it because of jet/tau/pixel code?
  • Still under investigation
  • At FNAL, I find a leak of 800 kB/event for Z???,
    and it is lumi dependent (600 kB/event for
    21033)
  • Nicola Amapane promises some results(fix?) this
    evening
  • Moreover, the DT reconstruction is known to have
    some deficiencies
  • So we have a two-fold plan
  • Continue with current Root Tree production at
    remote sites to get us a set of baseline results
    for HLT milestone
  • We already had Pt1, Pt10, and Pt4 (low lumi)
    done before CERN shut down.
  • Can Fermilab help? Run 1000 short jobs on PT4
    if leak not fixed
  • Push hard to get new HLT reconstruction code
    written to solve remaining problems in time for
    September deadline

5
Status of New Reconstruction Code
  • Stefano Lacaprara has authored a new version of
    DT reconstruction code
  • Corrects some naïve handling of the DT hits,
    incorrect pulls, new code organization,
  • Turns out DT reconstruction must know drift
    velocity to 1
  • This code has been examined (some bugs fixed) and
    cleaned up by Bart Van de Vyver (and also by
    Norbert and Nicola)
  • Aiming to get new results in August, hopefully
    with new reconstruction code

6
Muon Analysis at Fermilab
  • Request made to get the Muon Federations copied
    from CERN
  • Pt4 single muon sample at highest priority
  • Shafqat predicts copied by Monday
  • Pt1, Pt10, W, Z, and t-tbar to follow
  • Z??? (on-peak and above) already available
  • Root Trees will be copied as well, when available
  • US users thus have at least one local choice for
    an analysis center, in addition to CERN
  • Mechanism to obtain FNAL visitor id and computer
    accounts remotely works well (Thanks Hans)
  • When Pt4 sample ready, PRS/Muon group is
    interested in running large number of
    RootTreeMaker jobs at FNAL
  • INFN still trying to copy PT4 tracker digis
    (0.7TB)

7
Florida Prototype Tier-2 Center
  • Currently host the Z??? samples in Florida
  • but only for local accounts, I think, at the
    moment. Eventually should be accessible
    world-wide.
  • Limited by available disk space
  • Several TB of RAID ordered
  • Problematic analysis environment
  • Although the Production environment is working
    quite well with DAR distributions, the analysis
    environment (where users can compile and run
    jobs) is a little unstable
  • Some difficulties building code in ORCA6,
    perhaps traced to using RH6.2 vs. RH6.1 (loader
    version)
  • Need a more standard way to set up analysis
    environment
  • I think INFN also had some initial difficulties
    getting ORCA_6_2_0 installed and working
  • Should be solved once key people come back from
    vacation

8
Side Note
  • For better or worse, we are working with a
    complex set of software for CMS
  • Definitely not easy for newcomers to contribute
    to development or to debugging (or to create a
    DB)
  • Case in point how can a summer student plug a
    new L2 module into ORCA?
  • Many layers to ORCA software, difficult to
    navigate, little documentation of common
    classes
  • Sometimes counterintuitive rules must be
    followed
  • Complexity probably partly intrinsic to
    ORCA/COBRA, and partly due to inexperienced
    physicists working in this environment
  • That being the case, we MUST have professional
    tools for development and debugging
  • Must be able to debug and profile the code,
    check for memory leaks, corruption, etc.
  • This is standard for CDF, and reliability of
    production code has increased dramatically
  • Requires analysis workstations with enough
    memory to handle these tools
  • Should start defining a set of validation plots
    to show problems early in production
Write a Comment
User Comments (0)
About PowerShow.com