a fast hardware tracker for the atlas trigger system

1
A Fast Hardware Tracker for the ATLAS Trigger System Mark Neubauer 1 , Laura Sartori 2 1 University of Illinois at Urbana-Champaign, USA – [email protected] 2 Istituto Nazionale di Fisica Nucleare, and Marie Curie Fellow OIF, Italy– [email protected] on behalf of the ATLAS TDAQ Collaboration Why Add a Hardware Tracker? Controlling trigger rates at high-energy hadron collider experiments in a way that maintains the physics capabilities is very challenging The trigger system should be flexible & robust, with sufficient redundancy and operating margin Providing high-quality track reconstruction by the start of LVL2 processing is an important element in achieving these goals Physics motivation for high luminosity ( large interaction pileup) LHC running: 3 rd -generation fermions will play an important role in LHC physics given the central role mass likely plays in electroweak symmetry breaking. Bottom quarks and τ-leptons produce very specific tracking signatures which can be exploited in triggering. At high luminosity, sophisticated algorithms run at LVL2 will be needed to suppress backgrounds. In the current system, the LVL2 farm is burdened with reconstructing tracks within Regions-of- Interest (ROIs) Global tracking will be important at high interaction pile-up (high luminosity) o Use of primary vertexing and subsequent charged lepton association for improved isolation to reject hadronic jet backgrounds o b-decay events with final state hadrons identified outside of ROIs o Additional combined triggers (e.g. lepton + track) o Track-based missing transverse energy & jets at trigger level, possibly combined wtih calorimeter information We propose to enable early rejection of background events and more LVL2 execution time for sophisticated algorithms by moving track reconstruction into a hardware system (FTK) with massively parallel processing that produces global track reconstruction with nearly offline resolution near the start of LVL2 processing Pipeline memories Derandomizers Readout drivers Readout buffers Full-event buffers Processor sub-farms Hardware Software The Current ATLAS Trigger System ~10-100 MB/s ~2 μs (fixed) ~1-10 ms (variabl e) ID: ~5,800 silicon pixel and strip detectors (SCT) and ~1000 modules of straw tubes FTK uses pixel and SCT hit information to infer charged particle tracks FTK processes 8 overlapping Φ-regions in parallel 6.2 m 2.1 m The ATLAS Inner Detector (ID) IEEE09, March25-31, Orlando, Florida, USA Proposed FTK in Trigger System FTK Functional Diagram & Architecture Pattern Recognition using Associative Memory (AM) Some Expected Performance The FTK system receives data from ReadOut Drivers (RODs) ROD output is duplicated by a dual output board Tracks reconstructed by the FTK processors are written into ReadOut Buffers (ROBs) for use at the beginning of LVL2 trigger processing FTK operates in parallel with the silicon tracker readout following each LVL1 trigger We are currently evaluating three different possible architectures which balance good tracking efficiency and an acceptable number of track fits per event: 1)Two pattern banks offset by ½ superstrip (SS = binned silicon hits) Require road to be found in both. Reduced effective SS width decreases # of combinations w/ small increase (x2) in patten bank size 2)Tree Search Processor (TSP) Smaller SS binary search after AM. Reduces SS size fewer combinations 3)Split Architecture First find SCT tracks, then fit with pixels (ala CDF SVT) Bingo analogy A large bank of pre- computed hit patterns is used for very fast tracking (memory speed) Linearized track fit using full-resolution silicon hits after pattern recognition Comparable FTK tracking resolution to offline tracking (Simulated muons with pt>1 GeV and |η|<2.5) Light quark rejection versus b-tagging efficiency comparing FTK and offline tracking Offline FTK

Upload: sook

Post on 06-Jan-2016

32 views

Category:

Documents


1 download

DESCRIPTION

A Fast Hardware Tracker for the ATLAS Trigger System. Mark Neubauer 1 , Laura Sartori 2 1 University of Illinois at Urbana-Champaign, USA – [email protected] 2 Istituto Nazionale di Fisica Nucleare, and Marie Curie Fellow OIF, Italy – [email protected] on behalf of the ATLAS TDAQ Collaboration. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: A Fast Hardware Tracker for the ATLAS Trigger System

A Fast Hardware Trackerfor the ATLAS Trigger System

A Fast Hardware Trackerfor the ATLAS Trigger System

Mark Neubauer1, Laura Sartori2

1University of Illinois at Urbana-Champaign, USA – [email protected] Nazionale di Fisica Nucleare, and Marie Curie Fellow OIF, Italy– [email protected]

on behalf of the ATLAS TDAQ CollaborationWhy Add a Hardware Tracker?

Controlling trigger rates at high-energy hadron collider experiments in a way that maintains the physics capabilities is very challenging

The trigger system should be flexible & robust, with sufficient redundancy and operating margin

Providing high-quality track reconstruction by the start of LVL2 processing is an important element in achieving these goals

Physics motivation for high luminosity ( large interaction pileup) LHC running:

• 3rd-generation fermions will play an important role in LHC physics given the central role mass likely plays in electroweak symmetry breaking. Bottom quarks and τ-leptons produce very specific tracking signatures which can be exploited in triggering. At high luminosity, sophisticated algorithms run at LVL2 will be needed to suppress backgrounds. In the current system, the LVL2 farm is burdened with reconstructing tracks within Regions-of-Interest (ROIs)

• Global tracking will be important at high interaction pile-up (high luminosity)o Use of primary vertexing and subsequent charged lepton association for improved isolation to reject hadronic jet backgroundso b-decay events with final state hadrons identified outside of ROIso Additional combined triggers (e.g. lepton + track)o Track-based missing transverse energy & jets at trigger level, possibly combined wtih calorimeter information

We propose to enable early rejection of background events and more LVL2 execution time for sophisticated algorithms by moving track reconstruction into a hardware system (FTK) with massively parallel processing that produces global track reconstruction with nearly offline resolution near the start of LVL2 processing

Pipeline memories

Derandomizers

Readout drivers

Readout buffers

Full-event buffersProcessor sub-farms

Har

dwar

eSo

ftw

are

The Current ATLAS Trigger System

~10-100 MB/s

~2 μs(fixed)

~1-10 ms(variable)

• ID: ~5,800 silicon pixel and strip detectors (SCT) and~1000 modules of straw tubes

• FTK uses pixel and SCT hit information to infer charged particle tracks

• FTK processes 8 overlapping Φ-regions in parallel

6.2 m

2.1

m

The ATLAS Inner Detector (ID)

IEEE09, March25-31, Orlando, Florida, USA

Proposed FTK in Trigger System

FTK Functional Diagram & Architecture

Pattern Recognition using Associative Memory (AM)

Some Expected Performance

• The FTK system receives data from ReadOut Drivers (RODs)

• ROD output is duplicated by a dual output board

• Tracks reconstructed by the FTK processors are written into ReadOut Buffers (ROBs) for use at the beginning of LVL2 trigger processing

• FTK operates in parallel with the silicon tracker readout following each LVL1 trigger

We are currently evaluating three different possible architectures which balance good tracking efficiency and an acceptable number of track fits per event:1)Two pattern banks offset by ½ superstrip (SS = binned silicon hits)• Require road to be found in both. Reduced effective SS width

decreases # of combinations w/ small increase (x2) in patten bank size

2)Tree Search Processor (TSP)• Smaller SS binary search after AM. Reduces SS size fewer

combinations3)Split Architecture• First find SCT tracks, then fit with pixels (ala CDF SVT)

Bingo analogyA large bank of pre-computed hit patterns is used for very fast tracking (memory speed)

Linearized track fit using full-resolutionsilicon hits after pattern recognition

Comparable FTK tracking resolution to offline tracking (Simulated muons with pt>1 GeV and |η|<2.5)

Light quark rejection versus b-tagging efficiency comparing FTK and offline tracking

OfflineFTK