Summary. This report outlines preliminary studies performed into the optimization of the ATLAS Level 2 electron-photon ( \( e \gamma \) ) trigger with respect to the 2010 data run . Electrons from \( \ztoee \) are a major source of high \( \pt \) electrons, an understanding of which is vital for many physics analyses. This optimization focuses on electrons from \( \ztoee \). The optimization procedure identifies signal (actual electrons) and background (mis-identified electrons) Level 2 trigger electrons using geometrical matching. Reducing the number of these background trigger electrons has obvious benefits with respect to data storage. A number of possible variables useful in discriminating this signal from background are identified. A log-likelihood ratio discriminant method is developed based on these variables and tested using Monte Carlo. The performance is cross checked using 2010 data. The log-likelihood discriminant method is compared with a cuts based method, based on individual variables and is shown to give improved results.
At design luminosity the expected interaction rate of LHC collisions, taking into account the multiple interactions for each of the nominal 40 million bunch crossings occuring per second, is of the order of 1 GHz. The ATLAS Trigger system [1] is designed to reduce the nominal 40 MHz bunch crossing rate within ATLAS to the hundreds of Hz that can physically be written to disk. In doing so, it must identify rare interesting physics events with high efficiency. For this purpose the trigger system is divided into three levels Level 1 (L1), Level 2 (L2) and the Event Filter (EF), each designed to reduce the trigger rate to a new lower frequency threshold. At each successive level, more and more complicated and robust procedures are used in the selection due to the higher latency with which the algorithms can be run. Interesting physics events are generally referred as those with leptons, photons, jets and/or missing energy in the final state. Lists of such final states are organised into trigger menus which are tables of triggers used to select the corrresponding final states at each level. The trigger is divided into slices depending on the final state a set of triggers is supposed to identify, i.e. the \( e \gamma \) slice contains final states with electrons and photons.
Level 1 uses reduced granularity information from the RPCs and TGCs for measuring high \( \pt \) muons. A set of predefined thresholds are applied which must be passed for a muon to be identified and triggered on. The low \( \pt \) thresholds require a \( \pt \) of between 6-9 \( \Gc \) whereas the high \( \pt \) thresholds require a \( \pt \) in the range 9-35 \( \Gc \). The thresholds must be passed in coincidence in 2(3) stations of chambers for the low(high) \( \pt \) thresholds, in order to reduce backgrounds from cosmic rays.
All the calorimeter sub-systems are used to identify electromagnetic clusters associated to electrons, photons, jets and \( \tau \) leptons, as well as events with large \( \etmiss \) or large total transverse energy. This identification procedure is performed by level 1 algorithms which use information from calorimeter trigger towers. Calorimeter trigger towers are formed from summing the energy in all cells within 0.1 \( \times \) 0.1 ( \( \eta\times\phi \) ) radial towers through all calorimeter sampling layers. To be identified, an object must pass a predefined \( \pt \) threshold. The readout of the data from the calorimeter and the muon spectrometer is performed by Read Out Drivers (ROD's).
The Central Trigger Processor (CTP) uses information from the calorimeter and muon triggers to establish, based on comparison with sets of predefined selections, whether each event should be considered further. If this criteria is met, the kinematic information of each identified object is passed to the read out buffers in the form of Regions of Interest (RoIs), where they are held until requested to be used by the Level 2 Trigger.
The electromagtic calorimeter is divided into 23 layers in lateral \( \eta \phi \) and longitudinal directions of the shower. At high energy most of the energy is collected in the 2nd layer, which has lateral granularity of 0.025x0.025. The 1st layer has a higher resolution in the \( \eta \) direction than \( \phi \), which allows signal from prompt photons to be separated from two nearby photons originating from \( \pi^{0} \) decays. The 3rd layer is used to correct highly energetic showers initiated by \( e \gamma \). A low resolution presampler calorimeter layer is located in front of three layers and is used to correct for energy loss in upstream material.
The \( e \gamma \) trigger is collectively the trigger used to identify energetic electrons and photons. Electrons are identified by an energy deposit in the electromagtic calorimeter which is associated with a track in the inner detector.
At L1, electromagetic clusters are triggered using calorimeter information from trigger towers (TT's), defined as \( \eta \) x \( \phi \) regions of the ATLAS detector of size 0.1 x 0.1. The energy in each TT is calculated by summing the cells in the electromagnetic and hadronic calorimeters, where all layers are considered except the fourth layer of the hadronic endcap and the barrel-endcap gap scintillators. Electromagnetic clusters are formed by identifying local maxima, using a sliding window algorithm based on a 4 x 4 group of TT for \( |\eta| \) < 2.5. A trigger is satisfied if the energy within the central 2 x 2 TT contains one pair of TT that lie next to one another with a combined energy greater than the threshold of the given trigger.
At L2, for electrons, information of the inner detector is used. Seeded by L1 cluster position, L2 implements a calorimeter reconstruction algorithm and a inner detector track reconstruction algorithm for electrons. A 0.4 x 0.4 region around the L1 seed position is used, reducing traffic and increasing the processing speed of L2. At L2 cluster size is fixed to a group of calorimeter cells in the 2nd layer, 3x7 for whole \( \eta \) range.
At EF, cluster reconstruction is done with a sliding window algorithm acting on towers containing energy summed in depth. After this seed finding step, fixed window clusters are built starting from the second layer using the same cluster size as in offline ( 3x7 for \( \eta \) < 1.5 and 5x5 for other). Energy weighted \( \eta \) and \( \phi \) positions are calculated in the 2nd layer of calorimeter and cluster building is then extended to other layers. Track reconstruction is done as in offline using a combination of tracking starting from hits in the silcion and TRT detectors.
Selection algorithms are applied to reconstructed clusters and tracks at L2 and EF to identify good \( e \gamma \) candidates and reject fakes predominantly from jets. Selections are based on shower shapes, tracking information and cluster-track matching quantities.
L1 cuts are based on four variables to get maximum signal selection and background rejection. The \( \et \) of an electromagnetic cluster is identified as the maximum \( \et \) present in a 2x1 or 1x2 tower region within the central 2x2 electromagnetic calorimeter window. This is used to make a cluster \( \et \) threshold cut and primarily used to reject QCD. Other cuts test for isolation of the electromagnetic cluster. Isolated electrons leave a narrow energy deposit in electromagnetic calorimeter only. More background can be rejected by getting rid of clusters with broad electromagnetic clusters or those depositing energy in the hadronic calorimeter. Isolation is measured by the energy in a ring of 12 electromagnetic trigger towers surrounding the 2x2 central cluster. Energy leakage into the hadronic calorimter is measured by the energy in the four hadronic trigger towers behind the electromagnetic central cluster and in the hadronic trigger towers.
L2 is guided by RoIs found at L1, i.e. is seeded by them. It refines the L1 decision, implementing calorimeter only or combined calorimeter and inner detector hypotheses. There are systematic differences between L2 and offline algorithms, due to differences in latency. L2 cluster reconstruction begins using the most energetic cell in the 2nd layer within the central region 0.2x0.2, whereas offline uses a sliding window to find the cluster seed. Calorimeter cell energies can be corrected offline for transient hardware issues but cannot be corrected in real time triggering, which also applies to L2 and EF. For track reconstruction at L2, fast pattern recognition is used by first determining the z position of the interaction point along the beam axis and then performing combinatorial tracking only inside groups of space points that point back to the determined position. At L2 cluster size is fixed to a group of calorimeter cells from the 2nd layer with three cells along \( \eta \) and seven cells along \( \phi \). Offline algorithms define the electrons cluster as 3x7 for \( \eta \) < 1.5 and 5x5 for \( \eta \) < 2.5. They also have different cluster sizes for unconverted and converted photons. The L2 calorimeter hypothesis applies cuts based on electromagnetic \( \et \) and \( \et \) in the first layer of the hadronic calorimeter, shower shape in the \( \eta \) plane in the 2nd sampling of the electromagnetic calorimeter and the ratio of energy maxima found within the shower in strips of the first sampling electromagnetic calorimeter. Due to energy dependance of cross sections, jet backgrounds are mainly rejected by the \( \et \) cut. Electromagnetic showers are more contained in \( \eta \) and deposit most energy in electromagnetic calorimeter, so cuts on \( \eta \) shape and hadronic leakage give further jet electron discrimination. The ratio of the difference between the 1st and 2nd energy maxima in the electromagnetic calorimeter, will take values close to unity for isolated electrons but not for diffuse hadronic showers (i.e. \( \pi \) decays). This is becase multiple particles in hadronic showers can create many local maxima that may carry a significant fraction of the shower energy. L2 requires a track identified in inner detectors be matched in \( \Delta\eta \) x \( \Delta\phi \) to calorimeter cluster. EF uses offline algorithms as much as possible and more precise calibration and allignment constants. Electron reconstruction takes place within RoIs seeded by L2.
Looser time contraints allow more sophisticated algorithms to be applied in the EF. As at L2, selection is based on calorimeter, inner detector track and track-cluster matching information. Offline and EF electrons hypothesis cuts are reffered to as isEM. At offline, selection algorithms are applied to reconstructed clusters and tracks found at L2 and EF in order to identify good \( e \gamma \) candidates and reject fakes from jets. Selections based on cluster shower shapes, tracking information and cluster track matching quantities are used. Three reference sets of selections (loose, medium and tight) are defined with increasing rejection power for electrons.
In early 2010 the LHC luminosity was 10 \( ^{27} \Lum \) and only L1 triggers were being used. The HLT was being operated in pass through to allow for validation. By Oct 2010 LHC luminosity was 2 x 10 \( ^{32} \Lum \) and HLT rejection was being applied. The \( e \gamma \) triggers of the HLT were commissioned using events triggered by the L1 calorimeter and min-bias triggers and the L1 triggers validated using randomly generated triggers. Each was also validated using good offline \( e \gamma \) candidates as references. Early data was dominated by photons from \( \pi \) decays (electrons from photon conversions and hadrons faking electrons and photons). After 400 \( \nb \) of data was collected, events with electrons from W and jpsi and later \( \ztoee \) were included. Since august 2010, all e \( \gamma \) trigger levels were in use, including active HLT selections and with increasing luminosity, tighter selections were applied.
Relevant detector components for e \( \gamma \) selection were required to be fully operational. Criteria was used to select genuine pp collisions and reject possible background from beam gas and beam wall events and events where cosmic muons depositied energy in calorimeters. Approximately 45 \( \pb \) of proton-proton collision data was collected under stable beam conditions during the Autumn 2010 reprocessing (7 TeV (152166-167844) and 900 GeV (141748-154471) runs). In period I this number was 18.3 \( \pb \) without the trigger applied. Data from Period I was used and was selected using the \( e \gamma \) stream using the EGamma good run list (data10_7TeV.periodAllYear_DetStatus-v21-pro05_CoolRunQuery-00-04-00_Eg_standard.xml) [3].
This analysis looked at the EF_10_15_medium trigger which was seeded at L1 by EM10 and used as a high \( \pt \) physics trigger in 2010 data taking. Collision candidates were required to have atleast one primary vertex with atleast three tracks. Events containing very localized high-energy calorimeter deposits not originating from the proton-proton collision, e.g. from cosmic-ray muons undergoing a hard bremsstrahlung, were removed. In addition, any event is rejected if the candidate electromagnetic cluster is located in any problematic region of the electromagnetic calorimeter (owing to hardware problems, the signal could not be read out from \( \approx \) 2 \( \% \) of the electromagnetic calorimeter cells).
Offline electrons are selected if they are within the fiducial region, \( |\eta| \) <2.47, excluding the transition between the barrel and endcap of the calorimeter (1.37 < \( |\eta| \) < 1.52). Offline electrons were required to have passed the RobustMedium requirement [5]. Corrections to electron energy scale and resolution recommended by the \( e \gamma \) working group [6] were implemented.
The cluster \( \eta \) - \( \phi \) position of the L2 trigger objects (L2 trigger electrons, subsequently referred to as trig-electrons) is associated with the offline position of the electron or photon cluster. The distance in \( \eta \) - \( \phi \) space needs to be less than 0.15 for the event to be considered further, i.e. \( \Delta R = \sqrt{\Delta\eta^{2} + \Delta\phi^{2}} \) < 0.15.
The remaining selection cuts are as follows and were used to form the signal and background trig-electron samples for Monte Carlo and data respectively.
The following L2 e \( \gamma \) trigger discriminating variables were identified:
A Log-Likelihood Ratio (LLR) is used to separate signal from background. The distributions of the identified discriminating variables for signal and background are normalised to unity separately to obtain probability distribution functions (PDF). Using these the joint likelihoods \( L_{s} = \prod_{i} P^{i}_{s}(x_{j}) \) and \( L_{b} = \prod_{i} P^{i}_{b}(x_{j}) \) are constructed where \( P^{i}_{s}(x_{j}) \) and \( P^{i}_{b}(x_{j}) \) are the probabilities of the ith discriminating variable to have a value \( x_{j} \) in signal and background events respectively. The LLR is then defined as in Eqn. \eqref{eqn_LikelihoodRatio} $$ \begin{equation} LLR = log(\frac{L_{s}}{L_{b}}) \label{eqn_LikelihoodRatio} \end{equation} $$ 70 \( \% \) of the data in the \( \ztoee \) sample and in the real data sample were used for establishing the LLR cut. The likelihood ratios for these \( \ztoee \) and data samples for signal and background trig-electrons are shown in Figure 1 respectively.
Figure 1: LLRs for signal (black) and background (red -dotted) for a) \( \ztoee \) Monte Carlo and b) data.

The discrimination of the likelihood is improved by using un-correlated variables. The table below shows the correlation between variables used as input to the LLR, for \( \ztoee \) and data, signal and background.
| Signal | Background | |||||||||
| \( \ztoee \) | ||||||||||
| weta2 | fracs1 | \( \chi^{2} \) | \( \Delta\eta \) | \( \et/\pt \) | weta2 | fracs1 | \( \chi^{2} \) | \( \Delta\eta \) | \( \et/\pt \) | |
| weta2 | 1 | 0.11 | 0.09 | -0.19 | 0.14 | 1 | 0.1 | 0.03 | -0.05 | -0.16 |
| fracs1 | 0.11 | 1 | 0 | 0 | -0.05 | 0.1 | 1 | -0.03 | 0.05 | -0.13 |
| \( \chi^{2} \) | 0.09 | 0 | 1 | 0.05 | 0.13 | 0.03 | -0.03 | 1 | -0.1 | -0.06 |
| \( \Delta\eta \) | -0.19 | 0 | 0.05 | 1 | 0.03 | -0.05 | 0.05 | -0.1 | 1 | 0.2 |
| \( \et/\pt \) | 0.14 | -0.05 | 0.13 | 0.03 | 1 | -0.16 | -0.13 | -0.06 | 0.2 | 1 |
| Data | ||||||||||
| weta2 | 1 | 0.18 | 0.08 | -0.03 | 0.14 | 1 | 0.22 | 0.02 | -0.08 | -0.12 |
| fracs1 | 0.18 | 1 | 0.01 | -0.11 | -0.02 | 0.22 | 1 | 0.03 | 0.03 | -0.14 |
| \( \chi^{2} \) | 0.08 | 0.01 | 1 | 0.06 | 0.11 | 0.02 | 0.03 | 1 | -0.11 | -0.07 |
| \( \Delta\eta \) | -0.03 | -0.11 | 0.06 | 1 | 0.08 | -0.08 | 0.03 | -0.11 | 1 | 0.11 |
| \( \et/\pt \) | 0.14 | -0.02 | 0.11 | 0.08 | 1 | -0.12 | -0.14 | -0.07 | 0.11 | 1 |
Generally the PDF's will be dependent on the \( \eta \) and \( \pt \) of the object (in this case trig-electrons). Therefore ideally the PDF's should be obtained for each bin for each of the disriminating variables discussed. Due to a lack of statistics no attempt was made to perform the optimzation binned in \( \pt \) and \( \eta \) , although this is expected to be beneficial since the shower shape variables are generally \( \eta \) dependent. The variation of signal efficiency measured with respect to offline electrons is shown in Figure 2 as a function of cut on the LLR for \( \ztoee \) and data.
Figure 2: Plots showing a) efficiency of electron reconstruction relative to offline and b) percentage reduction in signal and background trig-electrons achieved for a given LLR cut for \( \ztoee \) Monte Carlo and data.

The results for 3 representative LLR cuts, -12, -4 and 4, are shown in table below in terms of efficiency with respect to offline and percentage reduction in the number of signal and background trig-electrons. In order to maintain 100 \( \% \) recontruction efficiency with respect to offline, a LLR cut of -12 enables the background trig-electrons to be reduced by 9 \( \% \) in the \( \ztoee \) sample and by 7 \( \% \) in the data. Far greater reductions in background trig-electrons of over 50 \( \% \) for both \( \ztoee \) and data can be achieved at the cost of around a 0.3 \( \% \) reduction in the number of signal trig-electrons, giving an efficiency with resepct to offline of over 99.7 \( \% \).
| L Cut | Sample | Efficiency (wrt offline) | \( \% \) fewer signal | \( \% \) fewer background |
| -12 | \( \ztoee \) | 100 \( \mypm \) 0.4 | 0 \( \mypm \) 0.1 | -9.1 \( \mypm \) 0.1 |
| Data | 100 \( \mypm \) 0.3 | 0 \( \mypm \) 0.1 | -7.3 \( \mypm \) 0.1 | |
| -4 | \( \ztoee \) | 99.8 \( \mypm \) 0.4 | -0.24 \( \mypm \) 0.02 | -54.5 \( \mypm \) 0.3 |
| Data | 99.7 \( \mypm \) 0.3 | -0.32 \( \mypm \) 0.01 | -62.0 \( \mypm \) 0.2 | |
| 4 | \( \ztoee \) | 43.7 \( \pm \) 0.3 | -56.4 \( \mypm \) 0.3 | -99.0 \( \mypm \) 0.4 |
| Data | 10.9 \( \mypm \) 0.1 | -89.2 \( \mypm \) 0.3 | -99.3 \( \mypm \) 0.3 |
Applying the -12 or -4 LLR cut is shown to have negligible impact to the reconstruction efficiency with respect to offline as a function of \( \pt \), which is shown in Figures 3a) and 3b) for \( \ztoee \) and data respectively. The same plots as a function of \( \eta \) are shown in Figures 3c) and 3d) for \( \ztoee \) and data respectively.
Figure 3: Efficiency of reconstruction with respect to offline electrons (for a),c) \( \ztoee \) and b),d) data) as a function of a,b) electron \( \et \) and c,d) \( \eta \) for no LLR cut and LLR cuts of -12, -4 and 4.

In the following the performance of the LLR method is compared to a cut based method, in which the cuts on the discriminating variables outlined in the section Discriminating Variables are optimized by means of the TMVA Genetic algorithm [7]. The samples of the datasets not used for establishing the LLR cut are used.
A comparison of the two methods is shown in Figures 4 for \( \ztoee \) and data respectively. Signal efficiency is defined as the ratio of the surviving number of signal trig-electrons divided by the original number and the background rejection is 1 minus the background efficiency which is defined in the same way as the signal efficiency.
Figure 4: Signal trig-electron efficiency vs background trig-electron rejection for the LLR and cuts methods for a) \( \ztoee \) Monte Carlo and b) data.

The methods appear to be fairly consistent over the range of the signal efficiencies explored. However, requiring all signal trig-electrons survive the selection, the performance of the LLR method is more performant than the cut based method, as shown in the table below. Requiring all trig-electrons survive, the LLR based method is able to reject 8 \( \% \) of background trig-electrons in the \( \ztoee \) Monte Carlo whereas this number is around \( \frac{1}{2} \% \) for the cut based method. In the data a similar trend is observed, with the LLR method reducing the number of background trig-electrons by 6 \( \% \) and the cut based method by less than \( \frac{1}{2} \% \).
| Method | Sample | \( \% \) fewer background |
| Likelihood | \( \ztoee \) | -8.4 \( \mypm \) 0.2 |
| Data | -6.2 \( \mypm \) 0.1 | |
| Cuts | \( \ztoee \) | -0.56 \( \mypm \) 0.04 |
| Data | -0.47 \( \mypm \) 0.03 |
1: This work has not been through the approval process within ATLAS and therefore it's results are not official.