Selection of the
\(D^{+}\longrightarrow K^{-}K^{+}K^{+}\) candidates at the LHCb experiment
Sebastian Ordoñez
jsordonezs@unal.edu.co
Supervisors: Alberto C. dos Reis & Diego A. Milanés C.
6th Colombian Meeting on High Energy Physics
29th of November 2021
CERN Summer Student Programme 2021
Introduction
- My work: Selection of \(D^{+}\longrightarrow K^{-}K^{+}K^{+}\) candidates using a multivariate analysis (MVA). The study is based on a sample of \(pp\)-collision data, collected at a centre-of-mass energy of 13 TeV with the LHCb detector during the run 2.
- Decays of \(D\) mesons into three mesons exhibit a rich resonance structure at low energies, involving heavy-quark weak transitions, hadron formation and final-state interactions.
The \(D\) meson is the lightest known particle containing charm quarks, its mass is known to be 1869.62\(\pm\)0.20 MeV
- An additional motivation for researching three-body hadronic decays of heavy-flavoured mesons is the large and pure datasets available nowadays coming from the B-factories, LHCb, ...
Tree-level diagram for the \(D^{+}\longrightarrow K^{-}K^{+}K^{+}\) decay
Data Analysis
Pre-selection: Clone Tracks
Dalitz Plot before and after cuts on the SDV
Invariant-mass spectrum of the \(K^{-}K^{+}K^{+}\) candidates
Before MVA it is necesasry to reduce the high levels of background: clone tracks and combinatorial.
- We use the slope difference variables (SDV):
and
Data Analysis
Pre-selection: PID
Invariant-mass spectrum of the \(K^{-}K^{+}K^{+}\) candidates
Dalitz Plot before and after cuts on the PID variables
- We impose requirements on PID variables in order to remove combinatorial background
- The two main sources in this channel are \(\Lambda^{+}_{c}\) decays into \(K^{-}K^{+}p\) and \(K^{-}p\pi\) final states.
Data Analysis
Multi-variate Analysis (MVA): Training-Input variables
Background from data
Signal from Monte Carlo
Input variables for the MVA Algorithms
- The discriminating variables chosen for the MVA methods are only related to the \(D^{+}\) candidate.
- MVA algorithm uses a set of discriminating variables for known background and signal events, with the purpose of building a new variable which provides an optimal signal-background discrimination.
Data Analysis
Multi-variate Analysis (MVA): Training-Input variables
Data Analysis
Multi-variate Analysis (MVA): Training-Booking and testing
- The following algorithms were considered: Multi Layer Perceptron (MLP), Gaussian Boosted Decision Tree (BDTG), BDT, Decorrelated BDT (BDTD).
Classifier output distributions
- Cutting on the value of the MVA variables, it is possible to find the one which maximises a given figure of merit, providing high signal efficiency and at the same time a significant background rejection.
Data Analysis
Multi-variate Analysis (MVA): Training-Evaluation
ROC curve for all the classifiers
Area under the ROC curve
- ROC curve gives a suitable performance evaluation for each classifier
Data Analysis
Multi-variate Analysis (MVA): Application
- Once the training is completed, the next phase is the application of these results to an independent data set with unknown signal and background composition.
BDTG response
BDT response
Results
Significance curves
Result of applying cuts in the classifiers that showed the best signal-background discrimination performance
- Cut values which give the maximum signal significance
Results
\(D^{+}\longrightarrow K^{-}K^{+}K^{+}\) invariant-mass spectrum
This is the final invariant-mass distribution of the \(K^{-}K^{+}K^{+}\) candidates after applying the cut on the MLP classifier, the one with the best performance.
Final Dalitz Plot and Projections
Thank you!
Details and references can be found in CDS: https://cds.cern.ch/record/2781351
[ComHEP] Selection D->KKK
By Sebastian Ordoñez
[ComHEP] Selection D->KKK
- 352