• Non ci sono risultati.

UNIVERSITÀ DEGLI STUDI DI PADOVA Dipartimento di Fisica e Astronomia “Galileo Galilei” Corso di Laurea in Fisica Tesi di Laurea

N/A
N/A
Protected

Academic year: 2021

Condividi "UNIVERSITÀ DEGLI STUDI DI PADOVA Dipartimento di Fisica e Astronomia “Galileo Galilei” Corso di Laurea in Fisica Tesi di Laurea"

Copied!
29
0
0

Testo completo

(1)

UNIVERSITÀ DEGLI STUDI DI PADOVA

Dipartimento di Fisica e Astronomia “Galileo Galilei”

Corso di Laurea in Fisica

Tesi di Laurea

Study of the decay

Λ

b0

→ Λ

c∗+

π

π

+

π

Relatore

Laureando

Prof. Gabriele Simi

Pietro Argenton

Correlatore

Dr.ssa Anna Lupato

(2)
(3)

Contents

1 Introduction 1

2 The LHCb experiment 3

2.1 The Large Habron Collider . . . 3

2.2 The LHCb Detector . . . 4

2.2.1 The VELO Detector . . . 5

2.2.2 Tracking Detectors . . . 5

2.2.3 RICH Detectors . . . 5

3 The Λb0 → Λc∗+π−π+π− decay 7 3.1 Topology of signal . . . 7

3.1.1 Relevant quantities . . . 8

4 The data sample analisys 9 4.1 Data sample . . . 9

4.2 Selection of signal and background sample . . . 9

4.3 Comparison between variable for signal and background samples . . . 10

5 Boosted Decision Tree 13 5.1 Multivariate analysis . . . 13 5.2 Decision Tree . . . 15 6 Results 17 6.1 BDT . . . 17 6.2 Significance . . . 20 7 Prospects 23

(4)
(5)

Chapter 1

Introduction

It is known that the Standard Model (SM) is not the ultimate description of elementary particle dynamics. Finding and identifying hints of New Physics (NP) in the quark flavour dynamics still represents a great challenge at the colliders. In the SM the couplings of the electroweak gauge bosons

Z0and W± to the leptons are independent of the lepton flavour and then, the branching fractions of

e, τ and µ can differ only by phase and helicity-suppressed contributions; this is the so called Lepton

Flavor Universality (LFU). The LFU is enforced in the SM by construction and therefore any violation of lepton universality would be a clear sign of physics beyond the SM. For this reason nowadays many semi-leptonic decay processes are studied in the colliders.

One of the methods to test LFU at colliders is to compare the semileptonic rate of a hadron

de-caying to different leptons in the final state R = B(H→lνlX)

B(H→l0ν0

lX). The BaBar, BELLE and LHCb

experiments compared in particular the semi-leptonic decays of a B meson into muons and tau

R(D∗−) = B(B0→D∗−τ+ντ)

B(B0→D∗−µ+ν

µ) [1] and they found a significant difference with respect to the Standard

Model prediction. A similar ratio can be measured by LHCb using the semi-leptonic decays of

B-baryons such as R(Λc∗+) = B(Λb

0→Λ

c∗+τ−ν¯τ)

B(Λb0→Λc∗+µ−ν¯µ)

1. In this measurement one of the dominant

contri-butions to the error on R comes from the contribution of the background from decays of the type

Λb0 → Λc∗+Ds−. These dacays are very similar to the Λb0 → Λc∗+π−π+π− decays, another possible

background. This is why in this thesis the Λb0 → Λc∗+π−π+π− decay has been investigated. In

par-ticular we study an optimized selection of the Λb0 → Λc∗+π−π+π− signal and compute its efficiency,

using a MultiVariate Analysis (MVA), in this case a Boosted Decision Tree (BDT). Moreover this decay has never been observed before and therefore, can potentially lead to the first measurement of the branching fraction of this decay. The structure of the thesis will be the following: a first introduc-tion of the LHCb experiment and the detectors, a descripintroduc-tion of the studied decay, the data sample with the selection of the signal and the background, the data analysis and in the end the results with conclusions.

(6)
(7)

Chapter 2

The LHCb experiment

This chapter describes the LHCb, which has been designed to systematically study the decays of hadrons containing b and c quarks (and the decays of tau leptons) . This experiment therefore allows to check the consistency of the SM through precision measurements of the sides and angles of the Cabibbo-Kobayashi- Maskawa triangle, and to search for NP in decays that are rare, or forbidden, in the SM.

2.1

The Large Habron Collider

The LHC is a circular collider of 26.66 Km circumference colliding two proton beams rotating in opposite directions. It is designed to collide protons onto protons at a center of mass energy of 14 TeV

at an unprecedented luminosity of 1034cm−2s−1 . In 2011 the centre of mass energy was kept at 7 TeV,

whereas in 2012 it was kept at 8 GeV. Bunches of 1011 protons each are obtained from hydrogen gas

and are firstly accelerated to 50 MeV with a linear accelerator called LINAC2. They are then passed to the Proton Synchrotron Booster where their energy is increased to 1.4 GeV. Following this, they are injected into the Proton Synchrotron, accelerated to 25 GeV and transmitted to the Super Proton Synchrotron (SPS). The SPS accelerates the bunches to 450 GeV and finally injects them clockwise and counter-clockwise into the LHC ring. A total of about seven minutes is needed to fill both LHC rings. When the rings are filled the LHC further accelerates the protons in a ramp phase. The four main detectors at the LHC: ATLAS, CMS, ALICE and LHCb, are located at the four collision points. Figure 2.1 shows a schematic view of the SPS, the LHC ring and the detectors.

(8)

2.2. THE LHCB DETECTOR CHAPTER 2. THE LHCB EXPERIMENT

2.2

The LHCb Detector

The LHCb detector is a forward arm spectrometer centered around the LHC beam pipe, 100 m un-derground. The aim of the LHCb experiment is to record the decay of particles containing b(c) and anti-b(c) quarks, known as B(charmed) mesons, and also to record τ leptons. At the interaction point, a proton-proton deep inelastic scattering occurs, producing a highly boosted virtual gluon and break-ing up the incombreak-ing protons at a primary vertex. Rather than flybreak-ing out in all directions, B mesons formed by the colliding proton beams (and the particles they decay into) stay close to the line of the beam pipe, as we can note in Figure 2.2. This explains the choice of detector design and geometry. LHCb measures particles which appear within its angular acceptance of 10 mrad to 250 mrad vertically, and 10 mrad to 300 mrad horizontally. Approximately, one third of B hadrons lie within the LHCb acceptance. In terms of pseudorapidity η = ln(tan(θ/2)), where θ is the polar angle with respect to the beam axis, the acceptance is 1.8 < η < 4.9.

Figure 2.2: Angular distribution of b-¯b pairs at LHCb. The axis show the polar angle θ of b and ¯b with respect

to the beam axis.

The LHCb detector is shown in Figure 2.3. Starting form the left side, the VErtex LOcator (VELO) is built around the proton interaction region. Directly after it, a Ring Imaging CHerenkov detector, RICH-1, is located. Then, there are the dipole magnet, the tracking system composed by stations TT, T1 and T2 and another Cherenkov detector, RICH-2. Afterwards the electromagnetic (ECAL) and hadronic (HCAL) calorimeters are present and finally there is a muon detector made of five stations (M1 - M5).

LHCb provides precise vertexing resolution and a precise momentum resolution. Moreover, it it char-acterized by a high trigger efficiency for the reconstruction of B mesons decays and background sup-pression, and an excellent particle identification. In this data analysis, the most important physical quantities, as we will see, are detected by VELO [4], the tracking system [6] and RICH (1 and 2) [5]. For this reason I am going to describe these detectors.

(9)

CHAPTER 2. THE LHCB EXPERIMENT 2.2. THE LHCB DETECTOR LHCb uses a right-handed coordinate system with y pointing upwards, x horizontal and pointing to the outside of the LHC-ring and the z-axis along the beam. The proton-proton collisions take place around z=0, located at the left side in the Figure 2.3.

2.2.1 The VELO Detector

The VErtex LOcator (VELO) is placed around the interaction point and measures particle trajectories close to the interaction region. The high resolution of the coordinate measurements of the tracks allows the reconstruction and separation of the primary interaction vertex from secondary decay vertexes of bottom and charmed mesons. These are essential for time dependent measurements and to determine the impact parameters of the decay products with respect to the primary vertex.

The Velo detector consists of silicon modules, 300µm thick, placed perpendicular to the beam. Charged particles produced by proton collisions traverse the silicon and generate electron-hole pairs; these are sensed using specific electronics.

The resolution on impact parameter (IP), e.g the distance of closest approach between a track and the reconstructed primary vertex, is about 20 µm for high transverse momentum tracks. The resolution on the decay length, the distance between the interaction point and the secondary vertex is within the range of 200-370 µm, depending on the decay channel.

2.2.2 Tracking Detectors

The tracking detector, combined with the dipolar magnet, provides efficient reconstruction of charged tracks and precise measurement of their momentum and direction. The latter is also needed to recon-struct Cherenkov rings in the RICH detectors.

Two different detector technologies are employed. The silicon tracker, which is placed close to the beam pipe, uses silicon microstrip detectors to detect passing particles. Ionization in the depleted region of the silicon sensors induces electric signals on the detector strips, which indicates the path of the original particle. The outer tracker is situated further from the beam pipe and is made up of thousands of gas-filled straw (drift) tubes. Whenever a charged particle passes through, it ionizes the gas molecules, producing electrons. The position of the track is found by timing how long the electrons take to reach an anode wire situated in the center of each tube.

The momentum resolution from the tracking system is about δp

p = 0.5%for momenta of 20 GeV/c and

rising to 1.0% at 200 GeV/c.

2.2.3 RICH Detectors

The LHCb detector uses two Ring Imaging CHerenkov (RICH) detectors. RICH-1 is placed directly downstream of the VELO and before the main tracking system. RICH-2 is positioned after the tracking stations and in front of the calorimeters. The RICH system is schematically drawn in Figure 4.1. Measurements of Cherenkov angles are used, together with momentum measurements by the main tracking system, to perform particle identification of charged tracks. RICH detectors are based on the Cherenkov effect: when a charge particle traverses a medium with a velocity β higher than the speed of light in that medium a cone of electromagnetic radiation, Cherenkov radiation, is emitted along the

trajectory. Such radiation is emitted coherently at an angle θC with respect to the direction of the

motion such as:

cosθC =

1 βn

where n is the refractive index of the medium. Particles can therefore be identified when their momen-tum and the opening angle of the Cherenkov radiation cone are known. The Ring Imaging CHerenkov

(RICH) detectors of LHCb measure θC by focusing the emitted light with spherical mirrors on a plane

of photodetectors (hybrid photodetectors). The photons emitted along the trajectory of the traversing

(10)

2.2. THE LHCB DETECTOR CHAPTER 2. THE LHCB EXPERIMENT The LHCb RICH detectors have to provide identification of charged particles over momentum range from 1 GeV/c up 150 GeV/c. In order to achieve this, several radiator materials with different refractive indices are used.

RICH-1 performs particle identification for momentum less than 60 GeV/c, over the full LHCb angular acceptance.

The RICH-2 detector is used to perform particle identification of high momentum tracks. This requires a lower refractive index, but a longer path length for the particles in order to manage to collect sufficient

Cherenkov light, since the number of photons emitted is proportional to Nγ= sin2θC. It has a reduced

angular acceptance because high momentum tracks are produced at small angles.

RICH detectors are fundamental in order to provides the pion-kaon separation, which reduces strongly the combinatory background.

(a) RICH 1 (b) RICH 2

(11)

Chapter 3

The

Λ

b

0

→ Λ

c

∗+

π

π

+

π

decay

As stated before the decay Λb0 → Λc∗+π−π+π− is a significant source of background in the search

for violation of LFU in the R(Λc∗+) measurement. In order to measure the BF of this background we

study in this thesis an optimized selection of the Λb0→ Λc∗+π−π+π−signal and compute its efficiency.

3.1

Topology of signal

The decay that was used in this thesis is:

Λb0→ Λc∗+π−π+π−

where

• Λc∗+= Λc+(2625).

• Λc∗+→ Λc+π−π+.

• Λc+→ pK−π+.

In principle two Λc∗+ resonances can be used, as we can see form the distribution of mas difference

between the Λc∗+ and the Λc+ baryons in Figure 4.1b. In this study we consider only the Λc+(2625).

In order to separate the signal decay from combinatoric background we exploit the excellent capabilities of the LHCb detector concerning momentum, impact parameter resolution and particles identification. This decay channel presents two important feature that cannot be used in semi-leptonic decays due to the presence of neutrinos: first it is kinematically closed i.e. all the particles in the final state can

be reconstructed, second the reconstructed Λb0 momentum points back to the Primary Vertex (PV).

Another advantage of this channel is that, thanks to resonance Λc∗+ which decays hadronically, the

Λb0 vertex can be found with a very good quality. Finally the reconstruction of the Λc∗+ → Λc+π−π+

decay is almost free from down-feed contamination.

The Λb0 baryons lifetime is large enough to allow them to fly on average about 1 cm before decaying.

Also Λc+ decay displaced with respect to Λb0 vertex, unlike Λc+(2625) resonance. These peculiarities

help us to identify the searched channel: large impact parameters (IP) of the reconstructed tracks and displaced vertices with good qualities.

(12)

3.1. TOPOLOGY OF SIGNAL CHAPTER 3. THE ΛB0 → ΛC∗+π−π+π− DECAY

3.1.1 Relevant quantities

The data used in this work have been collected with the LHCb detector in pp collisions after various processing phases. The online event selection is performed by a trigger system, which consists of a hardware stage, based on information from the detectors, followed by a software stage, which applies a full event reconstruction. After the software stage and the offline event reconstruction data are finally obtained with a selection of the interesting quantities and are available as root files; this work used roughly 450.000 events with about 60 variables each identifying all the main characteristics of the events.

It is useful to define some relevant quantities and variables used in the analysis. All quantities are calculated in the laboratory frame, and in general, can be applied to each particle in the decay.

• Primary vertex (PV): the point of the space where the primary proton-proton interaction is reconstructed.

• Secondary decay vertex (SV): the point in the space where the decay of long-lived particles occurs. The displacement vector ~d, defined as the spatial vector from primary to second vertex, is equal to ~d = γ~βct = (~p/m)ct , where c is the speed of light, m the mass, ~p is momentum and t the proper time decay of the decay particle.

• Impact Parameter (IP): given a vertex ~v , the impact parameter is the vector formed by the nearest point of a track to ~v.

• DIRA: the cosine of angle between the momentum of the particle ~p and the displacement vector,

DIRA = ~p · ~d/|~p|| ~d|For a fully reconstructed particle its total momentum tends to be aligned to

the displacement vector resulting into a DIRA close to unit (like in our case), differently for one partially reconstructed, like semileptonic decay.

• χ2

vertex and χ2track: the tracks and the vertexes are recostructed by minimizing the χ2 of the hints

in the detectors. Small χ2

vertex and χ2trackensures an agreement between the track and the vertex

model and the reality.

The large amount of variables that define this decay process make it difficult to separate Background from Signal because both populate complicated regions in a large multidimensional space. The AI approach is mandatory to find where the main part of the Signal lies. To obtain these results an AI technique is trained with a mixture of real data and Monte-Carlo simulations and, once fully trained, the algorithm can be applied to additional data to enhance the Signal component.

To form the training sample in this work the Signal is taken from a Monte-Carlo simulation while the Background is taken from real data.

(13)

Chapter 4

The data sample analisys

This chapter describes the first step of the analysis of the data sample, in particular the identification of the variables that discriminate between signal and background.

4.1

Data sample

We used the data collected by the LHCb detector in pp collisions at 14TeV of energy in the center of mass. Generic hadronic events were selected at the trigger level by using impact parameter selection and the topology of the events.

After the trigger level selection the complete event reconstruction is performed and a second loose selection is applied to the events. This selection, called "stripping", requires the events to contain

good tracks consistent with the decay of Λc+→ pK−π+ and 3 additional pions forming a good vertex.

Finally requires that the reconstructed Λb0 momentum points to the primary vertex associated to the

Λb0. After the stripping 7 million events are selected. In order to further reduce the number of events

and for practical purposes additional quality cuts are applied to the sample Table 4.1 reducing the number of selected events to 450000.

2260 < m(Λc+) < 2310M eV

Probability of the Λ∗+

c π− identified as a π>0.4

Probability of the Λ∗+

c π+ identified as a π>0.4

Probability of the other 3 π identified as a π>0.4

Λb0 DIRA PV>0.99999

Λb0 χ2vertex<15

Λc+ χ2vertex<10

Λc∗+ χ2vertex<10 Table 4.1: Pre-selection cuts

4.2

Selection of signal and background sample

In order to obtain the distribution of the discriminating variables for the signal we use a sample about 6000 signal events simulated with GEANT4 [4] and reconstructed with the same algorithms and pre-selection used to reconstruct real data. For simulated events we also require that reconstructed candidates are correctly matched to signal decays.

In order to obtain the same distributions for background events we use our data sample and apply a selection to exclude signal events.

(14)

4.3. COMPARISON BETWEEN VARIABLE FOR SIGNAL

AND BACKGROUND SAMPLES CHAPTER 4. THE DATA SAMPLE ANALISYS

The first cut is selected on the mass of the Λb0 which we know the exact value m(Λb0) = 5619.51 ±

0.23M eV. In Figure 4.1a is represented the distrubution of this variable and it is easily to see that

there is an important background signal. We decided to select a background cut on the m(Λb0)equal

to m(Λb0) < 5550M eV or m(Λb0) > 5680M eV.

The second cut is selected on the distribution of the ∆m = m(Λc∗+) − m(Λ+c). The exact values for

this two variables are m(Λc+) = 2286.46 ± 0.14M eV and m(Λc+(2625)) = 2628.11 ± 0.19M ev. Also

in this case (Figure 4.1b) the plot of the ∆m distribution has an important background signal. As we

can see there is also a peak near 306 MeV which correspond to the ∆m with the m(Λc+(2595)). So we

set another background cut, this time on the ∆m, equal to ∆m > 360MeV . Doing this we select the background sample: all the data with this characteristics are cataloged as background. In the Table 4.2 the cuts that define the background sample are summarized.

m(Λb0) < 5550M eV or m(Λb0) > 5680M eV

∆m > 360M eV

Table 4.2: Background selection cuts

htemp Entries 472750 Mean 4378 Std Dev 677.1 3000 4000 5000 6000 7000 8000 Lambda_b0_MM (MeV/c^2) 0 2000 4000 6000 8000 10000 12000 14000 16000 18000 20000 entries htemp Entries 472750 Mean 4378 Std Dev 677.1 Lambda_b0_MM

(a) Plot of the mass of the Λb0for the data sample

htemp Entries 472750 Mean 448.1 Std Dev 65.05 250 300 350 400 450 500 550 DeltaM (MeV/c^2) 0 2000 4000 6000 8000 10000 entries htemp Entries 472750 Mean 448.1 Std Dev 65.05 DeltaM (b) Plot of the ∆m

Figure 4.1: Plot of the variable used for the background cut

4.3

Comparison between variable for signal

and background samples

The next step is to find some variables that discriminate the signal from the background. This plots on Figure 4.2 represent the normalized distributions of the most relevant variables for the signal and background of data sample and for montecarlo, in the region of interest. The signal of data sample is defined using the cuts in the Table 4.3. Since we want to get a signal enriched sample, we apply on our data sample the inverse cuts used for the selection which exclude signal events.

5550 < m(Λb0) < 5680M eV

∆m < 360M eV

Table 4.3: Signal selection cuts

We select discriminating variables by requiring that their distribution on MC and the corresponding distribution on data is different. In addition we check that the distribution on MC is consistent with the signal enhanced data sample.

(15)

CHAPTER 4. THE DATA SAMPLE ANALISYS

4.3. COMPARISON BETWEEN VARIABLE FOR SIGNAL AND BACKGROUND SAMPLES

(a) Λ+

c z decay lenght over error (b) Λ+c decay time [ps]

(c) Probability that a p is identified as a π (d) Pair mass of the two π of the Λc∗+

(e) m(Λb0) [MeV] (f) ∆m [MeV]

Figure 4.2: Comparison of some variables Plots

For example as we see from the plot 4.2c the variable plotted is not a good discriminant variable. Instead the variable plotted in 4.2d is a very good discriminant variable.

(16)

4.3. COMPARISON BETWEEN VARIABLE FOR SIGNAL

AND BACKGROUND SAMPLES CHAPTER 4. THE DATA SAMPLE ANALISYS

Variable

m(Λ+c)

Probability of a Λ+

c p identified as a p

Λ+c z decay lenght over error

Λ+∗c z decay lenght over error

Λ+c decay time Λ+∗ c decay time Λ0b decay time Probability of Λ+ c χ2vertex Probability of Λ+∗ c χ2vertex Λ0b DIRA of PV

Pair mass of π0 and π1

Pair mass of π1 and π2

Table 4.4: Discriminant variables

The probability of a Λ+

c proton is identified as a proton is calculated by Neural Net from likelihood

ratios from the data of RICH detectors. The MC and signal distributions have a peak near 1 greater than the background one, as we can see in the Figure 6.1(where I report all the variables used for training the BDT).

Λ+c z decay length over error, and also Λ+∗c with the appropriate substitutions, is equal to

dz q σΛ0 bEN DV ERT EX 2 Λ+c EN DV ERT EX

2 where dz = −zΛ0bEN DV ERT EX+ zΛ+cEN DV ERT EX. The background

distribution of this variable stays closer to zero respect MC and signal distributions.

Λ+c decay time is equal to d·m(Λ+c)

c·p(Λ+c) where d =

q

d2

x+ d2y+ d2z, c is the speed of light, m is the mass and

p is the momentum. Also in this variable distribution the background stays closer to zero respect MC and signal, moreover the background distribution is symmetric to 0 while the MC and the signal ones are mostly displaced in the positive region.

The probabilities of χ2 for the vertexes of the Λ+

c and Λ+∗c are others important discriminant variables.

The background distribution has an exponential increase near 0 unlike MC and signal distribution, which have a constant trend as we expect.

As it was said before, the Λ0

(17)

Chapter 5

Boosted Decision Tree

This chapter introduces and briefly describes the important features of multivariate analysis (MVA), in particular boosted decision tree (BDT).

5.1

Multivariate analysis

An often faced problem is to predict the answer to a question based on different input variables. There are two different types of problems:

• Classification problems, which predict only a binary response like:

is it going to rain today? → yes/no; what is the measured data? → signal/background.

• Regression problems, which predict an exact value as an answer like: what will be the temperature

tomorrow? → 17◦C 21C . . .

BDT used in this thesis are instruments for classification problems. MVA is useful first of all because

allows to combine several discriminating variables into one final discriminator Rd→ R, because

pro-vides a better separation between signal and background (in our case) respect of using single variables, and also because correlations between variables become visible. A lot of available methods can be used: BDT, Neural Networks, Likelihood Functions ...

This simple example shows how MVA has a better separation power than simply cutting on the single variables:

(a) Cut on variable 0 (b) Cut on variable 1

(c) Single variable cuts (d) Better cut on MVA

(18)

5.1. MULTIVARIATE ANALYSIS CHAPTER 5. BOOSTED DECISION TREE An important step when aiming for a MVA approach is to train the selected algorithm. This can be achieved in two main ways: by using a supervised training or by using an unsupervised one; while the former aims to teach the program to recognize a specific case the second tries to find general unknown patterns. In particle physics the common approach is to use supervised training due to the possibility to use Monte-Carlo simulation to generate reliable sets of Signal. To start a supervised training a training set is needed. The training set is a classified sample of data which to be effective must be explanatory of all the possible situations the program could encounter. A MVA program commonly classifies data by associating to each event a continuous variable, usually in the range between -1 and 1, used to separate the initial sample in a Signal enriched region and a Background enriched region: a training sample must be provided for both the Signal and the Background categories and the same must be done for the testing sample. In Figure 6.3 we can see an example of classified events distributions for Background and Signal

Figure 5.2: Example of a Classified Signal and Background events distributions for a BDT method.

The training sample, before the actual training start, is split into two parts: one actually used for the training and one used as test sample, to calculate the score of the algorithm. The train of the program is repeated with multiple iterations; each time the algorithm tries to predict the outcome of the training sample, it recognize its errors, and tries to reconfigure itself to correct them. After each cycle the program is applied on the test sample and a score is calculated: if the score reach a fixed amount the train ends, if not the process is repeated another time; in case of over many iterations the program is unable to improve anymore the program can recognize the state as the best possible outcome and exit successfully the training anyway.

(19)

CHAPTER 5. BOOSTED DECISION TREE 5.2. DECISION TREE

5.2

Decision Tree

A Decision Tree (DT) is a consecutive set of questions (nodes), each of them has only two possible answers. Each question depends on the formerly given answers and the final verdict (leaf) is reached after a given maximum number of nodes, set previously.

Figure 5.3: Example of a Decision Tree

First of all a DT needs to be trained on a data set which already provides the outcome (e.g simulation data set with signal and background processes). The choice of nodes criterion is based on maximizing the separation gain between nodes, which is defined as: gain(parent cell) gain(daughter cell 1) -gain(daughter cell 2). Gain can be computed in different ways, a common used one is the Gini Index: gain(cell)= p(1 − p) where p is the purity.

An advantage of DTs, as we can see in Figure 5.3, is that are easy to understand and interpret, and also they have a very fast training. The only important disadvantage is that a single tree is not very strong. To ward off this problem we use Random Forests. A Random Forests is an ensemble method that combines different trees and in which the final output is determined by the majority vote of all trees.

One possible method to train Random Forest is the boosting; the most common used method is the AdaBoost (Adapting Boosting). This method enhances weights of missclassified events and reduces weight of correctly classified ones after each training, so that future trees learn those better.

A serious risk in a BDT training is the overtraining phenomena. If the training is done for an excessive number of iterations or the training sample is too small there is the possibility that the algorithm reaches an ad-hoc configuration. This case is particularly dangerous cause it could go undetected by the algorithm and classified as the only good result to be found. But when used on real data the program is unable to do proper classifications. The main method to avoid overtraining is done via the test sample. The first reason to have a test sample in fact is to test the program on slightly different data that have the same proprieties of the training sample hence avoiding excessively specific configurations and this is done by comparing the distribution in the BDT output variable for Signal/Background for the training/testing sample. Both Signal and Background distributions from training are taken and confronted with the respective ones from the test. Because they are random parts of the same sample there is no reason to have different distributions.

(20)
(21)

Chapter 6

Results

In this chapter the results of the BDT and the final value for the selection efficiency are presented.

6.1

BDT

2 4 6 8 10 12 14 16 lbDecTime [ps] 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.405 ps / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.2)%

Input variable: lbDecTime

0.2 0.4 0.6 0.8 1 lcDecVertChi2 [Probability] 0 0.5 1 1.5 2 2.5 3 0.0251 Probability / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: lcDecVertChi2

0.2 0.4 0.6 0.8 1 lcstarDecVertChi2 [Probability] 0 0.5 1 1.5 2 2.5 3 0.0251 Probability / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: lcstarDecVertChi2

0.99999 0.9999920.9999940.9999960.999998 1 Lambda_b0_DIRA_OWNPV 0 200 400 600 800 1000 1200 3 10 × 2.57e-07 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: Lambda_b0_DIRA_OWNPV

500 1000 1500 2000 2500 3000 pair0_3pi [MeV/c^2] 0 0.0002 0.0004 0.0006 0.0008 0.001 0.0012 0.0014 0.0016 0.0018 0.002 0.0022 79.8 MeV/c^2 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: pair0_3pi

500 1000 1500 2000 2500 3000 pair1_3pi [MeV/c^2] 0 0.0002 0.0004 0.0006 0.0008 0.001 0.0012 0.0014 0.0016 0.0018 0.002 81.2 MeV/c^2 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: pair1_3pi

(a) 2260 2270 2280 2290 2300 2310 Lambda_c_MM [MeV/c^2] 0 0.02 0.04 0.06 0.08 0.1 1.28 MeV/c^2 / (1/N) dN Signal Background U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: Lambda_c_MM

0.6 0.650.70.75 0.80.85 0.90.95 1 lc_p_ProbNNp [ ] 0 5 10 15 20 25 30 35 40 45 0.0102 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: lc_p_ProbNNp

20 − −10 0 10 20 30 lcZDecLSigma [MeV/c^2] 0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2 0.22 1.47 MeV/c^2 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.1)%

Input variable: lcZDecLSigma

3 − −2 −1 0 1 2 3 lcstarZDecLSigma 0 0.1 0.2 0.3 0.4 0.5 0.189 / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

Input variable: lcstarZDecLSigma

2 − −1.5 −1 −0.5 0 0.5 1 1.5 2 lcDecTime [ps] 0 0.5 1 1.5 2 2.5 0.11 ps / (1/N) dN U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.1)%

Input variable: lcDecTime

2 − −1 0 1 2 lcstarDecTime [ps] 0 0.5 1 1.5 2 2.5 0.134 ps / (1/N) dN U/O-flow (S,B): (0.0, 0.1)% / (0.0, 0.0)%

Input variable: lcstarDecTime

(b)

Figure 6.1: Plot of the variables used for training the BDT, in red the background sample and in the blu the signal sample

(22)

6.1. BDT CHAPTER 6. RESULTS As we said before, to training a BDT we need a signal and a background sample. The background sample is taken from the data with the cuts defined on the Table 4.2. For the signal sample we used the MC sample as it is said at the section 4.2. The number of the background events are 403158 while the signal events are 1490.

In the Figure 6.1 above, are represented all the variable used for training the BDT, and described previously. As we can see they all discriminate the signal from the background, some more than others, significantly. An other important fact to consider when we use a MVA method is the internal correlations between variables. A MVA is able to take into account these correlations, thus reducing the complexity of the problem. Here in Figure 6.2 is presented the Correlation Matrix for the signal

(the one for the background coincides). The only two variables correlated are the Λ+

c z decay length

over error and the Λ+

c decay time (both for Λ+c and for Λ+∗c ). This because the particles in this decay

travel for the most along the z-axis and so the decay time is strongly correlated with the z decay length.

100 − 80 − 60 − 40 − 20 − 0 20 40 60 80 100

Lambda_c_MMlc_p_ProbNNplcZDecLSigmalcstarZDecLSigmalcDecTimelcstarDecTimelbDecTimelcDecVertChi2lcstarDecVertChi2Lambda_b0_DIRA_OWNPVpair0_3pipair1_3pi Lambda_c_MM lc_p_ProbNNp lcZDecLSigma lcstarZDecLSigma lcDecTime lcstarDecTime lbDecTime lcDecVertChi2 lcstarDecVertChi2 Lambda_b0_DIRA_OWNPV pair0_3pi pair1_3pi

Correlation Matrix (signal)

100 -1 15 -1 20 -3 4 -1 -2 -1 100 4 4 3 4 7 1 -7 -6 15 4 100 1 85 2 -1 1 4 -4 -7 -1 4 1 100 4 84 2 2 2 -1 -8 20 3 85 4 100 5 -2 6 4 4 4 2 84 5 100 3 1 -1 2 -3 -7 -3 7 2 3 100 -3 1 32 3 -2 -1 2 -2 1 -3 100 4 5 2 4 1 6 -1 1 100 -3 -2 -1 1 4 2 2 32 4 -3 100 -3 -2 -7 -4 -1 4 -3 3 5 -2 100 14 -6 -7 -8 4 -7 -2 2 -3 14 100 Linear correlation coefficients in %

Figure 6.2: Correlation Matrix for the variables used

For this analysis we trained various method, 4 types of BDT, the differences are in method of boosting, a neural network and a linear discrimination analysis with Fisher discriminant. The Figure 6.4 represents for all the classifiers the power of separation and an overtraining check. Another method to compare the different classifiers is the ROC curve: the trend of the background rejection as a function of the signal efficiency. With the same signal efficiency the classifiers have a different background rejection. As we can see in Figure 6.3 the best classifier respect to this comparing method is the BTDF, a BDT trained with the usage of fisher discriminant for node splitting.

Also in the plots of Figure 6.4 proves that the BTDF method is the better one: there is not overtraining and the power of separation seems to be very good. For these reasons, for the final analysis, we use the BDTF classifier.

(23)

CHAPTER 6. RESULTS 6.1. BDT 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Signal efficiency 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Background rejection MVA Method: BDTF BDTD BDTG Fisher MLPBNN BDTB

Background rejection versus Signal efficiency

Figure 6.3: Plot of the ROC curves

1 − −0.80.60.40.2 0 0.2 0.4 0.6 BDTF response 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0.002 ( 0.04)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: BDTF

(a) 0.6 − −0.40.2 0 0.2 BDTD response 0 1 2 3 4 5 6 7 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0 (0.658)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: BDTD

(b) 1 − −0.5 0 0.5 1 BDTB response 0 2 4 6 8 10 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0.015 (0.534)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: BDTB

(c) 0.8 − −0.60.40.2 0 0.2 0.4 BDTG response 0 1 2 3 4 5 6 7 8 9 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0.004 (0.165)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: BDTG

(d) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 MLPBNN response 0 1 2 3 4 5 6 7 8 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0.159 (0.022)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: MLPBNN

(e) 5 − −4321 0 1 2 3 4 Fisher response 0 0.2 0.4 0.6 0.8 1 1.2 1.4 dx / (1/N) dN

Signal (test sample) Background (test sample)

Signal (training sample) Background (training sample)

Kolmogorov-Smirnov test: signal (background) probability = 0.679 (0.499)

U/O-flow (S,B): (0.0, 0.0)% / (0.0, 0.0)%

TMVA overtraining check for classifier: Fisher

(f)

(24)

6.2. SIGNIFICANCE CHAPTER 6. RESULTS

6.2

Significance

The Significance is a useful figure of merit defined as S = ss

ss+bb where s is the number of signal

events, b is the number of the background events, and  are the efficiencies. Significance is a useful quantity because maximizing that is equivalent to minimizing the relative error on the signal, and so on the branching ratio. Once we define the significance we can set the cut on the single variable of the BDTF. To estimate S we have to know n and b.

In Figure 6.5 it is represented the plot of the m(Λb

0) with some fits, before of course the BDTF cut.

In order to count the number of background and signal events we fit the m(Λb

0) data distribution

with a PDF built as: P (m(Λb

0)) = s · BW (m(Λb0); m0; Γ) + b · Expo(m(Λb0); λ) where BW is a

Breit-Wigner distribution centered at the value of m(Λb

0) m0 and defined also by a width Γ, and Expo is

an exponential distribution for the background eλx defined by the attenuation coefficient λ. The fit

values and the n and b are reported in the Table 6.1.

Name Value Error

b 4703 16

n 9623 18

λ -0.0047 0.0004

m0 5619.0 0.3

width 31.4 0.8

Table 6.1: Values of the fit

5540 5560 5580 5600 5620 5640 5660 5680 5700 2 ) MeV/c 0 b Λ m( 0 50 100 150 200 250 300 350 400 450 Events / ( 1.8 ) Figure 6.5: Fit of m(Λb0)

(25)

CHAPTER 6. RESULTS 6.2. SIGNIFICANCE Now we can define the value of the BDT cut. As we can see in the Figure 6.6 for the value of s and b of the Table 6.1 the value of the optimal BDT cut is -0.1143.

1

− −0.80.60.40.2 0 0.2 0.4 0.6

Cut value applied on BDTF output

0 0.2 0.4 0.6 0.8 1 Efficiency (Purity) Signal efficiency Background efficiency Signal purity Signal efficiency*purity S+B S/

For 9623 signal and 4703 background is S+B events the maximum S/ 90.7582 when cutting at -0.1143

Cut efficiencies and optimal cut value

0 10 20 30 40 50 60 70 80 90 Significance

Figure 6.6: Plot of the significance for the BDTF method

Applying the BDT cut on the MC sample we can compute the signal efficiency s = 0.94.

We can also apply the BDT cut on the data sample and compare the distribution of the m(Λb

0)before

and after the cut. The result is shown in Figure 6.7

5540 5560 5580 5600 5620 5640 5660 5680 ] 2 ) [MeV/c 0 b Λ m( 0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 (a) Normalized 5540 5560 5580 5600 5620 5640 5660 5680 ] 2 ) [MeV/c 0 b Λ m( 0 50 100 150 200 250 300 350 400 (b) Not normalized

(26)
(27)

Chapter 7

Prospects

This final chapter describes the future prospects of this work. First of all, to improve the BDT performance, we need to remove the pre-selection on the data sample. Second we have to include the

analysis also for the other resonance of the Λ+

c, Λ+c(2595).

Then finally we have to apply the BDT selection also on a control sample; in this case we can use the

normalization channel of the already known Λb0 → Λc+π−π+π− decay [2]. Doing this we can compute

the final branching ratio of our decay channel, without knowing the exact luminosity of LHCb. In order to account for small differences between the signal and normalization channel we also need to complete the computation of the efficiencies adding the trigger one, the stripping one, the acceptance one and the correction of the particles identification.

(28)
(29)

Bibliography

[1] R. Aaij et al. (LHCb Collaboration) (2018), Test of lepton flavor universality by the measurement

of the B0→ D∗−τ+ν

τ branching fraction using three-prong τ decays.

[2] T. Aaltonen, for the CDF Collaboration (2011), Measurement of the branching fraction Λb0 →

Λc+π−π+π− at CDF.

[3] Geant4 collaboration (2006), Geant4 developments and applications. IEEE Trans. Nucl. Sci. ,

(53-270).

[4] LHCb collaboration(2001), Lhcb velo: Technical design report. CERN-LHCC [5] LHCb collaboration(2001), Lhcb rich: Technical design report. CERN-LHCC

[6] LHCb collaboration(2003), Lhcb inner tracker: Technical design report. CERN-LHCC [7] A. Hocker et al. (2007), TMVA - Toolkit for Multivariate Data Analysis

Riferimenti

Documenti correlati

For example, starting from a given energy density profile, one should be able to compute numerically the radial pressure and from that obtain an equation of state which would give

Questi termini finiti devono però anche soddisfare le richieste che abbiamo fatto e in particolare dovranno avere supporto sulla linea di universo della particella, essere simmetrici

Now, let us analyze the three fine structure corrections to the hydrogen Hamiltonian obtained in Eq.( 2.47 ), in order to explain their physical meaning and discuss if and how

Un elemento del gruppo delle trecce può essere rappresentato graficamente come l’insieme delle linee d’universo delle N particelle che partono dalle posizione iniziali delle

Una volta estrapolata la formulazione hamiltoniana che meglio caratterizza il gas in questione, procediamo col discutere la forma delle soluzioni delle equazioni del moto che

Dopo aver introdotto il modello di Rouse per cui il centro di massa di un polimero immerso in un bagno termico diffonde con coefficiente di diffusione inversamente proporzionale

Equazioni di bilancio delle forze In questo capitolo ricaveremo, dalle equazioni MHD dell’equilibrio, un sistema di equazioni differenziali le cui incognite saranno le

Per Van Fraassen non è soddisfacente la lettura che la scuola di Copenhagen dà della misura: innanzitutto la teoria dovrebbe essere autosussistente, dunque capace di descrivere