• Nenhum resultado encontrado

Identification of LHC beam loss mechanism : a deterministic treatment of loss patterns

N/A
N/A
Protected

Academic year: 2023

Share "Identification of LHC beam loss mechanism : a deterministic treatment of loss patterns"

Copied!
199
0
0

Texto

Depending on the number of particles lost, the coils of the magnets may become normally conductive and/or be damaged. The BLM system's detectors are mainly ionization chambers located outside the cryostats.

La protection du LHC

La distance entre les mors est variable, contrôlée dans la limite de 5µm et exprimée en unités de l'écart type du rayon nominal, appelé σ. Les protons déviés par les collimateurs principaux seront absorbés par d'autres collimateurs plus ouverts.

Moniteurs de perte de faisceau

Dans le rôle d'un collimateur, la position des mâchoires dans les plans transversaux (horizontal, vertical ou diagonal) joue un rôle important. Les protons absorbés créent des gerbes de particules secondaires qui seront détectées par les moniteurs de perte de faisceau.

Figure 2: Exemple de l’installation de trois BLMs (chambres d’ionisation) sur le cˆ ot´e interne d’un quadrupole du LHC
Figure 2: Exemple de l’installation de trois BLMs (chambres d’ionisation) sur le cˆ ot´e interne d’un quadrupole du LHC

Int´ erˆ et du travail doctoral

Principe de la d´ ecomposition vectorielle

Le but de la décomposition vectorielle est de décomposer X~ en une combinaison linéaire de vecteurs (~vi), c'est-à-dire pour résoudre l'équation vectorielle : 1) Deux techniques sont utilisées pour cela : à partir d'une succession de projections vectorielles (~vi) (non orthogonales) dans l'espace vectoriel (processus de Gram-Schmidt, G-S) ; et avec des opérations matricielles. Puisque < m, la matrice des vecteurs (~vi) (notée M) n'est pas carrée et son inversion n'est pas simple.

Figure 3: Exemple simplifi´e de d´ecomposition vectorielle en 3 dimensions: n = 2, m = 3
Figure 3: Exemple simplifi´e de d´ecomposition vectorielle en 3 dimensions: n = 2, m = 3

Mise en place de la d´ ecomposition vectorielle

Les étoiles bleues correspondent à la valeur d’écart type normalisé pour les écrans sélectionnés. Des profils de perte pourraient être recréés en attribuant une valeur nulle aux moniteurs associés au faisceau non considéré.

Figure 6: Exemple de s´election des profils de pertes apr`es normalisation, pour les moniteurs du point 7, et le sc´enario correspondant au faisceau 2 horizontal
Figure 6: Exemple de s´election des profils de pertes apr`es normalisation, pour les moniteurs du point 7, et le sc´enario correspondant au faisceau 2 horizontal

Evolution temporelle

D´ ecomposition spatiale

Conclusion & Appendices

CERN

LHC PROTECTION

  • The Large Hadron Collider
  • Magnets
  • Collimators
  • Conclusion
  • The monitors

They are aperture limitations: the parts of the LHC closest to the beam. The cutoff threshold values ​​(losses above which the beam would be removed from the machine) can be tuned.

Figure 1.2: Schematics of the layout of the LHC, with the purpose of the different insertion regions
Figure 1.2: Schematics of the layout of the LHC, with the purpose of the different insertion regions

THE BLM SYSTEM

  • Time Structure: Running Sums
  • Electronics
  • Thresholds
  • Justification of the doctoral work
  • Principle

The simulation of the distribution of secondary particles according to the location of the loss is shown in fig. The input current iin(t) corresponds to the charge coming from the ionization chamber; the output is the frequency fout.

Figure 2.2: Example of installation of three BLMs (ionisation chambers) on the internal side of the LHC, on a quadrupole
Figure 2.2: Example of installation of three BLMs (ionisation chambers) on the internal side of the LHC, on a quadrupole

PRINCIPLE OF VECTOR DECOMPOSITION

  • Singular Value Decomposition (SVD)
  • Gram-Schmidt process
  • MICADO
  • Recomposition and error
  • Conclusion
  • Creation of the default vectors

However, there are not enough vectors (~vi) (n < m): they do not form a basis of the m-vector space. A negative factor in a vector decomposition indicates that the corresponding vector cannot contribute to the reconstruction of the vector X. The contribution is the projection of a vector onto another, calculated by a scalar product.

The error in the recomposition is estimated by calculating the norm of the difference between the original vector and its recomposition.

Figure 3.1: Structure of the matrices in the decomposition. M is the original matrix, made of the vectors (~v i ); U and W are the matrices of the left and right singular vectors respectively
Figure 3.1: Structure of the matrices in the decomposition. M is the original matrix, made of the vectors (~v i ); U and W are the matrices of the left and right singular vectors respectively

IMPLEMENTATION OF VECTOR DECOMPOSITION

Choice of the list of BLMs

In this work, each coordinate of the loss vectors corresponds to the signal of one BLM. The standard deviation quantifies the variation of the signal from one loss scenario to another. It was checked in advance whether the signal of one BLM does not change much between loss maps of the same scenario (cf. § 4.1.1).

However, both the standard deviation and the difference between minimum and maximum signals are proportional to the mean value of the signal.

Figure 4.3: Distributions of all BLMs of point 7 when considering all loss scenarios at the same time, for the loss maps of 2010
Figure 4.3: Distributions of all BLMs of point 7 when considering all loss scenarios at the same time, for the loss maps of 2010

Final loss map selection

Loss cards must be sorted by another criterion: the order in which they were measured. As described in § 4.1.1, loss maps are measurements of losses during the crossing of the resonance of the melody for one of the transverse planes. The difference can already be assessed visually: there is more variation within the values ​​of one monitor when the loss maps were measured at another location (see Figure 4.5 below).

It shows the beam 2 horizontal loss maps that were measured before the beam 2 vertical loss maps.

Figure 4.4: All 2011 loss maps (at the time of the study) for Beam 1 horizontal, normalised.
Figure 4.4: All 2011 loss maps (at the time of the study) for Beam 1 horizontal, normalised.

Adding more cases: longitudinal scenarios and TCTs

In the case of the 2011 transverse loss charts, it was shown that only the loss charts that were executed first should be kept. Most of the loss maps are mixed: the level of the losses is the same for both beams (cf. fig. 4.7). In order to compare the quality of the loss maps with each other, all the relative standard deviations were collected on the same plot (cf. fig. 4.14).

All longitudinal loss maps present higher RSD than transversal ones, which is a consequence of the various problems presented in §.

Figure 4.6: Values of signals in all monitors of IP7 for the B2H loss maps of 2011, and values of the normalised standard deviation of the signal of the selected BLMs (in blue), for all loss maps of the loss scenario B2H taken before the loss maps for B2V
Figure 4.6: Values of signals in all monitors of IP7 for the B2H loss maps of 2011, and values of the normalised standard deviation of the signal of the selected BLMs (in blue), for all loss maps of the loss scenario B2H taken before the loss maps for B2V

Validation of results: Centers of Mass

The values ​​used for the centers of mass are the combinations of the signals on the four primary collimators (cf. Each of the 4 plots corresponds to the loss scenario (known before the calculation) given in the title of each plot. The results of the centers of mass perfectly match the type of each individual vector.

Each of the 4 plots corresponds to a scenario (known before the calculation) indicated in the title.

Table 4.6: List of primary collimators of IP7 used in the Centers of Mass.
Table 4.6: List of primary collimators of IP7 used in the Centers of Mass.

Choice of the algorithm

Various corrections were considered, such as subtracting the signal from the vertical collimator from the horizontal one, or stretching the variation interval. The implementation of the first method showed that a factor of approximately twice the vertical signal had to be subtracted to get a correct variation interval; but it created a pole for the function (h, v) 7→ (h−α ·v, v), i.e. the arbitrary stretch of the variational interval, which corresponds to the precedent method with no pole and a correction factor of higher value, led to the problem that some values ​​of the mass centers are higher than one, in cases where the crosstalk between the two monitors is overestimated.

In the end, the choice was made to keep the centers of mass uncorrected, knowing that the horizontal/vertical would vary between -1 and -0.5.

Evaluation of the correctness of a decomposition

Averages are lower for SVD (better recomposition), and correct and incorrect are close to each other. The averages are lower with SVD (better recomposition), and regular and irregular are still too close to each other to allow good separation. For incorrect parses, 26 entries are above average (79% true negatives) and 7 below (21% false positives).

For the correct decomposition, there are 38 entries below the mean (54% . true positives) and 33 above (46% false negatives). for the incorrect breakdowns, 24 above (73% true negatives) and 9 below (27% false positives).

Figure 4.18: Distributions of correct (left) and incorrect (right) decompositions for G-S (top) and SVD (bottom) of every loss map of 2010 on every vector set (of the same loss maps)
Figure 4.18: Distributions of correct (left) and incorrect (right) decompositions for G-S (top) and SVD (bottom) of every loss map of 2010 on every vector set (of the same loss maps)

Application to LHC data

The first value calculated here is the rate of difference (subtraction) between the normalized vector representing losses in one second and the normalized vector representing losses in the previous second (cf. fig. 4.21):. Calculating the Euclidean norm of the vector is a good way to estimate the magnitude of the overall losses at the LHC (cf. fig. 4.21). In addition, the vector rate is several orders of magnitude higher than the "shape change" (cf. fig. 4.21).

So the scalar product is dominated by the effect of the norm and evolves like the norm.

Figure 4.21: Top: values of the Euclidean norm of the current loss profile (purple), the scalar product of two consecutive loss profiles (green), the sum of all losses in the current loss profile (red).
Figure 4.21: Top: values of the Euclidean norm of the current loss profile (purple), the scalar product of two consecutive loss profiles (green), the sum of all losses in the current loss profile (red).

Default Loss Profile

The work presented in this chapter focused on studying the time evolution of the signals from the Beam Loss Monitors during nominal operation. The point is to learn how the losses evolve; check where significant losses occur and which parts of the LHC are free of losses. It will also provide a better understanding of the behavior of the vectors and what information can be extracted from their time evolution.

TIME EVOLUTION

  • Individual evolution of the signal of a BLM
  • Evolution of the overall losses
  • Conclusion
  • Results of decomposition

It will describe the total loss in a similar way to the norm of the loss vector. It was performed by calculating the distribution of the difference between the signal of one BLM in one second and the previous second (cf. Figure 5.6). For the degree of precision of the loss measurement, the variations of the signal are Gaussian.

The norm of the difference also decreases in the first 3 hours; this is shown in more detail in fig.

Figure 5.2: Top left: Norm of the current loss profile. Top right: sum of all losses at current second
Figure 5.2: Top left: Norm of the current loss profile. Top right: sum of all losses at current second

SPATIAL DECOMPOSITION

Examples of decomposition versus time

The error onxj, written asσxj, is the error on the BLM size. This is known: it is the standard deviation of each coordinate of the reference vector over all the loss maps that make up this vector. They are dominated by the error on the loss maps and are at least 3 orders of magnitude lower than the values ​​of the factors, so they are not shown in the results.

All errors are less than 3 orders of magnitude or more of the values ​​of the factors.

Figure 6.2: Top: SVD versus time, for transversal and longitudinal loss scenarios and the error on decomposition, for the stable beam of the 15 th of October 2011
Figure 6.2: Top: SVD versus time, for transversal and longitudinal loss scenarios and the error on decomposition, for the stable beam of the 15 th of October 2011

Validation of results

At the beginning of the filling, the total derivative is dominated by the derivative of Beam 1: the center of mass is positive. When the intensity derivative is dominated by beam 1 (CoM closer to 1), the decomposition is dominated by factors of beam 1 (CoM closer to 1). Bottom: correlation between the derivative of the intensity in beam 1 (Y-axis) and the SVD factor associated with B1H (X-axis).

For higher values ​​of the factor, many more protons were lost (higher negative values ​​of derivative).

Figure 6.8: Left: short period of evolution of the intensity in the LHC as measured by the Fast Beam Current Transformers, without smoothing, from the 3 rd of August 2011
Figure 6.8: Left: short period of evolution of the intensity in the LHC as measured by the Fast Beam Current Transformers, without smoothing, from the 3 rd of August 2011

Conclusion

But during scraping the losses increase greatly; and the shape of the vector is then closer to the standard and the loss is better compounded. This shows that the error is an effective way to evaluate the quality of the recomposition. The reason for this is visible in the plot of the beam intensity versus time (cf. Fig. 6.15 center).

New reference vectors could be constructed for other scenarios, such as breaking the collimator hierarchy.

Introduction

DATA ACCESS

Types of required data and corresponding databases

Loss thresholds are different from BLM to BLM, depending on the element they protect; they also vary depending on the 32 beam energy levels and LHC cycle phases. The actual thresholds used at the same time are stored in the databases, under a variable name similar to that of the loss: BLM_EXPERT_NAME:THRESH_RSXX where. The structure of the Layout database is that of a standard SQL database: several tables hold all related information in different columns.

Two identically behaved databases are used to store the values ​​coming from the LHC's beam instruments.

Database Access Techniques

But once created, they are fixed: they do not follow the evolution of the variables in the database. Since not all variables are logged on the same time stamp (cf. § A.2.3), it was chosen to link each value to its time stamp, as this has already been done in the database. To link the list of values ​​to the name of the variable, a mapping is performed thanks to a layout object.

The fact that all lists are the same length is guaranteed by the database structure.

Figure A.1: Data flow in the toolbox, from the user to the databases and back. A timestamp is passed to the python object, which can start the java command line  applica-tion and request logging and measurement data
Figure A.1: Data flow in the toolbox, from the user to the databases and back. A timestamp is passed to the python object, which can start the java command line applica-tion and request logging and measurement data

Display

This is used as the X coordinate, where the Y coordinate is the second element of the 2-tuple. All this and creating the TLegend is done by the display.rainbowplot function. The format of the data recorded by this plot is as follows: the dual (pair) of base values ​​is: (DCUM, value of the loss at the displayed timestamp).

The plot time is an important value displayed in the title of the plot.

Figure A.5: Example of “plot time” object: signal of BLMs versus time. One graph and one colour is associated to each BLM
Figure A.5: Example of “plot time” object: signal of BLMs versus time. One graph and one colour is associated to each BLM

The wrapping module: analysis

This list (ordered by DCUM) is downloaded from one of the layout databases called MTF. The name of the graph describes the criteria validated by the BLMs on the list, e.g. The display of the signal from the BPMs requires a similar process to the BLMs.

The mapping of the value with the name and position must be done after the data query.

Figure A.9: Global structure of “checkdict” object. The name of the graph describes the criteria validated by the BLMs in the list, e.g
Figure A.9: Global structure of “checkdict” object. The name of the graph describes the criteria validated by the BLMs in the list, e.g

Conclusion

CROSS-CHECKS

Any monitor, sensitive to a secondary shower of mixed particles originating from protons lost in the LHC. Short parts of the accelerator where the majority of the beam operations take place (collision, acceleration, cleaning, measurements..) by opposition to the arcs, where the beam is carried. Predefined integration interval used as a sliding integration window by the BLM electronics to evaluate the duration of a beam loss (cf.

Zamantzas, "The Real-Time Data Analysis and Decision System for Particle Flux Detection in LHC Accelerator at CERN", CERN-thesis-2006-037.

Figure B.2: Factors and error of the decompositions of all loss maps of 2011 on the average set of selected loss maps used for decomposition of real data
Figure B.2: Factors and error of the decompositions of all loss maps of 2011 on the average set of selected loss maps used for decomposition of real data

Imagem

Figure 1: Sch´ema du syst`eme de collimation en plusieurs ´etages au LHC. Les materiaux sont:
Figure 2: Exemple de l’installation de trois BLMs (chambres d’ionisation) sur le cˆ ot´e interne d’un quadrupole du LHC
Figure 6: Exemple de s´election des profils de pertes apr`es normalisation, pour les moniteurs du point 7, et le sc´enario correspondant au faisceau 2 horizontal
Figure 8: Exemple de repr´esentation des r´esultats de la d´ecomposition (SVD) en fonction du temps
+7

Referências

Documentos relacionados

First the classification algorithm builds a predictive model from the training data set and then measure the accuracy of the model by using test data set.This work proposes a