Studying Same-Sign Top Pair Production in Flavor Changing Scalar Models at the HL-LHC

We investigate the potential of the HL-LHC for discovering new physics effects via the same-sign top pair signatures. We focus on the semi leptonic (electron and muon) decay of the top quarks and study the reach for a simplified model approach where top quark flavor changing could occur through a neutral scalar exchange. A relatively smaller background contribution and clean signature are the advantages of the leptonic decay mode of the same-sign $W$ bosons in the same-sign production processes of top quark pairs. Assuming the FCNC between top quark, up type quark and scalar boson from the new physics interactions the branchings could be excluded of the order ${\cal O}(10^{-4})$. We use angular observables of the same-sign lepton pairs and the top quark kinematics in the process which provide the possibility of separation of new physics signal from the SM backgrounds using machine learning teqniques. We find that the same-sign top quark pair production is quite capable of testing the top-Higgs FCNCs at the HL-LHC.


I. INTRODUCTION
Among all fundamental fermions in the standard model (SM), top quark has the largest mass and causes the most serious hierarchy, and plays an essential role in the metastability of the Higgs boson potential [1]. Top quark is also the last corner stone of the family structure of the SM with a huge mass gap with other members of quark content of the SM. It is the most sensitive particle for TeV scale physics in SM with Higgs boson, therefore researching the interactions of top quark is a crucial part of BSM physics.
The flavor changing neutral currents (FCNCs) among the up or down sector quarks are not present at leadingorder in both Yukawa and gauge interactions within the standard model (SM) framework. However, extremely small FCNC couplings could be generated from looplevel diagrams which are strongly suppressed due to the Glashow-Iliopoulos-Maiani (GIM) mechanism [2] and it is one of the unique characteristics of the SM. Besides it sets a new horizon for new researches.
The essence and importance of GIM mechanism's veto and studying FCNC interactions lies in the decision of dropping or keeping the FCNC preventing unique feature of SM model to new physics. FCNC searches will deduce its ultimate fate without any doubt. If one can show its possibility, that would be a great progress at BSM researches.
To be specific about the branching ratios models are expected we would like to summarize them at a Table I. Table I. Expected FCNC branching ratios from models [16]. Production of two positively charged top quarks via uu → tt resulting in an excess of same-sign lepton pairs have already been searched by the ATLAS [17]. Systematic uncertainties for the main backgrounds, including charge misidentification, fake/non-prompt leptons, etc. are presented about 28% , 33% and 30% for the ee ,eµ ,µµ channels, respectively. The limit on the cross section leads to a limit of BR(t → uH) < 0.01. Another search for flavour changing neutral current processes in top quark decays have been presented again by the ATLAS Collaboration from proton-proton collisions at the LHC with √ s = 13 TeV [18]. The observed (expected) upper limits are set on the t → cH branching ratio of 1.1 × 10 −3 (8.3 × 10 −4 ) and on the t → uH branching ratio of 1.2 × 10 −3 (8.3 × 10 −4 ) at the 95% confidence level. A search for flavor-changing neutral currents (FCNC) in events with the top quark and the Higgs boson is presented by the CMS collaboration [19]. The observed (expected) upper limits at 95% confidence level are set on the branching ratios of top quark FCNC decays, BR(t → uH) < 7.9 × 10 −4 (1.1 × 10 −3 ) and BR(t → cH) < 9.4×10 −4 (8.6×10 −4 ), assuming a single non-zero FCNC coupling. These prior works are the basically constitutes the starting point of FCNC researches at post Higgs era researches for FCNC. More recently, the limits given by CMS collaboration have been further improved as BR(t → uH) < 1.9 × 10 −4 (3.1 × 10 −4 ) and Especially we focus on the limits by CMS collaboration [20]: we use same effective Lagrangian up to a factor of weak coupling constant g and follow their limits on coupling constants which are 0.037 for η u and 0.071 for η c . In the case of ATLAS collaboration [18], they give an lower bound arround 0.065 for both coupling constants. However, bear in mind that the differences mainly come from using different effective Lagrangian which is discussed at model framework section. We investigate the problem by setting coupling constant 0.07 for scenarios about to introduce at following sections and try to improve the limits.
In recent years many new collider ideas such as HL-LHC/HE-LHC/FCC [21][22][23] have been reported and technical design report of HL-LHC has been published. Most promising feature of the HL-LHC collider for BSM searches is increased COM energy (14 TeV) and especially its luminosity [21] of up to 3 ab −1 . Some phenomenological researches for future colliders and HL-LHC have already been started up; for tqH couplings have been explored at a high-luminosity ab −1 ep colliders (with the possibility of electron beam having a polarization of 80% and electron energy is typical 60 GeV), the 2σ upper limits on Br(t → uH) have been obtained as 1.5 × 10 −3 and 2.9 × 10 −4 at the future colliders LHeC and FCC-eh, respectively [24].
Development of such a collider has notable effects on the BSM literature evidently since it to offer new possibilities for phenomenological studies and gives a large room for potential discoveries/exclucions. It offers an opportunity to rule out flavor violating 2HDM for t → cH case and penetrate the other regions forseen by other models (such as Randall-Sundrum model) [25]. As a consequence exploiting the physics potential of HL-LHC is crucial for next phase of BSM searches.
The phenomenological researches and simulations based on new colliders started to making predictions about new physics scenarios and set new limitations. To be specific at HL-LHC for FCNC interactions [26][27][28][29][30], branching ratios are updated as BR(t → qh) < O(10 −4 ) using various different analyses from different channels and processes; thus couplings are expected to go below η q = 0.04 which is rougly below the known limits from experiments. Expected FCNC decay widths and branching ratios are given at Fig. 1 and 2 respectively according to three scenarios which are important and handled seperately to set limits for couplings in following sections. In this study, we would like to investigate the problem and seek for the new limits at HL-LHC. To do so, we restrict ourselves to production mechanisms of same sign tt(tt) pairs (signal processes pp → tt including the exchange of Higgs boson at the HL-LHC. In addition, because the analysis was carried out in the HL-LHC, with the anticipation that the systematic uncertainties described in the literature would, in general, reduce, a value of 20% was used for the systematic uncertainties, which is still near to the limitations given in the literature and discussed in the findings section. This was done in light of the fact that the analysis was carried out in the HL-LHC.We introduce the kinematical variables to enhance the signal (S) and background (B) ratio. Angular separation of the two same sign leptons could indicate the new physics effects in tt(tt) production process, and separate the signal from background processes.
In order to make this research more detailed and similar to other studies in the literature (for comparison purposes), three different scenarios were designed for the signaling process. These are the u+c, only u and only c scenarios. As the nomenclature suggests, FCNC transitions are made possible from the top quark to the other two quarks in the u+c case, while in other cases the transitions are limited to just one quark.

II. MODEL FRAMEWORK
The flavour changing neutral current interactions of the top quark with other particles of the SM have been described in a general way as an extension [14,15]. This    provides a direct connection between experimental observables and the new anomalous couplings. The Lagrangian describing FCNC tqH interactions in model independent manner is given as Ht(η L u P L + η R u P R )u + h.c.
where the η L/R q couplings set the strength of the coupling between the top quark, the Higgs boson and up or charm quark, as well as the chirality of this coupling. They can be complex in general, however we take into account real parts of the couplings to reduce the free parameters. In literature this interaction can be seen as modeled without the constant 1 √ 2 thus gives higher top branchings a factor of 2 [31]. The FCNC processes that corresponds to tqh interactions have been described by a similar Lagrangian [19] with an extra factor of weak coupling constant. To switch between models we just need to remind this conversion factor. We keep that constant here in order to make bounds more strict mean while keeping the conversion to other models in our mind. Note that it effects cross section and number of evens naturally too, thus makes signal processes even harder and realistic. The decay width for FCNC channels can be calculated as and its numerical value depends on the coupling values related to Γ(t → qh) 0.1904(η 2 qL + η 2 qR ) GeV. The branching ratio to an FCNC channel can be expressed as BR(t → qh) = Γ(t → qh)/Γ(t → all). Since the dominant decay mode of top quark is Γ(t → W b), this branching ratio mostly related to (η 2 qL + η 2 qR ) factor especially for smaller coupling values.
The model framework can also be compared with the formalism assumed that the FCNC interactions occur via a weak sector. The relevant effective interaction Lagrangian including a new flavor changing scalar (φ) is given where the coupling parameters a u,c and b u,c denote the scalar and axial couplings between top quark and up-type light quarks (u, c) which proceeds through the exchange of a scalar φ. To compare different formalism for the topscalar FCNC we find the correspondance of the couplings Assuming no specific chirality dependence (same value for left and right handed couplings) of the process we may set a q = η q / √ 2 and b q = 0. In this study, we use a template model. The paremeters that appear in the topFCNC_UFO [32,33] model are complex numbers in general and their real and imaginary parts can be set manually. In this work, we restrict ourselves to real parameters in order to reduce the free parameters.

III. CROSS SECTIONS OF SIGNAL AND BACKGROUND
At the first step before event generation we calculate the cross section for FCNC processes including tqh which vertices leads to same sign signal final state as shown schematicaly at Fig. 3. Since the cross-section is proportional to the modulo quartic of the value of the anomalous couplings. In figure 4 and 5 we can see due to presence of up type quarks in proton, pp → tt process is much more favorable than pp →tt. Although the contribution from the signal pp →tt to same sign lepton signal compared to the signal from pp → tt is nearly less than one order of magnitude, we also use that contribution to enhance the signal.  . Estimated cross sections according to coupling constant of two same sign lepton signal in FCNC processes. As we can see main contribution comes from positively charged top pair due to higher parton distributions of valance quarks at proton which differs nearly one order of magnitude. Nevertheless addingtt production we use negatively charged lepton pair to enhance the signal process. We assume all FCNC coefficients are the same and all channels are open (to state exactly we use u + c → ηu = ηc case).
After setting model parameters the signal samples and background samples are generated with MadGraph5 [34]. In the partonic and hadronic level simulations we use the parton distribution function (PDF) set NNPDF2.3 [35] at MadGraph5's default energy scale. PYTHIA8 [36] is used for shower and hadronisation processes and finally DELPHES 3 [37] is used for detector level simulation. Result files are analyzed with Root6 [38].
As mentioned before same sign lepton signal has relatively low backgroud, which is advantageous and many of the background processes fall into reducible background category which means although they are present due to similarities between signal process by applying proper analyze cuts their contributions can be well reduced. However there still exist tough irreducible backgrounds. The contributions from various backgrounds are listed below.
Characteristics of signal events are two jets (b-tagged if possible), two same sign leptons, and missing transverse energy. We choose our background processes by considering three fundamental features: • Similarity of final state particles as much as possible The matrix element of the FCNC process includes two new physics vertices which are proportional to η 2 q , hence the cross section is proportional to η 4 q . If c quark does not involve in interactions then cross section depends only on η 4 u . Since the PDF of valence u quarks are high, the contribution to cross section from u quarks are consideribly high. In the case of a forbidden u quark interaction, cross section completely depends on η 4 c while cross section is lower. At u + c scenario when both u and c actively take part at interaction in addition to former coefficients plus additional cross terms present and total cross section is relatively higher from only u case showing that dominant part of interactions carries the fingerprint of u quark distributions in proton.
with signal processes.
• High cross section compared to signal.
• Having same reconstruction inputs as for the signal. Table II have at least one of these properties, besides some of them have two. As long as all processes have their own unique nature, they more or less differ at least one criterion or partly one or two criterion.

Backgrounds given in
The processes pp → W ± W ± jj; pp → ttW ± ; pp → ttl + l − ; pp → ttW + W − with same-sign dilepton decay modes which are most similar to our signal process are directly background to our signal process and they are all irreducible. Although they give same final state content with the signal, pp → ttW ± reconstruction region is slightly different. This process also gives similar products at final state. However its cross section is high. In the case pp → W ± W ± jj, on the one hand reconstruction region is significantly different, on the other hand its particle content is exactly the same. In addition to previous two discussions as an advantage for analysis pp → ttl + l − procces has low cross section compared to other two. Nevertheless its reconstruction region is fairly same. pp → ttW + W − with leptonic decay modes directly produce signal content, however its reconstruction region noticeably distinct. Besides its cross section is quite high. Similar arguments can easily be expanded to Table II. Signal and background processes with leptonic decay channels: We consider the positively and negatively charged leptonic final states for maximal imitation of signal process. So background events generated with this regard. We force particles to give l + /l − final states if possible. Otherwise, we let particles to decay any channel. For W boson at intermediate states we always take leptonic decay modes for get the maximal similarity with signal.
Process Cross section(pb) Intermediate states with/without jets pp → ZZjj other backgrounds. Others are reducible backgrounds: even though their particle contents are similar to signal, either their cross sections are low and reconstruction region significantly different. In that regard they satisfy only one criterion while irreducible ones fulfill two or more.
Further we select decay channels of background events as such to give same sign 2l ± with 2j and MET. Jets includes at least one b-tag jet. This ensures the maximum cross section for background and gives more contribution to histograms when we consider the detector effects such as misidentification and over counting of particles.
Inability to distinguish between signal and background processes increases with misidentification of particles and loss of particles due to detector effects. These effects causes the fuzzing of characteristics of signal while imitating the features of signal for background processes. Moreover b-tag efficiency plays also an important role for analyzing the signal and background events. Since two btagged jets are a major property of signal. Nevertheless two b-tagged jets requirement is so strict for observability of signal while reducing background effects too. Therefore we confined ourselves to at least one b-tagged jet while recognizing characteristics of our signal and background processes. It is also important to note that there is no interference between signal and background at this level of calculation.

IV. ANALYSIS
At first stage we have started with the known limits from current LHC experiments that put a limit on the FCNC coupling constant value η q = 0.07 which is already reached and then use benchmark value η q = 0.07 to insvestigate the limits for upgrading HL-LHC detector to search for a possible FCNC signal outcome. After that we seek edge values to limit and finalize our research. We will look forward to push the limits for η u+c , η u and η c separately.
We use the statistical significance SS disc and SS exc for discovery and for exclusion as given in [39][40][41][42]. For exclusion of a parameter value we are looking SS exc > 1.645 corresponding to a confidence level of 95% CL. In order to make it complete we will give limits for discovery relation too. Both relations reduces to S √ B at large background limit. In addition, we will conduct an evaluation in which systematic uncertainties are estimated in order to comprehend how systematic uncertainties influence our results. For these calculations, we will use the following formula for discovery with systematic uncertainties and for exclusion case we use the equation where x being When it comes to our analysis path, the first thing we'll do is focus on the key points for analysis and talk about the unique characteristics of the signaling process. These characteristics will then be disclosed by presenting the kinematic variables, and the method to be used in the study will be determined. Following that, the analysis will be performed, and the results will be provided.
Here we will track exactly two positively/negatively charged leptons as same sign lepton pairs since we investigate the case W ± → l ± ν l ± followed after t(t) → W + b(W −b ). Missing transverse energy is also an essential charactersitics of the process too. We note that despite we have only two b-jets in our signal when we consider the nature of interaction, more jets must be generated and we need to distinguish them from bottom quarks to reconstruct two top quarks. That point needs a little bit attention when we think of backgrounds and to make it clear we would like to go deeper: as we know our background events have more particles, in addition the nature of interaction also dictates numerous jets which gives more hadronic transverse energy. When we consider both, a cut that is limiting the number of jets seem to be advantageous. The best choice at first glance is limiting jet number as two, so we conclude with exact event selection. Nonetheless, taking into account detector effects, in a situation where two leptons are detected individually, if there are no jets or only one jet, these jets are more likely to escape from the detector. Working with a small number of jets is useful in this regard, as backdrops are highly prominent when working with a large number of jets. Furthermore, because the top quark is the source of leptons in the processes, it can be assumed that every case in which two same-sign leptons are seen belongs to the signal event, again taking charge conservation into account. Again, b-tagging serves a purpose here. Of course, without jets, this labeling is not conceivable. However, in single-jet (or fat jet) scenarios, this criterion can be used to provide the analysis a boost.
For lepton flavors we have 2 possibilities namely e ± and µ ± for l ± case since τ lepton disintegrates before reach the detector so its analysis is out of scope. In that respect we divide analysis region to three which includes three possibilities of same sign lepton pairs (e ± e ± , µ ± µ ± , e ± µ ± ) with exactly two jets while at least one of them b-tagged and lastly presence missing transverse energy in events.
Decay of top quarks in their rest frame give rise to high p T b-jets larger than about 80 GeV as a prediction in addition same happens for W + bosons and daughter particles should have at least 40 GeV. These particles also carries momentum, thus we expect boosted behaviour at histograms for mother and daughter particles.
To sum up at the begining of the analysis we have divided signal region to three analysis region with exact event selection, followed by simple cuts given in Table  IV. Here, the η cuts were choosen to work with the more sensitive regions of the detector for leptons especially. Furthermore, ∆R cuts were established as the minimum lepton isolation criteria. For trigerring and good object selection criteria for jets and leptons, missing transverse energy and p T cuts are minimally incorporated. To avoid a fat-jet scenario, a ∆R cut was regarded appropriate for jets, although jets reaching the detector were tolerated by avoiding an η limiting cut. As previously stated, it was noted that at least one of the jets entering the depicted histograms was b-tagged. Here we give the kinematical distributions for lepton p T at Fig. 6, 7 and 8, for lepton η at Fig. 9 and 10, H T and MET at Fig. 11 and 12, and lastly jet p T and η at Fig. 13, 14, 15, 16 belonging the signal process.
Finally we present histograms showing the characteristics of jets produced and little comment on them.
These histograms compare the behavior of signal and background events without delving into a detailed investigation. Except for a few cuts relevant to the study, the segments utilized for event production have been transferred to the detector level in order to provide the histograms in their simplest form. While some variables reflect differences. Although it is possible to separate the signal process, which includes two new physics vertices (and lowers cross section drastically). In the background, due to the high cross-section of the ttW ± background and its similarity to the signal, it is not possible to make a discrepancy after a point and provide the desired improvement in the analysis.
Although this makes the investigated process more appealing for exclusion, because the path forward with cutbased analysis is limited, better results can be obtained by utilizing machine learning techniques with the help of variables defined after these basic cuts.
The most fundamental variables in this analysis are the p T , η and φ components of the jets up to 4th as well as the same kinematic variables as the first two leptons. Furthermore, variables such as missing E T , H T , and ∆R are used together with, the invariant masses of the two jets and two leptons, the invariant and transverse masses in the quadruple state (l 1 , l 2 , j 1 , j 2 ), and finally m  Figure 6. For the signal process, lepton pT distributions e ± e ± , µ ± µ ± , e ± µ ± event regions: Histogram clearly shows that e ± µ ± final state is more favorable. The e ± µ ± pair comes from disintegration of W ± pairs which have about 80 GeV rest mass. Hence that energy and momentum shared by final particles and gives a peak arround 40 GeV with boosted behavior. However same flavor final states shows an asymmetry originates from the following reasons: Detector always discriminates lower and higher pT particle which gives a gap between first and second highest pT object. Nevertheless they all have boosted behavior and give peaks close to 40 GeV as well.
and   During the generation of these variables, cuts identical to those in Table IV were utilized, with minor modifications. First, the event selection regions are separated immediately into two lepton regions with same signs, regardless of the number of jets. While the fundamental sections of the p T cuts were kept, the criteria for lepton and jet separation were abandoned. In addition, the p T cuts for the fifth jet remained at 15 GeV. Each η variable is set to a value less than 2.5. To contrast the prominent signal regions in low jet number and low H T states with the dominant background processes in high jet number and high H T states, the number of jets is included up to a maximum of five and processes begin in non-jet states. In addition, the parsing of the ∆R variable was to be performed solely using machine learning techniques [43].
Finally, the findings of the BDT analysis are presented in Fig. 19, 20. Observing the nonlinear behavior of the signal and background, suitable approaches were chosen. Since a method based on fluctuations, such as BDT, is employed. A decision tree takes a set of input features and splits input data recursively based on these features.   Figure 11. Scalar HT distribution for the signal and background processes. Since more jet generated at backgrounds they have relatively shitfted to forward.
Boosting is a method of conbining many weak learnings (trees) into a strong classifier. It has been confirmed that the rate of discriminating gradually increases between the training and testing phases. Fig. 19 demonstrates that, as a result of the employment of several variables with high event numbers, the distribution and height of the signal's curve are significantly superior to the back one.

V. RESULTS AND CONCLUSIONS
In this study we have searched for accesible limits for top-Higgs FCNC couplings using same sign lepton channel at the HL-LHC (see Fig. 21, 22). This channel gives clean signal signature in addition to its low reducible/irreducible background. However, this channel suffers from two new physics vertices. Thus, these effects lower the cross section drastically which is a disadvantageous feature of this analysis. Keeping these in mind we can conclude that: simulation of this process with the same sign lepton channel turns into a laboratory for testing the mentioned scenarios in the text. In this respect this channel determines the upper limit for couplings and benefits exclusion limits rather than discovery.
We have started with coupling constant η q = 0.07 to demonstrate the characteristics of signal an catch the limits given in Ref. [18] whose limits are more or less same as our benchmark value. Then, as stated at introduction section we have tried to improve our results and get better limits for FCNC couplings.
In Table V we summarize our analysis results with the discovery and exclusion significance. Due to the sensi-  tivity of the study to exclusion, examining exclusion instances first will expose the results more clearly. First, the results of the initial review have improved the known limits [18][19][20], with the exception of some channels [20]. In addition, although how the results will be compared with one another is discussed in the second chapter, these changes will be discussed in greater detail here.
As expected, the η u+c instance produced the best results. Cases η u and η c followed these outcomes, respectively. Theoretically and empirically, the results vary little in the idealized scenario, assuming a total of 20% systematic uncertainty. This means that the background in the study has been eliminated with great success, and these uncertainties will not significantly impact the outcomes. In the case of exclusion, we have improved the coupling constant limits for η u+c and η u situations, although our limits for η c are more stringent. However, the branching ratios for the η c situation appear to have already been exceeded. Similarly, the η u and η u+c scenario exceeds the LHC constraints by a small amount. At this point, as the effective Lagrangiande utilized by  Figure 17. ∆R(j1, j2) distribution between two jets. Although there is no significant difference between the signal and background ∆R distributions at this point, it stands out as one of the most important variables since the signal process is symmetrical in its stationary frame of reference. Jets have the direct top quarks' back-to-back scattering structure.
CMS contains a weak interaction constant, it is apparent that the results may stray further from the known limitations with this factor, despite the fact that the analysis indicates the reverse. From the obtained coupling constants, the resulting branching ratios are determined. The minimum values coupling constants can attain are proportional to the number of events in the analysis, or indeed the cross section. When the analyzed process consists of two vertices, it is dependent on constants of the fourth order and has an advantage proportionate to the inverse square of the coupling constant size when compared directly. In fact, this circumstance nullifies the influence of the weak interaction constant and drastically decreases the results below the known levels. In this regard, it becomes evident why the channel is superior for exclusion and why it establishes very strict upper limits. Nonetheless, these constraints also limit other studies of top quark-Higgs FCNC interactions. Since these restric- w/wo boson(s) t t Bosons w/wo jets All Figure 18. ∆R(l + 1 , l + 2 ) distribution between two leptons. The ∆R variable in leptons is very crucial for discrimination, as it is in jets, but their behavior is much looser in comparison to jets due to additional energy-momentum conservation constraints from the decay of the W boson. Figure 19. By employing an appropriate cut, it is achievable to distinguish the signal from the background with great efficiency. (The curve was determined by analyzing ηuc = 0.07.) tions are precluded for this channel with two vertices, it stands to reason that studies with a single vertex will go below this limit, at least proportional to the obtained coupling constant value. Note that, we have also caught the phenomenelogical limits for HL-LHC expected [26][27][28][29][30]. At that studies η q varies near 0.04 (Bear in mind that models does not include additional 1 √ 2 factor [27][28][29]). [? ] Concerning the scenario of discovery, the limits reached for discovery coincide with the limits reached by the CMS and ATLAS collaborations at lower total luminosity values. In this instance, if further data is obtained, this value indicates that exploration is feasible up to the region's limit. However, it should not be forgotten that due to the nature of the analysis channel, they are still upper limitations. In this context, it has been proved that the values for an analysis involving a single FCNC vertex can be reduced.
Lastly upcoming colliders will provide better visions Figure 20. In the analysis, nonlinear approaches were predominantly employed. This is due to the fact that the structure of the signal and background distributions is more accurately reflected in this manner. In this regard, linear approaches such as Fisher's and its derivatives are unsuitable for analysis. Again, similar procedures were not adopted since they did not produce successful results. However, other nonlinear approaches were incorporated in this context so that the analysis could be compared to other ways and its evolution could be observed. The results indicate that the inputs are uniformly distributed, no overtraining was seen, and the analysis produced a high level of diversification overall.
for FCNC interactions [23]. To compare them with each other, we may say HL-LHC and FCC-eh are expected to work at same region. Moreover, HL-LHC offers better limits when we compare it with ILC/CLIC [23]. So our results have some implications on the analysis have been done for both FCC-eh and ILC/CLIC. In support of this, studies gives similar results to ours done for FCC-eh [24]. Even though results of HL-LHC will give direction to new researches without any doubt, there is a gap in COM and luminosity values between HL-LHC and FCC-hh. It is expected to taken down limits even further by FCC-hh. FCC-hh can possibly rule out RS models, and start to penetrate the MSSM region.
To sum up our findings we may say, while these limits are compatible with the expectations from HL-LHC, which enforce limitations for findings on other channels: Since on the one hand this channel gives its clean signal fingerprint, on the other hand even lower cross section, the same-sign lepton channel provides upper limits and provides hints to other detectors and thanks to its clean signal fingerprint, it also imposes partial limitations on other channels. Our limits can also be combined with the other sensitive channels for similar scenarios. reading of the manuscript. We wish to acknowledge the support of the AUHEP group, offering suggestions and encouragement. The numerical calculations reported in this paper were partially performed at TUBITAK ULAK-BIM, High Performance and Grid Computing Center (TRUBA resources).