Search for pair production of a new heavy quark that decays into a W boson and a light quark in pp collisions at √s = 8 TeV with the ATLAS detector

A search is presented for pair production of a new heavy quark ( Q ) that decays into a W boson and a light quark ( q ) in the final state where one W boson decays leptonically (to an electron or muon plus a neutrino) and the other W boson decays hadronically. The analysis is performed using an integrated luminosity of 20 . 3 fb − 1 of pp collisions at ﬃﬃﬃ s p ¼ 8 TeV collected by the ATLAS detector at the LHC. No evidence of Q ¯ Q production is observed. New chiral quarks with masses below 690 GeVare excluded at 95% confidence level, assuming BR ð Q → Wq Þ ¼ 1 . Results are also interpreted in the context of vectorlike quark models, resulting in the limits on the mass of a vectorlike quark in the two-dimensional plane of BR ð Q → Wq Þ versus BR ð Q → Hq Þ .

Search for pair production of a new heavy quark that decays into a W boson and a light quark in pp collisions at ffiffi s p ¼ 8 TeV with the ATLAS detector G. Aad et al. * (ATLAS Collaboration) (Received 14 September 2015;published 22 December 2015) A search is presented for pair production of a new heavy quark (Q) that decays into a W boson and a light quark (q) in the final state where one W boson decays leptonically (to an electron or muon plus a neutrino) and the other W boson decays hadronically. The analysis is performed using an integrated luminosity of 20.3 fb −1 of pp collisions at ffiffi ffi s p ¼ 8 TeV collected by the ATLAS detector at the LHC. No evidence of QQ production is observed. New chiral quarks with masses below 690 GeVare excluded at 95% confidence level, assuming BRðQ → WqÞ ¼ 1. Results are also interpreted in the context of vectorlike quark models, resulting in the limits on the mass of a vectorlike quark in the two-dimensional plane of BRðQ → WqÞ versus BRðQ → HqÞ.

I. INTRODUCTION
The recent observation of the long-predicted Higgs boson by the ATLAS [1] and CMS [2] collaborations now completes the Standard Model (SM). In spite of its success, the SM cannot account for dark matter and the matter/antimatter asymmetry in the Universe and also fails to provide insight into the fermion mass spectrum, nonzero neutrino masses, why there are three generations of fermions, or why parity is not violated in the strong interaction. Furthermore, the observed Higgs boson is unnaturally light, requiring fine-tuning to cancel radiative corrections that would naturally result in a mass many orders of magnitude larger, a discrepancy known as the hierarchy problem [3].
A variety of models has been proposed to address the various shortcomings of the SM. For example, a primary motivation for supersymmetry (SUSY) is to solve the hierarchy problem [4]. In SUSY models, the quadratically divergent radiative corrections to the Higgs-boson mass due to SM particles are automatically canceled by the corrections from the supersymmetric partners. Models such as little Higgs, composite Higgs, and topcolor take a different approach, proposing that electroweak symmetry breaking happens dynamically as the result of a new strong interaction [5][6][7][8][9]. Grand unified theories (GUTs) go further, unifying the electroweak and strong forces by proposing that the SM gauge symmetry SUð3Þ C × SUð2Þ L × Uð1Þ Y is the low-energy limit of a single fundamental symmetry group such as SOð10Þ or E 6 [10,11], which could potentially explain the observed spectrum of fermions and even provide insight into the unification of the electroweak and strong forces with gravity. A feature in many of these, and other models [12][13][14], is the prediction of vectorlike quarks (VLQs), hypothetical spin- 1 2 particles that are triplets under the color gauge group and have identical transformation properties for both chiralities under the electroweak symmetry group. Furthermore, massive VLQs would respect gauge invariance without coupling to the Higgs field. This allows VLQs to avoid constraints from Higgs-boson production [15]; if the Higgs sector is minimal, these constraints rule out additional chiral quarks. However, some two-Higgs-doublet models are able to avoid those constraints and accommodate a fourth generation of quarks [16].
In the models of interest, the VLQs have some mixing with the SM quarks, allowing them to decay to SM quarks and either a W, Z, or Higgs boson; however, the exact nature of the coupling depends on the model. For example, in composite Higgs models, the VLQs are involved in a seesaw mechanism with the SM quarks, so the lightest VLQ couples almost exclusively to the heaviest SM quarks (t-and b-quarks) [6]. However, there are also models that predict TeV-scale VLQs that could preferentially decay to light SM quarks (q ¼ u; d; s, or c) [10,11,17]. For example, the left-right mirror model (LRMM) [17] predicts three generations of heavy "mirror" quarks, with the lightest mirror generation coupling to the lightest SM generation. The two lightest mirror quarks could be pair-produced at the LHC via the strong interaction and would then decay to Wq, Zq, or Hq (q ¼ u or d). The LRMM would provide an explanation for tiny neutrino masses, parity violation in weak interactions, parity conservation in strong interactions, and could be the first step toward uncovering the symmetry structure of a GUT. Another model predicting VLQs that decay to light quarks is the E 6 GUT with isosinglet quarks [10,11]. In this model, after the E 6 symmetry is broken down to the SM group structure, VLQ partners to the d-, s-, and b-quarks are predicted. If the VLQs have the same mass ordering as their SM partners, the lightest VLQ would couple predominately to first-generation SM quarks (q ¼ u or d). The values for the branching ratios to the three decay modes (Wq, Zq, Hq) depend on parameters in the model. The values in the E 6 isosinglet model range from approximately (0. 6, 0.3, 0.1) to (0.5, 0.25, 0.25), while the LRMM allows branching ratios from approximately (0. 6, 0.4, 0) to (0, 0, 1), depending on the VLQ mass and mixing angles.
If such new quarks exist, they are expected to be produced predominantly in pairs via the strong interaction for masses up to Oð1 TeV) in LHC collisions with a centerof-mass energy of 8 TeV. Single production of a new heavy quark, Q, would be dominant for very high quark masses, but the production rate is model dependent and could be suppressed if the coupling to SM quarks is small. To date, there have been two analyses of LHC data sensitive to VLQs that decay to light quarks, both using 1.04 fb −1 of ATLAS data at ffiffi ffi s p ¼ 7 TeV: one search for pair-production [18] and one for single production [19]. The pair-production search set a lower limit on the Q mass of 350 GeV at 95% confidence level, assuming BRðQ → WqÞ ¼ 1. Such a signal was also ruled out by the Tevatron for masses up to 340 GeV [20]. This paper presents a search for new heavy quarks that couple to light SM quarks using data collected by the ATLAS detector. The analysis focuses on model-independent pair production of a new heavy quark and its antiparticle, which then decay through a charged-current interaction to a final state with a single electron or muon, missing transverse momentum and light SM quarks, making it complementary to searches for VLQs that decay to thirdgeneration quarks [21][22][23][24][25][26][27][28]. The dominant Feynman diagrams for the signal process are shown in Fig. 1. Background rejection is achieved through event topology and kinematic requirements. In particular, a kinematic variable motivated by the splitting scale of a heavy object into daughter particles [29,30] is used to reduce the background when selecting the pair of jets consistent with a hadronically decaying W boson. The statistical interpretation of the data uses the invariant mass distribution of the Q candidate formed from the hadronically decaying W boson and a light quark. The results are interpreted both in the context of VLQs and a chiral fourth-generation quark.

II. THE ATLAS DETECTOR
The ATLAS detector [31] at the LHC covers nearly the entire solid angle around the interaction point (IP) 1 . It consists of an inner tracking detector surrounded by a thin superconducting solenoid, electromagnetic and hadronic calorimeters, and a muon spectrometer incorporating three large superconducting toroid magnets.
The inner tracking detector system is immersed in a 2 T axial magnetic field and provides charged-particle tracking in the range jηj < 2.5. A high-granularity silicon pixel detector covers the interaction region and typically provides three measurements per track. It is followed by a silicon microstrip tracker, which usually provides four twodimensional measurement points per track. These silicon detectors are complemented by a transition radiation tracker, which enables radially extended track reconstruction up to jηj ¼ 2.0. The transition radiation tracker also provides electron identification information based on the fraction of hits (typically 30 in total) above a higher threshold for energy deposits corresponding to transition radiation.
The calorimeter system covers the pseudorapidity range jηj < 4.9. Within the region jηj < 3.2, electromagnetic calorimetry is provided by barrel and endcap highgranularity lead/liquid-argon (LAr) calorimeters, with an additional thin LAr presampler covering jηj < 1.8, to correct for energy loss in material upstream of the calorimeters. Hadronic calorimetry is provided by the steel/scintillator-tile calorimeter, segmented into three barrel structures within jηj < 1.7, and two copper/LAr hadronic endcap calorimeters with jηj < 3.2. The solid angle coverage is completed with forward copper/LAr and tungsten/LAr calorimeter modules optimized for electromagnetic and hadronic measurements respectively.
The muon spectrometer comprises separate trigger and high-precision tracking chambers measuring the deflection of muons in a magnetic field generated by superconducting air-core toroids. The precision chamber system covers the region jηj < 2.7 with three layers of monitored drift tubes, FIG. 1. Leading-order Feynman diagrams for QQ → WqWq production at the LHC. 1 ATLAS uses a right-handed coordinate system with its origin at the nominal IP in the center of the detector and the z axis along the beam pipe. The x axis points from the IP to the center of the LHC ring, and the y axis points upwards. Cylindrical coordinates ðr; ϕÞ are used in the transverse plane, ϕ being the azimuthal angle around the z axis. The pseudorapidity is defined in terms of the polar angle θ as η ¼ − ln tanðθ=2Þ. Angular distance is measured in units of ΔR ≡ ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ffi ðΔηÞ 2 þ ðΔϕÞ 2 p . and Z þ jets samples were produced using SHERPA v1.4.1 [48] with the CT10 PDF set, taking the cand b-quarks as massive. Samples of tt and single-top events were generated with POWHEG-BOX 3.0 [49,50] interfaced to PYTHIA v6.426 [47] using the Perugia2011C set of tunable parameters [51] for the underlying event and the CTEQ6L1 PDF set. Diboson production was modeled using ALPGEN v2.13 [52] interfaced to HERWIG v6.520 [53] with the CTEQ6L1 PDF set, except for the lepton þ jets final state (WV → lνqq with V ¼ W; Z), which used the SHERPA v1.4.1 event generator with the CT10 PDF set. The contribution from multijet events originates from the misidentification of a jet or a photon as an electron, or from the semileptonic decay of a bor c-quark, and the matrix method [54] is used to determine the kinematic distributions for the multijet background. The W=Z þ jets and multijet backgrounds are normalized to data and a data-driven correction is applied to the transverse momentum p T of the boson as described in Sec. VA. The tt cross section is determined by the Top þþ prediction, with the top-quark mass taken to be 172.5 GeV. The single-top samples are normalized to the approximate NNLO theoretical cross sections [55][56][57] calculated using the MSTW 2008 NNLO PDF set. The diboson background processes are normalized to NLO theoretical cross sections [58].
All simulated samples include multiple pp interactions per bunch crossing and simulated events are weighted such that the distribution of the average number of interactions per bunch crossing agrees with data. The generated samples are processed through a simulation [59] of the detector geometry and response using GEANT4 [60], then reconstructed using the same software as used for data. Simulated events are corrected so that the object identification efficiencies, energy scales, and energy resolutions match those determined in data control samples.

IV. EVENT SELECTION
The data analyzed in this search were collected with the ATLAS detector in 2012 and correspond to an integrated luminosity of 20.3 fb −1 . Data quality requirements are applied to remove events with incomplete, corrupted, or otherwise compromised subdetector information. Events are required to pass a single-electron or single-muon trigger. The p T thresholds are 24 GeV and 60 GeV for the electron triggers and 24 GeV and 36 GeV for the muon triggers. The lower-threshold triggers include isolation requirements on the candidate leptons, resulting in inefficiencies at higher p T that are recovered by the higherthreshold triggers without an isolation requirement.

A. Preselection
The basic object selection is called preselection and requires exactly one charged lepton (electron or muon), at least four jets, and large missing transverse momentum (E miss T ), as described below. The criteria are similar to those used in recent ATLAS top-quark studies [61], except that this analysis requires that there are no jets identified as originating from a b-quark. The expected and observed numbers of events after preselection are shown in Table I. There is negligible sensitivity to heavy quark production because the signal expectations for all masses are much smaller than the uncertainty on the background, dominated at this stage by systematic uncertainties.

Charged-lepton requirements
Electron candidates are required to have p T > 25 GeV and either jηj < 1.37 or 1.52 < jηj < 2.47 to exclude the transition between the barrel and endcap calorimeters. Muon candidates are required to have p T > 25 GeV and jηj < 2.5. Nonprompt leptons and nonleptonic particles may be reconstructed as leptons and satisfy the selection criteria, giving rise to nonprompt and fake lepton backgrounds. In the case of electrons, these include contributions from semileptonic decays of band c-quarks, photon conversions, and jets with a large electromagnetic energy fraction. Nonprompt or fake muons can originate from semileptonic decays of band c-quarks, from chargedhadron decays in the tracking volume or in hadronic showers, or from punchthrough particles emerging from high-energy hadronic showers [54]. The nonprompt and fake lepton backgrounds are reduced by requiring the lepton candidates to be isolated from other energy deposits or high-p T tracks. The tracks used in the isolation calculation are required to originate from the primary interaction vertex and have p T > 1 GeV. For electrons, an η-dependent limit is placed on the amount of energy measured in the calorimeter within a ΔR ¼ 0.2 cone around the candidate which is neither from the electron candidate itself nor from additional pp interactions. A similar requirement is placed on the scalar sum of the p T of tracks within a ΔR ¼ 0.3 cone around the track of the electron candidate. Each requirement has an average efficiency of 90% for electrons from Z → ee. Muons are required to have a p T -dependent track isolation [62], requiring the scalar sum of the p T from tracks with ΔR < 10 GeV=p μ T to be less than 0.05 · p μ T , where p μ T is the p T of the candidate muon track. This requirement has an efficiency of approximately 95% for muons from W → μν.
In this analysis, τ leptons are not explicitly reconstructed. Because of the high p T threshold, only a small fraction of τ leptons decaying leptonically are reconstructed as electrons or muons, while the majority of τ leptons decaying hadronically are reconstructed as jets.

Jet requirements
Events must contain at least four jets with p T > 25 GeV and jηj < 2.5 reconstructed using the anti-k t algorithm [63] with a radius parameter R ¼ 0.4. The jets are constructed from calibrated topological clusters built from energy deposits in the calorimeters, and they are calibrated to the hadronic scale [64]. Prior to jet finding, a local cluster calibration scheme [65] is applied to correct the topological cluster energies for the effects of calorimeter noncompensation, dead material and out-of-cluster leakage. The jet energy scale was determined using information from testbeam data, LHC collision data, and simulation [64,66]. All jets are required to have at least two tracks that originate from the primary interaction vertex. In order to suppress jets that do not originate from the primary vertex, jets with p T < 50 GeV and jηj < 2.4 are required to have a jet vertex fraction (JVF) above 0.5, where JVF is the ratio of the summed scalar p T of tracks originating from the primary vertex to that of all tracks associated with the jet. An overlap removal procedure [61] is applied to remove jets that were already identified as electrons.
To identify jets originating from the hadronization of a bquark ("b-tagging"), a continuous discriminant is produced by an algorithm using multivariate techniques [67] to combine information from the impact parameter of displaced tracks and topological properties of secondary and tertiary vertices reconstructed within the jet. The efficiency for a jet containing a b-hadron to be b-tagged is 70%, while the light-jet (c-jet) efficiency is less than 1% (20%) as determined in simulated tt events, where light jets are jets initiated by a u-, d-, s-quark, or gluon. If any jets are b-tagged, the event is rejected.  angle between the charged-lepton transverse momentum vector andẼ miss T .

B. Final selection
With the final-state objects identified, additional kinematic requirements are applied to exploit the distinct features of the signal, assumed to be heavier than the previously excluded mass of 350 GeV. The W boson and light quark originating from Q → Wq would be very energetic and have a large angular separation due to the large mass of the Q quark. On the other hand, the decay products of the W boson would tend to have a small separation due to the W boson's boost. By selecting only events that are consistent with these properties, the W þ jets background yield is reduced by orders of magnitude. To facilitate the discussion of the kinematic selection, the following objects are defined: (i) W had is the candidate for a W boson in the decay Q → Wq → qqq; (ii) W lep is the candidate for a W boson in the decay Q → Wq → lνq; (iii) q-jet is a candidate for the jet originating from the q in Q → Wq (i.e., from the decay of the heavy quark, not the hadronically decaying W boson). There are two q-jets per event, so q 1 (q 2 ) is used to denote the one with higher (lower) p T . The W had candidate is defined as a dijet system with p T > 200 GeV, angular separation ΔR < 1.0, and an invariant mass in the range of 65 to 100 GeV. All possible jet combinations are considered and, if multiple pairs satisfy the above requirements, the pair with the mass closest to that of the W boson [69] is chosen. If no W had candidate is found, the event is removed. This requirement results in 94% background rejection while keeping 53% of the signal if m Q ¼ 700 GeV. The mass distribution of the W had candidates prior to the mass requirement is shown in Fig. 2.
The W lep candidate is reconstructed using the lepton and E miss T , which is taken to be the neutrinop T . The longitudinal momentum of the neutrino is determined up to a two-fold ambiguity by requiring the invariant mass of the leptonneutrino system to equal the mass of the W boson. When no real solution exists, the neutrino pseudorapidity is set equal to that of the lepton because the decay products of the W boson tend to be nearly collinear for the kinematic regime of interest. In simulated samples, the rate of events with no real solution is approximately 35%. Signal events are expected to have energetic W bosons, so the W lep candidate is required to have p T > 125 GeV.
The W candidates (W had and W lep ) are then each paired with a different one of the remaining jets to create the two heavy quark candidates, Q1 and Q2. This step involves testing all possible pairings of q-jet candidates with the W had and W lep candidates. In addition, the W lep candidate may have two real solutions for the longitudinal momentum of the neutrino. Among the possible combinations of neutrino momentum solutions and Wq pairings, the one yielding the smallest absolute difference between the two reconstructed heavy quark masses is chosen. In simulated samples, the rate of correct Wq pairing is approximately 40% (48%) for a signal of mass 400 GeV (800 GeV). Once the heavy quark candidates are determined, the q-jets are required to have p T ðq 1 Þ > 160 GeV and p T ðq 2 Þ > 120 GeV and the difference between the reconstructed heavy quark masses must be less than 120 GeV. With candidate objects identified for two heavy quarks and their decay products, the following additional kinematic criteria are applied. Each event must have H T > 1100 GeV, where H T is the scalar sum of the transverse momenta of the lepton, neutrino, W had , and the two q-jets. The angular separation between the lepton and neutrino candidates must satisfy ΔRðl; νÞ < 1.4. The heavy quarks would tend to be central and back-to-back, so the angular separation between the two reconstructed heavy quarks is required to satisfy 2.0 < ΔRðQ1; Q2Þ < 4.2.
To further capitalize on the presence of a hadronically decaying W boson with high p T in the signal, the analysis makes use of a splitting variable [30] defined as where j 1 and j 2 are the two jets from the W had candidate and m j1j2 is the mass of the W had candidate. The two jets from a hadronically decaying W boson tend to have roughly equal energies, while dijets from QCD processes are likely to be asymmetric in energy. Furthermore, jets from W → qq tend to have a larger opening angle due to the mass of the W boson. Dividing by the dijet mass provides discrimination for cases in which a QCD dijet system happens to have large values for minðp Tj1 ; p Tj2 Þ 2 and ΔRðj 1 ; j 2 Þ 2 , as such systems are likely to have a very large mass. The background is reduced by a factor of 3.2 by requiring y 12 > 0.25. This requirement has a signal efficiency of approximately 50%, although the precise value depends on the mass of the heavy quark. The modeling of the splitting variable with a large number of events can be seen in the top histograms of Fig. 3, which depict the y 12 distribution after preselection and requiring a W had candidate. The bottom of Fig. 3 contains the distributions of y 12 immediately before the y 12 > 0.25 requirement is applied.
The final selection criteria are motivated by the fact that the decay products from the Q quark are well separated. The final requirements are ΔRðW had ; q 1 Þ > 1.0, ΔRðW had ; q 2 Þ > 1.0, ΔRðW lep ; q 1 Þ > 1.0, and ΔRðW lep ; q 2 Þ > 1.0. Table II presents a summary of the expected and observed numbers of events after the final selection, for which the signal (background) efficiency compared to preselection is approximately 8% (0.004%). The small contributions from tt, Z þ jets, dibosons, single-top, and multijet events are combined into a single background source referred to as "non-W þ jets." Uncertainties on the yields include the uncertainty due to the size of the signal and background samples and the cumulative effect of the systematic uncertainty described in Sec. VI.
The final discriminant variable used in this search is m reco , the reconstructed heavy quark mass built from the W had candidate and the paired q-jet candidate.

V. BACKGROUND MODELING
A. Correction to W þ jets modeling It is observed after applying the preselection criteria that the simulated W þ jets sample does not accurately model the p T spectrum of the leptonically decaying W boson candidate. This mismodeling leads to an overestimation of the W þ jets yields in the highmomentum tails of the E miss T , lepton p T , jet p T , and H T distributions. The dominant background for this analysis is W þ jets events in which the transverse momentum of the W boson, p T ðWÞ, is high. Therefore, it is important to have accurate predictions for both the overall normalization and the p T ðWÞ distribution. This section describes the procedure used to derive the W þ jets and multijet normalizations and a reweighting function to correct the vector-boson p T in the W þ jets TABLE II. Expected yields for the backgrounds and the VLQ signal with m Q ¼ 700 GeV, along with the observed number of data events, after applying all selection criteria. The uncertainties on the predicted yields correspond to the statistical uncertainty due to finite sample size and the systematic uncertainty, respectively.

Data 9 11
Events / 40 GeV and Z þ jets 2 simulated amples. All steps in the procedure rely on fits to the data using the p T distribution of the W lep candidate, p reco T ðW lep Þ, after applying the preselection requirements.
First, the normalizations for the W þ jets templates and the multijet templates are fit to the data, with all other background processes fixed at their expectations. The difference between the observed and predicted number of W þ jets events is assumed to be due to the cross section differing from its predicted value, so the electron and muon channels are fit simultaneously to determine a single W þ jets scale factor of 0.82. A correction for p truth T ðVÞ is then derived that minimizes the χ 2 between data and simulation for the p reco T ðW lep Þ distribution, where p truth T ðVÞ is the p T of the generated vector boson in the W þ jets or Z þ jets sample just before its leptonic decay. The reweighting is approximately unity for low p truth T ðVÞ, decreasing to 0.86 for p truth T ðVÞ ¼ 200 GeV and 0.65 for p truth T ðVÞ ¼ 500 GeV. Finally, the normalizations for the multijet samples and the corrected W þ jets samples are fit to the data, with all other background processes fixed. The fit is done simultaneously in the electron and muon channels. Figure 4 shows the p reco T ðW lep Þ distribution in the electron and muon channels after applying the p truth T ðVÞ correction and scale factors. The corresponding distributions for H T and E miss T are shown in Figs. 5 and 6. The uncertainties on the normalizations and p truth T ðVÞ reweighting are described in Sec. VI.

B. Validation regions
The following validation regions are used to verify the modeling of the background processes: (i) VR1: preselection, plus one W had and m reco < 350 GeV; (ii) VR2: preselection, plus one W had and H T < 800 GeV; (iii) VR3: final selection, but with the requirements on p T ðq 1 Þ, p T ðq 2 Þ, and H T changed to p T ðq 1 Þ< 160GeV, p T ðq 2 Þ < 80 GeV, and H T < 800 GeV and with no requirements on ΔRðW had ; q 1 Þ, ΔRðW had ; q 2 Þ, ΔRðW lep ; q 1 Þ, and ΔRðW lep ; q 2 Þ. The validation regions are orthogonal to the signal region but are nevertheless useful for checking with a larger number of events that the background normalization and kinematics are well modeled. The expected signal contribution in each validation region is approximately the size of the background uncertainty for signal masses around the previous limit (350 GeV) and decreases very rapidly for higher masses. VR2 and VR3 are used to validate the modeling of the final discriminant, m reco , as shown in Figs. 7 and 8. VR1 is used to verify the modeling of variables other than m reco , such as the H T distribution and the p T spectra for the individual objects.

VI. SYSTEMATIC UNCERTAINTIES
The uncertainties considered in this analysis can affect the normalization of signal and background and the shape of the final discriminant, m reco . Each source of systematic uncertainty is assumed to be 100% correlated across all samples, but the different sources are treated as uncorrelated with one another. Table III shows the impact of the dominant uncertainties on the normalization of the background processes and a signal sample with a mass of 700 GeV.

A. Normalization uncertainties
Uncertainties affecting only the normalization include those on the integrated luminosity (AE2.8%) and the cross sections for various background processes. The uncertainty on the integrated luminosity is derived following the same methodology as that detailed in Ref. [71]. This uncertainty is applied to all simulated signal and background processes.
After the final selection, the non-W þ jets background has a total normalization uncertainty of 15%. The predicted contribution from the multijet background is negligible compared to the uncertainty on the non-W þ jets background, so it is neglected.
As described in Sec. VA, the normalization of the W þ jets background is determined from a fit to data using both lepton channels. The uncertainty on the W þ jets normalization is determined by comparing that result to the normalization one would obtain if only the electron or only the muon channel were used. This is motivated by the fact that it is not known whether the normalization from the electron channel or the muon channel (or something in between) is correct and this procedure leads to an uncertainty of þ2.7=− 4.4%. The statistical uncertainty from the fit is negligible (0.03%) in comparison.
The rest of the systematic uncertainties can modify both the normalization and shape of the m reco distribution.

B. Shape uncertainties
Uncertainties on the trigger, reconstruction, and isolation efficiencies for the selected lepton are estimated using Z → ee and Z → μμ events [72,73]. In addition, high jetmultiplicity Z → ll events are studied, from which extra uncertainties on the isolation efficiency are assigned to account for the difference between Z boson and tt events. Uncertainties on the E miss T reconstruction and the energy scale and resolution of the leptons were also considered; however, these have a very small impact on the results.
The jet energy resolution is measured by studying dijet events in data and simulation. The simulation is found to agree with data to better than 10% [74]. The differences in resolutions between data and simulations are used to determine the relative systematic uncertainty. The uncertainty on the jet energy scale is evaluated by repeating the analysis with the jet energy scale shifted by AE1σ [64,66]. The jet reconstruction efficiency is estimated using trackbased jets and is well described in simulation. To account for differences in the efficiency for reconstructing jets in simulated events compared to collider data, the efficiencies are measured in both samples and the difference is taken as the uncertainty. The uncertainty due to the JVF requirement is evaluated by comparing the signal and background distributions with the JVF cut shifted up and down by 10%, a variation that spans the difference observed between data and simulation in this quantity. The b-tagging efficiency for b-jets, as well as c-jets and light jets, is derived in data and a simulated tt sample, parametrized as a function TABLE III. Overall normalization changes (expressed in %) in signal and background yields for the dominant systematic uncertainties considered. The selection presented here is the combination of the e þ jets and μ þ jets channels after the final selection.  67,75]. The corresponding (in)efficiencies in the simulated samples are corrected to match those in data and the uncertainties from the calibration are propagated through the analysis.
The W þ jets sample is assigned a p truth T ðVÞ-dependent shape uncertainty due to the correction described in Sec. VA. Four sources of uncertainty on the p truth T ðVÞ correction are considered: (1) the statistical uncertainty on the parameters of the reweighting function; (2) the difference between the nominal reweighting function and alternative corrections obtained by only considering the electron or muon channel; (3) the dependence of the fit on the choice of bin width when deriving the reweighting function; (4) the difference between alternative parametrizations of the reweighting function. A closure test is performed to verify that any residual differences between data and prediction are well within the assigned uncertainty.

VII. RESULTS
The final m reco distribution for the combined electron and muon channels is shown in Fig. 9 for three signal scenarios: m Q ¼ 600 GeV, m Q ¼ 700 GeV, and m Q ¼ 800 GeV. The observed distribution shows a slight excess over the SM expectation, but the excess is broader than expected for signal and is consistent with the background-only prediction at the level of 2 standard deviations. Therefore, the analysis proceeds to setting limits on the signal hypothesis.
The m reco distribution for the combined electron and muon channels after the final selection ( Fig. 9) is used to derive 95% confidence level (C.L.) limits on the QQ production cross section using the CL s method [76,77].
Limits on the pair production of new chiral quarks are evaluated by setting BRðQ → WqÞ ¼ 1. Figure 10 shows the observed and expected limits on a heavy chiral quark as a function of m Q , compared to the theoretical prediction [33][34][35][36][37][38]. The total uncertainty on the theoretical cross section includes the contributions from the scale variations and PDF uncertainties. Using the central value of the theoretical cross section, the observed (expected) 95% C.L. limit on the mass of a new chiral quark is m Q > 690 GeV (780 GeV). This represents the most stringent limit to date on the mass of a new quark decaying exclusively into a W boson and a light quark (u; d; s). This limit is also applicable to the production of pairs of down-type vectorlike quarks with electric charge of −4=3 which each decay into a W boson and a light quark (d; s).
Next, the VLQ signal samples are used to set limits on the mass of a heavy quark that decays to a light quark (u; d; s) and either a W, Z, or H boson. The results are given as a function of the branching ratios BRðQ → WqÞ versus BRðQ → HqÞ, with the branching ratio to Zq fixed by the requirement BRðQ → ZqÞ ¼ 1 − BRðQ → WqÞ − BRðQ → HqÞ. The analysis loses sensitivity at low masses due to the tight selection requirements optimized for the decay Q → Wq, so the results are presented as the upper and lower bounds on the mass range that is excluded at 95% C.L. The expected limits on m Q as a function of the branching ratios are shown in Fig. 11 and the observed limits are shown in Fig. 12. For example, for the branching

VIII. CONCLUSION
A search for new heavy quarks that decay to a W boson and a light quark ðu; d; sÞ is performed with a data set corresponding to 20.3 fb −1 that was collected by the ATLAS detector at the LHC in pp collisions at ffiffi ffi s p ¼ 8 TeV. No significant evidence of a heavy quark signal is observed when selecting events with one charged lepton, large E miss T , and four non-b-tagged jets. The results exclude new heavy chiral quarks with masses up to 690 GeV at 95% C.L. This search is also interpreted in the context of vectorlike quark models, for which the new heavy quark can decay to a light quark and either a W, Z, or H boson. New exclusions limits are presented in the twodimensional plane of BRðQ → WqÞ versus BRðQ → HqÞ.

ACKNOWLEDGMENTS
We thank CERN for the very successful operation of the LHC, as well as the support staff from our institutions without whom ATLAS could not be operated efficiently. We acknowledge the support of ANPCyT    (2015) 112007-13 [28] CMS Collaboration, Search for a Vectorlike Quark with Charge 2=3 in t þ Z Events from pp Collisions at ffiffi ffi s p ¼ 7 TeV, Phys. Rev. Lett. 107, 271802 (2011Lett. 107, 271802 ( ). A 506, 250 (2003.
[61] ATLAS Collaboration, Measurement of the top pair production cross section in 8 TeV proton-proton collisions using kinematic information in the lepton þ jets final state with ATLAS, Phys. Rev. D 91, 112013 (2015