Backup Slides - PowerPoint PPT Presentation

About This Presentation
Title:

Backup Slides

Description:

Study biases to kinematics from tracking. Study IP resolution(data/MC): Primary, B & D vertices ... Kinematics. Can we rely on kinematical biases estimated from MC? ... – PowerPoint PPT presentation

Number of Views:3005
Avg rating:3.0/5.0
Slides: 39
Provided by: Ale8223
Learn more at: https://www-cdf.lbl.gov
Category:

less

Transcript and Presenter's Notes

Title: Backup Slides


1
Backup Slides
2
Moments Correlations
Stat StatExpBR StatExpBRTheo
m 74 48 (no BR) -
M 92 62 -
? -99 -88 -77
3
Vcb exclusive determination
  • Measure absolute scale of B?Dl?
  • D states also important for Vcb exclusive
    determination
  • end-point in q2 for B?Dl? decays
  • systematic uncertainty from B?Dl? background

4
Channels with neutral B
  • B0 ? D l- ?
  • D ? D0 ? OK
  • D ? D ?0 Not reconstructed. Half the rate of
    D ?-
  • D ? D0 ?
  • D0 ? D0 ?0 Not reconstructed. Background to D0
    ?
  • D0 ? D0 ? Not reconstructed. Background to D0
    ?
  • D ? D ?0 Not reconstructed. Half the rate of
    D ?-
  • We will not deal with neutral B

5
Data Stability
A (152595-154012) Before winter 2003 shutdown B
(158826-165297) After winter 2003 shutdown C
(164303-165297) SVT 4/5
6
Kinematic Comparisons lD, D0?K???
7
Kinematic Comparisons lD, D0?K??0
8
Kinematic Comparisons D
9
Can we predict yields?
Two methods (a,b) to derive this BR
  1. Based on inclusive b?D()l?
  2. Based on exclusive B?D()l?, Dl?

PDG BR MC efficiency ratios
10
What background model for what?
  • WS is often used in this kind of analyses as a
    model for the background
  • We can also use our fully recod B from other
    triggers
  • We choose to use WS for the optimization
  • Embedding is being used as a cross-check for
    systematics

11
Whats available on the market
D?D
D?D0
  • No background subtraction
  • 80 events in D
  • 80 events in D
  • 215 events on D0
  • uncertainty gt sqrt(n)

D?D
12
Estimator Behaviour
13
K? Optimization
14
D Optimization
15
Combinatorial Background
  • WS ?
  • Already used for the optimization
  • Physics can be different
  • Fully reco. B
  • independent emulation of the background
  • Limited statistics
  • Needs some machinery for emulating a semileptonic
    decay!
  • Eliminate the B daughters
  • Replace the B with a semileptonic B with the same
    4-momentum a template montecarlo where the B
    decay comes from EvtGen and the rest of the event
    comes from the data!

16
Background Modeling II
  • Tight cuts (avoid subtractions)
  • Exclude B tracks
  • Replace with MC B
  • QuickCdfObjects/GenTrig
  • Re-decay N times
  • Same analysis path from there on

Lxygt500?m
17
Signal Fits
18
Sample Consistency
19
(No Transcript)
20
Embedded MC vs Semileptonics
MC yield scaled to number of data events
21
Pl
  • Theory prediction depends on Pl cuts. We cannot
    do much but
  • see how our analysis bias looks like
  • Use a threshold-like correction
  • Evaluate systematics for different threshold
    values

22
MC efficiencies
  • ?(M) is dependent on
  • D decay Model
  • Pl cut
  • Use different models/cuts to evaluate systematics

23
MC Validation
  • ? is an unique probe
  • Large statistics
  • Low background
  • Similar spectrum to ?
  • Can reconstruct with minimal cuts (e.g. COT only)
  • Technique
  • Search for ? with very loose cuts
  • Do not include in B vertex
  • Study biases to kinematics from tracking
  • Study IP resolution(data/MC) Primary, B D
    vertices
  • Study ?(data/MC) vs selection criteria

24
MC validation
  • Cross-check kinematic variables
  • B spectrum modeling
  • Trigger emulation
  • Validate CdfSim model of tracking resolution
  • Relative efficiencies
  • ? selection/bias
  • Compare many data/MC distributions using binned
    ?2
  • Every possible decay mode
  • Sideband subtracted before comparison
  • Duplicate removal (D0?K???)

25
Kinematics
  • Can we rely on kinematical biases estimated from
    MC?
  • Rem we dont care about absolute scales
  • Pt dependent MC/data ratio

MC/Data vs Pt
? Pt
MC Data
400 MeV
26
Impact Parameters
148/34
40/33
134/32
26/30
61/30
42/42
151/45
118/43
58/52
27
Impact Parameters (covr)
40/33
146/34
133/32
550/40
39/29
25/30
124/49
137/43
146/45
28
?(MC), ?(data) vs selection criteria
29
Another perspective MC(after/before) /
data(after/before)
30
MC(after/before) / data(after/before)Plan for
the evaluation of systematics
31
Ds Background
  • Use ? peak in D candidates to set the scale
  • Measure the relative contribution of Ds decays to
    D fakes using MC
  • Extrapolate to the total size of the
    contribution 4
  • Build a suitable D background model
  • Subtract

32
Cross-feeds
33
Cross-feeds (details, sat.)
34
Cross-feeds (details, K???)
35
Cross-feeds (details K?)
36
D Moments
  • Combine all events of all types in all channels
    (D,D,SRS,SBRS,feed-down, etc.)
  • Compute mean (m1) and variance (m2) of M2
    distribution with weighted events.
  • Errors and correlation computed with MC (for toy
    MC) or bootstrap (for data).
  • For some realizations, one finds a negative value
    for m2 Var(M2) ltM4gt - ltM2gt2.

37
Inputs for the D0 and D0 Contributions
  • For the BRs, results from charged and neutral B
    decays are combined using isospin partial widths
    are assumed equal.
  • BRs, ratio of lifetimes and ratio of production
    fractions are taken from PDG.
  • Toy Monte Carlo is used to propagate the
    uncertainties from m1, m2, the BRs, etc., to
    uncertainties on M1 and M2 and their correlation.

38
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com