Section 4: Conditional Likelihood: Sufficiency and Ancillarity in the Presence of Nuisance Parameters
|
|
- Λίγεια Κρεστενίτης
- 6 χρόνια πριν
- Προβολές:
Transcript
1 70 Section 4: Conditional Likelihood: Sufficiency and Ancillarity in the Presence of Nuisance Parameters In this section we will: (i) Explore decomposing information in terms of conditional and marginal likelihood parts (ii) Give conditions (ancillarities) under which the conditional likelihood is efficient
2 71 Section 4.1 Partial sufficiency and ancillarity Suppose that X p(x; θ 0 ) P= {p(x; θ) :θ Θ}. We assume θ =(γ,λ),θ=γ Λ, where Γ is an open interval of the real line and Λ is an open set in R k. The parameter of interest is γ and the nuisance parameter is λ. Let T (X) be a statistic. Then, we know that p(x; θ) =h(x T (x);θ)f(t (x);θ) where h is the conditional density of X given T (X) andf is the marginal density of T (X).
3 72 Partial Sufficiency Definition: If h(x T (x); θ) = h(x T (x); λ) and f(t (x); θ) =f(t (x); γ), i.e., p(x; θ) =h(x T (x);λ)f(t (x);γ) then we say that T (X) ispartially sufficient for γ. The term partial is introduced because we must establish sufficiency for each fixed λ. Inference about γ can be made using only the marginal distribution of T (X). There will be no loss of information. If the score for γ is in the class of unbiased estimating functions, then the score is the most efficient member of the class. see Basu (JASA, 1977, pgs ).
4 73 Example 4.1: Partial Sufficiency Suppose that X = X where (1 γ)λ exp(λx) x 0 p(x; θ) = γλexp( λx) x>0 Here Γ = (0, 1) and Λ = R +. Show that T (X) =I(X >0) is partially sufficient for γ. Toseethis,notethat h(x T (x); θ) = (λ exp( λx)) T (x) 1 T (x) (λ exp(λx)) f(t (x); θ) = γ T (x) 1 T (x) (1 γ)
5 74 Partial Ancillarity Suppose h(x T (x); θ) =h(x T (x); γ) andf(t (x); θ) =f(t (x); θ), i.e., p(x; θ) =h(x T (x);γ)f(t (x);θ) Make inference about γ via the conditional distribution. We will make assumptions about the marginal distribution of T (X) in order to guarantee that there will not be information loss.
6 75 Section 4.2 Important types of partial ancillarity S-Ancillarity Definition: T (X) issaidtobes-ancillary for γ if f(t (x); θ) depends only on λ, i.e., p(x; θ) =h(x T (x);γ)f(t (x);λ) This definition is equivalent to letting T (X) be partially sufficient for λ. see Basu (JASA, 1977, pgs ).
7 76 R-Ancillarity Definition: T (X) issaidtober-ancillary for γ if there exists a reparameterization between θ =(γ,λ) and(γ,φ) such that f(t (x); θ) depends on θ only through φ. Thatis, p(x; θ) =h(x T (x);γ)f(t (x);φ) see Basawa (Biometrika, 1981, pgs ).
8 77 C-Ancillarity Definition: T (X) issaidtobec-ancillary for γ if for all γ Γ, the class {f(t (x); γ,λ):λ Λ} is complete. Completeness: E θ [m(t (X); γ)] = 0 for all λ Λ implies P θ [m(t (X); γ) =0]=1forallλ Λ see Godambe (Biometrika, 1976, pgs ).
9 78 A (Weak)-Ancillarity Definition: T (X) issaidtobea-ancillary if for any given θ 0 Θ and any other γ Γ, there exists a λ = λ(γ,θ 0 ) such that If this condition holds, then f(t; θ 0 )=f(t; γ,λ) for all t {f(t (x); γ,λ); λ Λ} isthesamewhateverbethevalueofγ. Intuitively, observation of T (X) cannot give us any information about γ when λ is unknown. see Andersen (JRSS-B, 1970, pgs ).
10 79 Example 4.2: Partial Ancillarity Let X =(Y,Z), where Y and Z are independent normal random variables with variance 1 and means γ and γ + λ, respectively. Here Γ=Λ=R. Let T (X) =Z. Then, we know that Y T (X) N(γ,1) and T (X) N(γ + λ, 1). T (X) isnot S-ancillary for γ. Let φ = γ + λ, sothatt (X) N(φ, 1) Then, T (X) is R-ancillary for γ. For fixed γ, weseethat f(t; θ) = (2π) 1/2 exp( 1 2 (t2 2γt 2λt +(γ + λ) 2 )) = (2π) 1/2 exp( 1 2 (t2 2γt)) exp( 1 2 (γ + λ)2 )exp(λt)
11 80 By exponential family results, we know that for fixed γ {f(t; γ,λ):λ R} is complete. Thus, we know that T (X) is C-ancillary. For fixed θ 0 and given γ, we define λ = λ(γ,θ 0 )= γ + γ 0 + λ 0. Then for all t, we know that So, T (X) is A-ancillary. f(t; θ 0 )=f(t; γ,λ)
12 81 Section 4.3 Information decomposition Information Decomposition Suppose h(x T (x); θ) =h(x T (x); γ) andf(t (x); θ) =f(t (x); θ), i.e., p(x; θ) =h(x T (x);γ)f(t (x);θ) Recall: The Fisher information for γ was defined as I γ (θ) = I γγ (θ) I γλ (θ)i λλ (θ) 1 I λγ (θ) = E θ [(ψ γ (X; θ) Π[ψ γ (X; θ) Λ θ ]) 2 ] The generalized Fisher information for γ from a class of unbiased estimating function G was defined as I γ(θ) =E θ [(ψ γ (X; θ) Π[ψ γ (X; θ) G γ ]) 2 ]
13 82 Let ψγ C (X; γ) andψγ M (X; θ) bethescoresforγ based on the conditional and marginal parts of the factorization of the density of X. Letψ λ (X; θ) =ψλ M (X; θ) be the score for λ from the marginal part of the factorization. Note that the nuisance tangent space Λ θ is the same as the nuisance tangent space from the marginal density of T (X). The Fisher information for γ contained in the conditional distribution h is I C γ (θ) =E θ [ψ C γ (X; γ) 2 ] The generalized Fisher information for γ from a class of unbiased estimating function G and based on the conditional distribution h is I C γ (θ) =E θ [(ψ C γ (X; γ) Π[ψ C γ (X; γ) G γ ) 2 ]
14 83 If ψγ C (X; γ) G γ,theniγ C (θ) =Iγ C (θ). The Fisher information for γ contained in the marginal distribution f is I M γ (θ) =E θ [(ψ M γ (X; θ) Π[ψ M γ (X; θ) Λ θ ]) 2 ] The generalized Fisher information for γ from a class of unbiased estimating function G based on the marginal distribution f is I M γ (θ) =E θ [(ψ M γ (X; θ) Π[ψ M γ (X; θ) G γ ]) 2 ]
15 84 Theorem 4.1: If ψ C γ (X; γ) G γ, I γ(θ) =I C γ Proof: (θ)+iγ M (θ). I γ(θ) = E θ [(ψ γ (X; θ) Π[ψ γ (X; θ) G γ ]) 2 ] = E θ [(ψ C γ (X; γ)+ψ M γ (X; θ) Π[ψ M γ (X; θ) G γ ]) 2 ] = I C γ (θ)+iγ M (θ)+ 2E θ [ψ C γ (X; γ)(ψ M γ (X; θ) Π[ψ M γ (X; θ) G γ ])] = Iγ C (θ)+iγ M (θ)+2e θ [ψγ C (X; γ)ψγ M (X; θ)] = I C γ (θ)+iγ M (θ)
16 85 Theorem 4.2: I γ (θ) =I C γ (θ)+i M γ (θ). Proof: Note that ψ C γ (X; γ) Λ θ. I γ (θ) = E θ [(ψ γ (X; θ) Π[ψ γ (X; θ) Λ θ ]) 2 ] = E θ [(ψγ C (X; γ)+ψγ M (X; θ) Π[ψγ M (X; θ) Λ θ ]) 2 ] = Iγ C (θ)+iγ M (θ)+ 2E θ [ψγ C (X; γ)(ψγ M (X; θ) Π[ψγ M (X; θ) Λ θ ])] = Iγ C (θ)+iγ M (θ)+2e θ [ψγ C (X; γ)ψγ M (X; θ)] = Iγ C (θ)+iγ M (θ) The information decompositions clearly spell out how the information about γ is partitioned between the conditional and marginal parts of the factorization.
17 86 Section 4.4 Optimality of the conditional score under ancillarities What is the efficiency of ψ C γ (X; γ)? Assume that ψ C γ (X; γ) G γ. Eff θ [ψ C γ (X; γ)] E θ[ ψ C γ (X; γ)/ γ] 2 E θ [ψ C γ (X; γ) 2 ] = {E θ[ 2 log h(x T (X); γ)/ γ 2 ]} 2 E θ [( log h(x T (X); γ)/ γ) 2 ] = E θ [ψγ C (X; γ) 2 ]=Iγ C (θ) =Iγ C (θ) Now, we say an UEF g 0 is optimal for γ if Eff θ [g 0 ]=Iγ(θ). The conditional score function will be optimal if Iγ M (θ) = 0. We will now give conditions under which this latter condition holds.
18 87 Lemma 4.3: If T (X) is S-ancillary, then the conditional score function is optimal. Proof: I M γ (θ) =E θ [(ψ M γ (X; θ) Π[ψ M γ (X; θ) G γ ]) 2 ]=0 since ψ M γ (X; θ) =0
19 88 Lemma 4.4: If T (X) is R-ancillary, then the conditional score function is optimal. Proof: It is sufficient to show that Iγ M (θ) =0sinceweknowthat 0 Iγ M (θ) Iγ M (θ). Since T (X) is R-ancillary, we know that there exists a one-to-one reparameterization between θ =(γ,λ) and (γ,φ) such that f(t (x); θ) depends on θ only through φ. Thatis, f(t; θ) =f (t; φ(θ)) Under suitable regularity conditions, we know that ψ M γ (θ) = log f (T (X); φ(θ))/ φ φ(θ)/ γ ψ M λ (θ) = log f (T (X); φ(θ))/ φ φ(θ)/ λ Assuming that φ(θ)/ λ is positive definite, we know that log f (T (X); φ(θ))/ φ = ψ M λ (θ) [ φ(θ)/ λ] 1
20 89 This implies that ψ M γ (θ) = ψ M λ (θ) [ φ(θ)/ λ] 1 φ(θ)/ γ = φ(θ)/ γ [ φ(θ)/ λ] 1 ψ M λ (θ) Note that ψγ M (θ) Λ θ.thisimpliesthatπ[ψγ M (θ) Λ θ ]=ψγ M (θ). So, Iγ M (θ) =E θ [(ψγ M (θ) Π[ψγ M (θ) Λ θ ]) 2 ]=0
21 90 Lemma 4.5: If T (X) is C-ancillary, then the conditional score function is optimal. Proof: It suffices to show that ψ M γ (X; θ) G γ.sincet (X) is C-ancillary, we know that for all γ Γ, the family {f(t; γ,λ):λ Λ} is complete. That is, E θ [m(t (X); γ)] = 0 for all λ Λ P θ [m(t (X); γ) =0]=1forallλ Λ Now, for an UEF g(x; γ), E θ [ψ M γ (X; θ)g(x; γ)] = E θ [ψ M γ (X; θ)e θ [g(x; γ) T (X)]] Note that E θ [g(x; γ) T (X)] is a function of T (X) andγ with mean zero. This implies that E θ [g(x; γ) T (X)] = 0 a.e. So, E θ [ψ M γ (X; θ)g(x; γ)] = 0 which implies that ψ M γ (X; θ) G γ.
22 91 Lemma 4.6: If T (X) is A-ancillary, then it is R-ancillary. Proof: If T (X) is A-ancillary, then for any given θ 0 Θandany other γ Γ, there exists a λ = λ(γ,θ 0 ) such that f(t; θ 0 )=f(t; γ,λ(γ,θ 0 )) for all t For any γ, we know that the distribution of T (X) depends only on φ = λ(γ,θ 0 ). So, there is a transformation between θ 0 and (γ,φ) such that f(t (x); θ) depends only on φ. Thatis,T (X) is R-ancillary. Corollary 4.7: If T (X) is A-ancillary, then the conditional score function is optimal. Proof: If T (X) is A-ancillary, then it is R-ancillary. R-ancillarity implies that the conditional score is optimal.
23 92 Four Examples of Ancillarity Example 4.3: Let X =(Y 1,Y 2 ) be independent Poisson random variables with means µ 1 and µ 2.Letγ = µ 1 µ 1 +µ 2 and λ = µ 1 + µ 2. Let T (X) =Y 1 + Y 2. Show that T (X) is S-ancillary for γ. We know that T (X) P oisson(λ). The conditional distribution of X given T (X) isequalto
24 93 h(x T (x); θ) = P [Y 1 = y 1,Y 2 = y 2,T(X) =t] P [T (X) =t] = P [Y 1 = y 1,Y 2 = t y 1 ]I(y 1 t) P [Y 1 + Y 2 = t] = exp( µ 1)µ y 1 1 exp( µ 2)µ t y 1 2 I(y 1 t)/y 1!(t y 1 )! exp( µ 1 µ 2 )(µ 1 + µ 2 ) t /t! = = y1 t y1 µ1 µ2 µ 1 + µ 2 µ 1 + µ 2 t! y 1!(t y 1 )! γy 1 (1 γ) t y 1 I(y 1 t) t! y 1!(t y 1 )! I(y 1 t) So, p(x; θ) =h(x T (x);γ)f(t (x);λ) and T (X) is S-ancillary.
25 94 Example 4.4: Let X =(X 1,X 2,...,X n ) be i.i.d. Normal random variables with mean µ and variance σ 2.Letγ = σ 2 and λ = µ. Let T (X) =X. We know that X N(λ, γ/n). By exponential family results, we know that for fixed γ, T (X) is complete for λ. The conditional distribution of X given T (X) isequalto h(x T (X) =t; θ) = (2πγ) n/2 exp( P n i=1 (x i λ) 2 /(2γ))I( P n i=1 x i = nt) ( n 2πγ )1/2 exp( n(t λ) 2 /(2γ)) = (2πγ) (n 1)/2 n 1/2 exp( ( nx i=1 x 2 i nt 2 )/(2γ)) So, p(x; θ) =h(x T (x);γ)f(t (x);θ) and T (X) is C-ancillary.
26 Example 4.5: Suppose that X 1 Binomial(n 1,p 1 )and X 2 Binomial(n 2,p 2 )wheren 1 and n 2 are fixed sample sizes. We assume X 1 is independent of X 2.LetX =(X 1,X 2 ), q 1 =1 p 1, q 2 =1 p 2, γ =log( p 2/q 2 p 1 /q 1 )andλ =log(p 1 /q 1 ). There is a one-to-one mapping between (p 1,p 2 )and(γ,λ). Let T (X) =X 1 + X 2. What is the distribution of T (X)? 95
27 96 f(t; θ) = = min(t,n 2 ) X u=0 min(t,n 2 ) X u=max(0,t n 1 ) P [X 1 + X 2 = t X 2 = u]p [X 2 = u] P [X 1 = t u]p [X 2 = u] = min(t,n 2 ) X u=max(0,t n 1 ) t u 0 n 1 0 A p t u 1 q n 1 t+u n 2 u 1 A p u 2 q n 2 u 2 = min(t,n 2 ) X u=max(0,t n 1 ) t u 0 n u 1 n 0 A 1 A exp(γu + λt)q n 1 1 q n 2 2 T (X) is C-ancillary, so the conditional score function is optimal. T (X) is not S (obvious), R, A- ancillary.
28 97 To be thorough, we compute the conditional density of X given T (X). This is given by h(x T (X) =t; θ) = P [X 1 = x 1,X 2 = x 2,T(X) =t] P [T (X) =t] = = x 1 0 n x 2 1 n 0 A 0 P min(t,n2 ) 1 u=max(0,t n 1 ) t u 0 n x 1 1 A 0 n 0 P min(t,n2 ) 1 u=max(0,t n 1 ) t u 1 A exp(γx2 + λt)q n 1 1 qn 2 2 u 1 n 0 A 1 A exp(γx2 ) x 2 u 1 n 0 A 1 A exp(γu + λt)q n 1 1 qn A exp(γu)
29 98 One advantage of using the conditional score function is that we may consider situations where there is an infinite-dimensional nuisance parameter (i.e., a semiparametric model). Suppose that there are n independent tables. Let λ i be the baseline log odds for the ith table, but let γ be the common odds ratio across the n tables. Assume that λ i L. Inthiscase,the parameters are (γ,l). Let X =(X 1,...,X n ), T (X) =(T 1 (X 1 ),...,T n (X n )), T i (X i )=X 1i + X 2i, X i =(X 1i,X 2i ), and X 1i and X 2i be independent Binomial random variables with fixed sample sizes n 1i and n 2i and random success probabilities p 1i and p 2i, respectively. Let q 1i =1 p 1i and q 2i =1 p 2i. So, we know that λ i =log(p 1i /q 1i )and γ =log( p 2i/q 2i p 1i /q 1i ) for all i. The conditional distribution of X given T (X) doesn t depend on L.
30 Example 4.6: Consider the semiparametric truncation model. Let Y and T be independent non-negative random variables. Suppose that Y k(k) andt l(l). In the right truncation problem, (Y,T ) is only observed if Y T. Lagakos (Biometrika, 1988) describes a study population of subjects infected with the HIV virus from contaminated blood transfusions. The date of infection was known for all subjects and only those subjects who contracted AIDS by a fixed date were included in the analysis. Interest is in the incubation time of AIDS, i.e., time from infection to AIDS. Let Y denote the incubation time and T be the time from infection to the fixed cutoff date. Note that Y and T are only observed if Y T. 99
31 100 Let X =(Y,T) betheobserved(y,t ). Then, p(x) =P [Y = y, T = t] =P [Y = y, T = t Y T ]= k(y)l(t)i(y t) β where P [Y T ]=β = 0 K(t)l(t)dt = 0 (1 L(y))k(y)dy Let T (X) =T be the conditional statistic. Then h(x T (x) =t) =P [Y = y T = t] =P [Y = y T = t, Y T ]= k(y)i(y t) K(t) and f(t) =P [T = t] =P [T = t Y T ]= K(t)l(t) β
32 101 Suppose that we parameterize the law of Y via a parameter γ and leave the distribution of T unspecified. So, we have a semiparametric model with parameters (γ,l). So, p(x; γ,l)= k γ(y)i(y t) K γ (t) K γ (t)l(t) β γ,l where β γ,l = 0 K γ (t)l(t)dt = 0 (1 L(y))k γ (y)dy
33 102 Claim 4.8: T (X) is A-ancillary for γ. Proof: For any given (γ 0,l 0 )andanygivenγ, thereexists l = l(γ,γ 0,l 0 ) such that f(t; γ 0,l 0 )=f(t; γ,l(t; γ,γ 0,l 0 ) for all t, where f(t; γ,l)= K γ(t)l(t) β γ,l If we take where l(t; γ,γ 0,l 0 )= K γ 0 (t)l 0 (t) K γ (t)c c = 0 K γ0 (t)l 0 (t) dt K γ (t) then the above equality holds. This implies that the conditional score function is optimal.
ST5224: Advanced Statistical Theory II
ST5224: Advanced Statistical Theory II 2014/2015: Semester II Tutorial 7 1. Let X be a sample from a population P and consider testing hypotheses H 0 : P = P 0 versus H 1 : P = P 1, where P j is a known
Other Test Constructions: Likelihood Ratio & Bayes Tests
Other Test Constructions: Likelihood Ratio & Bayes Tests Side-Note: So far we have seen a few approaches for creating tests such as Neyman-Pearson Lemma ( most powerful tests of H 0 : θ = θ 0 vs H 1 :
Statistical Inference I Locally most powerful tests
Statistical Inference I Locally most powerful tests Shirsendu Mukherjee Department of Statistics, Asutosh College, Kolkata, India. shirsendu st@yahoo.co.in So far we have treated the testing of one-sided
Example Sheet 3 Solutions
Example Sheet 3 Solutions. i Regular Sturm-Liouville. ii Singular Sturm-Liouville mixed boundary conditions. iii Not Sturm-Liouville ODE is not in Sturm-Liouville form. iv Regular Sturm-Liouville note
C.S. 430 Assignment 6, Sample Solutions
C.S. 430 Assignment 6, Sample Solutions Paul Liu November 15, 2007 Note that these are sample solutions only; in many cases there were many acceptable answers. 1 Reynolds Problem 10.1 1.1 Normal-order
2 Composition. Invertible Mappings
Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan Composition. Invertible Mappings In this section we discuss two procedures for creating new mappings from old ones, namely,
Every set of first-order formulas is equivalent to an independent set
Every set of first-order formulas is equivalent to an independent set May 6, 2008 Abstract A set of first-order formulas, whatever the cardinality of the set of symbols, is equivalent to an independent
Problem Set 3: Solutions
CMPSCI 69GG Applied Information Theory Fall 006 Problem Set 3: Solutions. [Cover and Thomas 7.] a Define the following notation, C I p xx; Y max X; Y C I p xx; Ỹ max I X; Ỹ We would like to show that C
Uniform Convergence of Fourier Series Michael Taylor
Uniform Convergence of Fourier Series Michael Taylor Given f L 1 T 1 ), we consider the partial sums of the Fourier series of f: N 1) S N fθ) = ˆfk)e ikθ. k= N A calculation gives the Dirichlet formula
6. MAXIMUM LIKELIHOOD ESTIMATION
6 MAXIMUM LIKELIHOOD ESIMAION [1] Maximum Likelihood Estimator (1) Cases in which θ (unknown parameter) is scalar Notational Clarification: From now on, we denote the true value of θ as θ o hen, view θ
Phys460.nb Solution for the t-dependent Schrodinger s equation How did we find the solution? (not required)
Phys460.nb 81 ψ n (t) is still the (same) eigenstate of H But for tdependent H. The answer is NO. 5.5.5. Solution for the tdependent Schrodinger s equation If we assume that at time t 0, the electron starts
5.4 The Poisson Distribution.
The worst thing you can do about a situation is nothing. Sr. O Shea Jackson 5.4 The Poisson Distribution. Description of the Poisson Distribution Discrete probability distribution. The random variable
Lecture 21: Properties and robustness of LSE
Lecture 21: Properties and robustness of LSE BLUE: Robustness of LSE against normality We now study properties of l τ β and σ 2 under assumption A2, i.e., without the normality assumption on ε. From Theorem
SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM
SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question 1 a) The cumulative distribution function of T conditional on N n is Pr T t N n) Pr max X 1,..., X N ) t N n) Pr max
Reminders: linear functions
Reminders: linear functions Let U and V be vector spaces over the same field F. Definition A function f : U V is linear if for every u 1, u 2 U, f (u 1 + u 2 ) = f (u 1 ) + f (u 2 ), and for every u U
Econ 2110: Fall 2008 Suggested Solutions to Problem Set 8 questions or comments to Dan Fetter 1
Eon : Fall 8 Suggested Solutions to Problem Set 8 Email questions or omments to Dan Fetter Problem. Let X be a salar with density f(x, θ) (θx + θ) [ x ] with θ. (a) Find the most powerful level α test
Chapter 6: Systems of Linear Differential. be continuous functions on the interval
Chapter 6: Systems of Linear Differential Equations Let a (t), a 2 (t),..., a nn (t), b (t), b 2 (t),..., b n (t) be continuous functions on the interval I. The system of n first-order differential equations
Solution Series 9. i=1 x i and i=1 x i.
Lecturer: Prof. Dr. Mete SONER Coordinator: Yilin WANG Solution Series 9 Q1. Let α, β >, the p.d.f. of a beta distribution with parameters α and β is { Γ(α+β) Γ(α)Γ(β) f(x α, β) xα 1 (1 x) β 1 for < x
Theorem 8 Let φ be the most powerful size α test of H
Testing composite hypotheses Θ = Θ 0 Θ c 0 H 0 : θ Θ 0 H 1 : θ Θ c 0 Definition 16 A test φ is a uniformly most powerful (UMP) level α test for H 0 vs. H 1 if φ has level α and for any other level α test
Ordinal Arithmetic: Addition, Multiplication, Exponentiation and Limit
Ordinal Arithmetic: Addition, Multiplication, Exponentiation and Limit Ting Zhang Stanford May 11, 2001 Stanford, 5/11/2001 1 Outline Ordinal Classification Ordinal Addition Ordinal Multiplication Ordinal
Section 8.3 Trigonometric Equations
99 Section 8. Trigonometric Equations Objective 1: Solve Equations Involving One Trigonometric Function. In this section and the next, we will exple how to solving equations involving trigonometric functions.
EE512: Error Control Coding
EE512: Error Control Coding Solution for Assignment on Finite Fields February 16, 2007 1. (a) Addition and Multiplication tables for GF (5) and GF (7) are shown in Tables 1 and 2. + 0 1 2 3 4 0 0 1 2 3
Lecture 2: Dirac notation and a review of linear algebra Read Sakurai chapter 1, Baym chatper 3
Lecture 2: Dirac notation and a review of linear algebra Read Sakurai chapter 1, Baym chatper 3 1 State vector space and the dual space Space of wavefunctions The space of wavefunctions is the set of all
Numerical Analysis FMN011
Numerical Analysis FMN011 Carmen Arévalo Lund University carmen@maths.lth.se Lecture 12 Periodic data A function g has period P if g(x + P ) = g(x) Model: Trigonometric polynomial of order M T M (x) =
Lecture 2. Soundness and completeness of propositional logic
Lecture 2 Soundness and completeness of propositional logic February 9, 2004 1 Overview Review of natural deduction. Soundness and completeness. Semantics of propositional formulas. Soundness proof. Completeness
5. Choice under Uncertainty
5. Choice under Uncertainty Daisuke Oyama Microeconomics I May 23, 2018 Formulations von Neumann-Morgenstern (1944/1947) X: Set of prizes Π: Set of probability distributions on X : Preference relation
SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM
SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question 1 a) The cumulative distribution function of T conditional on N n is Pr (T t N n) Pr (max (X 1,..., X N ) t N n) Pr (max
STAT200C: Hypothesis Testing
STAT200C: Hypothesis Testing Zhaoxia Yu Spring 2017 Some Definitions A hypothesis is a statement about a population parameter. The two complementary hypotheses in a hypothesis testing are the null hypothesis
Chapter 6: Systems of Linear Differential. be continuous functions on the interval
Chapter 6: Systems of Linear Differential Equations Let a (t), a 2 (t),..., a nn (t), b (t), b 2 (t),..., b n (t) be continuous functions on the interval I. The system of n first-order differential equations
= λ 1 1 e. = λ 1 =12. has the properties e 1. e 3,V(Y
Stat 50 Homework Solutions Spring 005. (a λ λ λ 44 (b trace( λ + λ + λ 0 (c V (e x e e λ e e λ e (λ e by definition, the eigenvector e has the properties e λ e and e e. (d λ e e + λ e e + λ e e 8 6 4 4
Solutions to Exercise Sheet 5
Solutions to Eercise Sheet 5 jacques@ucsd.edu. Let X and Y be random variables with joint pdf f(, y) = 3y( + y) where and y. Determine each of the following probabilities. Solutions. a. P (X ). b. P (X
4.6 Autoregressive Moving Average Model ARMA(1,1)
84 CHAPTER 4. STATIONARY TS MODELS 4.6 Autoregressive Moving Average Model ARMA(,) This section is an introduction to a wide class of models ARMA(p,q) which we will consider in more detail later in this
Fractional Colorings and Zykov Products of graphs
Fractional Colorings and Zykov Products of graphs Who? Nichole Schimanski When? July 27, 2011 Graphs A graph, G, consists of a vertex set, V (G), and an edge set, E(G). V (G) is any finite set E(G) is
Μηχανική Μάθηση Hypothesis Testing
ΕΛΛΗΝΙΚΗ ΔΗΜΟΚΡΑΤΙΑ ΠΑΝΕΠΙΣΤΗΜΙΟ ΚΡΗΤΗΣ Μηχανική Μάθηση Hypothesis Testing Γιώργος Μπορμπουδάκης Τμήμα Επιστήμης Υπολογιστών Procedure 1. Form the null (H 0 ) and alternative (H 1 ) hypothesis 2. Consider
Finite Field Problems: Solutions
Finite Field Problems: Solutions 1. Let f = x 2 +1 Z 11 [x] and let F = Z 11 [x]/(f), a field. Let Solution: F =11 2 = 121, so F = 121 1 = 120. The possible orders are the divisors of 120. Solution: The
Mean-Variance Analysis
Mean-Variance Analysis Jan Schneider McCombs School of Business University of Texas at Austin Jan Schneider Mean-Variance Analysis Beta Representation of the Risk Premium risk premium E t [Rt t+τ ] R1
Math 446 Homework 3 Solutions. (1). (i): Reverse triangle inequality for metrics: Let (X, d) be a metric space and let x, y, z X.
Math 446 Homework 3 Solutions. (1). (i): Reverse triangle inequalit for metrics: Let (X, d) be a metric space and let x,, z X. Prove that d(x, z) d(, z) d(x, ). (ii): Reverse triangle inequalit for norms:
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture 7: Information bound Lecturer: Yihong Wu Scribe: Shiyu Liang, Feb 6, 06 [Ed. Mar 9] Recall the Chi-squared divergence
Inverse trigonometric functions & General Solution of Trigonometric Equations. ------------------ ----------------------------- -----------------
Inverse trigonometric functions & General Solution of Trigonometric Equations. 1. Sin ( ) = a) b) c) d) Ans b. Solution : Method 1. Ans a: 17 > 1 a) is rejected. w.k.t Sin ( sin ) = d is rejected. If sin
CHAPTER 25 SOLVING EQUATIONS BY ITERATIVE METHODS
CHAPTER 5 SOLVING EQUATIONS BY ITERATIVE METHODS EXERCISE 104 Page 8 1. Find the positive root of the equation x + 3x 5 = 0, correct to 3 significant figures, using the method of bisection. Let f(x) =
Lecture 7: Overdispersion in Poisson regression
Lecture 7: Overdispersion in Poisson regression Claudia Czado TU München c (Claudia Czado, TU Munich) ZFS/IMS Göttingen 2004 0 Overview Introduction Modeling overdispersion through mixing Score test for
Congruence Classes of Invertible Matrices of Order 3 over F 2
International Journal of Algebra, Vol. 8, 24, no. 5, 239-246 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.2988/ija.24.422 Congruence Classes of Invertible Matrices of Order 3 over F 2 Ligong An and
Exercises to Statistics of Material Fatigue No. 5
Prof. Dr. Christine Müller Dipl.-Math. Christoph Kustosz Eercises to Statistics of Material Fatigue No. 5 E. 9 (5 a Show, that a Fisher information matri for a two dimensional parameter θ (θ,θ 2 R 2, can
Parametrized Surfaces
Parametrized Surfaces Recall from our unit on vector-valued functions at the beginning of the semester that an R 3 -valued function c(t) in one parameter is a mapping of the form c : I R 3 where I is some
Homework 8 Model Solution Section
MATH 004 Homework Solution Homework 8 Model Solution Section 14.5 14.6. 14.5. Use the Chain Rule to find dz where z cosx + 4y), x 5t 4, y 1 t. dz dx + dy y sinx + 4y)0t + 4) sinx + 4y) 1t ) 0t + 4t ) sinx
Homework 3 Solutions
Homework 3 Solutions Igor Yanovsky (Math 151A TA) Problem 1: Compute the absolute error and relative error in approximations of p by p. (Use calculator!) a) p π, p 22/7; b) p π, p 3.141. Solution: For
6.3 Forecasting ARMA processes
122 CHAPTER 6. ARMA MODELS 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss a linear
Concrete Mathematics Exercises from 30 September 2016
Concrete Mathematics Exercises from 30 September 2016 Silvio Capobianco Exercise 1.7 Let H(n) = J(n + 1) J(n). Equation (1.8) tells us that H(2n) = 2, and H(2n+1) = J(2n+2) J(2n+1) = (2J(n+1) 1) (2J(n)+1)
SCHOOL OF MATHEMATICAL SCIENCES G11LMA Linear Mathematics Examination Solutions
SCHOOL OF MATHEMATICAL SCIENCES GLMA Linear Mathematics 00- Examination Solutions. (a) i. ( + 5i)( i) = (6 + 5) + (5 )i = + i. Real part is, imaginary part is. (b) ii. + 5i i ( + 5i)( + i) = ( i)( + i)
Math221: HW# 1 solutions
Math: HW# solutions Andy Royston October, 5 7.5.7, 3 rd Ed. We have a n = b n = a = fxdx = xdx =, x cos nxdx = x sin nx n sin nxdx n = cos nx n = n n, x sin nxdx = x cos nx n + cos nxdx n cos n = + sin
Jesse Maassen and Mark Lundstrom Purdue University November 25, 2013
Notes on Average Scattering imes and Hall Factors Jesse Maassen and Mar Lundstrom Purdue University November 5, 13 I. Introduction 1 II. Solution of the BE 1 III. Exercises: Woring out average scattering
Απόκριση σε Μοναδιαία Ωστική Δύναμη (Unit Impulse) Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο. Απόστολος Σ.
Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο The time integral of a force is referred to as impulse, is determined by and is obtained from: Newton s 2 nd Law of motion states that the action
Bounding Nonsplitting Enumeration Degrees
Bounding Nonsplitting Enumeration Degrees Thomas F. Kent Andrea Sorbi Università degli Studi di Siena Italia July 18, 2007 Goal: Introduce a form of Σ 0 2-permitting for the enumeration degrees. Till now,
Lecture 12: Pseudo likelihood approach
Lecture 12: Pseudo likelihood approach Pseudo MLE Let X 1,...,X n be a random sample from a pdf in a family indexed by two parameters θ and π with likelihood l(θ,π). The method of pseudo MLE may be viewed
b. Use the parametrization from (a) to compute the area of S a as S a ds. Be sure to substitute for ds!
MTH U341 urface Integrals, tokes theorem, the divergence theorem To be turned in Wed., Dec. 1. 1. Let be the sphere of radius a, x 2 + y 2 + z 2 a 2. a. Use spherical coordinates (with ρ a) to parametrize.
6.1. Dirac Equation. Hamiltonian. Dirac Eq.
6.1. Dirac Equation Ref: M.Kaku, Quantum Field Theory, Oxford Univ Press (1993) η μν = η μν = diag(1, -1, -1, -1) p 0 = p 0 p = p i = -p i p μ p μ = p 0 p 0 + p i p i = E c 2 - p 2 = (m c) 2 H = c p 2
Nowhere-zero flows Let be a digraph, Abelian group. A Γ-circulation in is a mapping : such that, where, and : tail in X, head in
Nowhere-zero flows Let be a digraph, Abelian group. A Γ-circulation in is a mapping : such that, where, and : tail in X, head in : tail in X, head in A nowhere-zero Γ-flow is a Γ-circulation such that
A Two-Sided Laplace Inversion Algorithm with Computable Error Bounds and Its Applications in Financial Engineering
Electronic Companion A Two-Sie Laplace Inversion Algorithm with Computable Error Bouns an Its Applications in Financial Engineering Ning Cai, S. G. Kou, Zongjian Liu HKUST an Columbia University Appenix
Approximation of distance between locations on earth given by latitude and longitude
Approximation of distance between locations on earth given by latitude and longitude Jan Behrens 2012-12-31 In this paper we shall provide a method to approximate distances between two points on earth
The Simply Typed Lambda Calculus
Type Inference Instead of writing type annotations, can we use an algorithm to infer what the type annotations should be? That depends on the type system. For simple type systems the answer is yes, and
Matrices and Determinants
Matrices and Determinants SUBJECTIVE PROBLEMS: Q 1. For what value of k do the following system of equations possess a non-trivial (i.e., not all zero) solution over the set of rationals Q? x + ky + 3z
HW 3 Solutions 1. a) I use the auto.arima R function to search over models using AIC and decide on an ARMA(3,1)
HW 3 Solutions a) I use the autoarima R function to search over models using AIC and decide on an ARMA3,) b) I compare the ARMA3,) to ARMA,0) ARMA3,) does better in all three criteria c) The plot of the
ORDINAL ARITHMETIC JULIAN J. SCHLÖDER
ORDINAL ARITHMETIC JULIAN J. SCHLÖDER Abstract. We define ordinal arithmetic and show laws of Left- Monotonicity, Associativity, Distributivity, some minor related properties and the Cantor Normal Form.
Fourier Series. MATH 211, Calculus II. J. Robert Buchanan. Spring Department of Mathematics
Fourier Series MATH 211, Calculus II J. Robert Buchanan Department of Mathematics Spring 2018 Introduction Not all functions can be represented by Taylor series. f (k) (c) A Taylor series f (x) = (x c)
Lecture 34 Bootstrap confidence intervals
Lecture 34 Bootstrap confidence intervals Confidence Intervals θ: an unknown parameter of interest We want to find limits θ and θ such that Gt = P nˆθ θ t If G 1 1 α is known, then P θ θ = P θ θ = 1 α
HOMEWORK 4 = G. In order to plot the stress versus the stretch we define a normalized stretch:
HOMEWORK 4 Problem a For the fast loading case, we want to derive the relationship between P zz and λ z. We know that the nominal stress is expressed as: P zz = ψ λ z where λ z = λ λ z. Therefore, applying
557: MATHEMATICAL STATISTICS II RESULTS FROM CLASSICAL HYPOTHESIS TESTING
Most Powerful Tests 557: MATHEMATICAL STATISTICS II RESULTS FROM CLASSICAL HYPOTHESIS TESTING To construct and assess the quality of a statistical test, we consider the power function β(θ). Consider a
Section 7.6 Double and Half Angle Formulas
09 Section 7. Double and Half Angle Fmulas To derive the double-angles fmulas, we will use the sum of two angles fmulas that we developed in the last section. We will let α θ and β θ: cos(θ) cos(θ + θ)
MINIMAL CLOSED SETS AND MAXIMAL CLOSED SETS
MINIMAL CLOSED SETS AND MAXIMAL CLOSED SETS FUMIE NAKAOKA AND NOBUYUKI ODA Received 20 December 2005; Revised 28 May 2006; Accepted 6 August 2006 Some properties of minimal closed sets and maximal closed
Abstract Storage Devices
Abstract Storage Devices Robert König Ueli Maurer Stefano Tessaro SOFSEM 2009 January 27, 2009 Outline 1. Motivation: Storage Devices 2. Abstract Storage Devices (ASD s) 3. Reducibility 4. Factoring ASD
HOMEWORK#1. t E(x) = 1 λ = (b) Find the median lifetime of a randomly selected light bulb. Answer:
HOMEWORK# 52258 李亞晟 Eercise 2. The lifetime of light bulbs follows an eponential distribution with a hazard rate of. failures per hour of use (a) Find the mean lifetime of a randomly selected light bulb.
3.4 SUM AND DIFFERENCE FORMULAS. NOTE: cos(α+β) cos α + cos β cos(α-β) cos α -cos β
3.4 SUM AND DIFFERENCE FORMULAS Page Theorem cos(αβ cos α cos β -sin α cos(α-β cos α cos β sin α NOTE: cos(αβ cos α cos β cos(α-β cos α -cos β Proof of cos(α-β cos α cos β sin α Let s use a unit circle
derivation of the Laplacian from rectangular to spherical coordinates
derivation of the Laplacian from rectangular to spherical coordinates swapnizzle 03-03- :5:43 We begin by recognizing the familiar conversion from rectangular to spherical coordinates (note that φ is used
2. Let H 1 and H 2 be Hilbert spaces and let T : H 1 H 2 be a bounded linear operator. Prove that [T (H 1 )] = N (T ). (6p)
Uppsala Universitet Matematiska Institutionen Andreas Strömbergsson Prov i matematik Funktionalanalys Kurs: F3B, F4Sy, NVP 2005-03-08 Skrivtid: 9 14 Tillåtna hjälpmedel: Manuella skrivdon, Kreyszigs bok
The challenges of non-stable predicates
The challenges of non-stable predicates Consider a non-stable predicate Φ encoding, say, a safety property. We want to determine whether Φ holds for our program. The challenges of non-stable predicates
Chapter 3: Ordinal Numbers
Chapter 3: Ordinal Numbers There are two kinds of number.. Ordinal numbers (0th), st, 2nd, 3rd, 4th, 5th,..., ω, ω +,... ω2, ω2+,... ω 2... answers to the question What position is... in a sequence? What
Math 6 SL Probability Distributions Practice Test Mark Scheme
Math 6 SL Probability Distributions Practice Test Mark Scheme. (a) Note: Award A for vertical line to right of mean, A for shading to right of their vertical line. AA N (b) evidence of recognizing symmetry
MATH423 String Theory Solutions 4. = 0 τ = f(s). (1) dτ ds = dxµ dτ f (s) (2) dτ 2 [f (s)] 2 + dxµ. dτ f (s) (3)
1. MATH43 String Theory Solutions 4 x = 0 τ = fs). 1) = = f s) ) x = x [f s)] + f s) 3) equation of motion is x = 0 if an only if f s) = 0 i.e. fs) = As + B with A, B constants. i.e. allowe reparametrisations
Tridiagonal matrices. Gérard MEURANT. October, 2008
Tridiagonal matrices Gérard MEURANT October, 2008 1 Similarity 2 Cholesy factorizations 3 Eigenvalues 4 Inverse Similarity Let α 1 ω 1 β 1 α 2 ω 2 T =......... β 2 α 1 ω 1 β 1 α and β i ω i, i = 1,...,
F A S C I C U L I M A T H E M A T I C I
F A S C I C U L I M A T H E M A T I C I Nr 46 2011 C. Carpintero, N. Rajesh and E. Rosas ON A CLASS OF (γ, γ )-PREOPEN SETS IN A TOPOLOGICAL SPACE Abstract. In this paper we have introduced the concept
k A = [k, k]( )[a 1, a 2 ] = [ka 1,ka 2 ] 4For the division of two intervals of confidence in R +
Chapter 3. Fuzzy Arithmetic 3- Fuzzy arithmetic: ~Addition(+) and subtraction (-): Let A = [a and B = [b, b in R If x [a and y [b, b than x+y [a +b +b Symbolically,we write A(+)B = [a (+)[b, b = [a +b
An Introduction to Signal Detection and Estimation - Second Edition Chapter II: Selected Solutions
An Introduction to Signal Detection Estimation - Second Edition Chapter II: Selected Solutions H V Poor Princeton University March 16, 5 Exercise : The likelihood ratio is given by L(y) (y +1), y 1 a With
SCITECH Volume 13, Issue 2 RESEARCH ORGANISATION Published online: March 29, 2018
Journal of rogressive Research in Mathematics(JRM) ISSN: 2395-028 SCITECH Volume 3, Issue 2 RESEARCH ORGANISATION ublished online: March 29, 208 Journal of rogressive Research in Mathematics www.scitecresearch.com/journals
Lecture 15 - Root System Axiomatics
Lecture 15 - Root System Axiomatics Nov 1, 01 In this lecture we examine root systems from an axiomatic point of view. 1 Reflections If v R n, then it determines a hyperplane, denoted P v, through the
A General Note on δ-quasi Monotone and Increasing Sequence
International Mathematical Forum, 4, 2009, no. 3, 143-149 A General Note on δ-quasi Monotone and Increasing Sequence Santosh Kr. Saxena H. N. 419, Jawaharpuri, Badaun, U.P., India Presently working in
Second Order Partial Differential Equations
Chapter 7 Second Order Partial Differential Equations 7.1 Introduction A second order linear PDE in two independent variables (x, y Ω can be written as A(x, y u x + B(x, y u xy + C(x, y u u u + D(x, y
Bayesian statistics. DS GA 1002 Probability and Statistics for Data Science.
Bayesian statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Frequentist vs Bayesian statistics In frequentist
PARTIAL NOTES for 6.1 Trigonometric Identities
PARTIAL NOTES for 6.1 Trigonometric Identities tanθ = sinθ cosθ cotθ = cosθ sinθ BASIC IDENTITIES cscθ = 1 sinθ secθ = 1 cosθ cotθ = 1 tanθ PYTHAGOREAN IDENTITIES sin θ + cos θ =1 tan θ +1= sec θ 1 + cot
Survival Analysis: One-Sample Problem /Two-Sample Problem/Regression. Lu Tian and Richard Olshen Stanford University
Survival Analysis: One-Sample Problem /Two-Sample Problem/Regression Lu Tian and Richard Olshen Stanford University 1 One sample problem T 1,, T n 1 S( ), C 1,, C n G( ) and T i C i Observations: (U i,
Partial Differential Equations in Biology The boundary element method. March 26, 2013
The boundary element method March 26, 203 Introduction and notation The problem: u = f in D R d u = ϕ in Γ D u n = g on Γ N, where D = Γ D Γ N, Γ D Γ N = (possibly, Γ D = [Neumann problem] or Γ N = [Dirichlet
Exercise 2: The form of the generalized likelihood ratio
Stats 2 Winter 28 Homework 9: Solutions Due Friday, March 6 Exercise 2: The form of the generalized likelihood ratio We want to test H : θ Θ against H : θ Θ, and compare the two following rules of rejection:
Optimal Parameter in Hermitian and Skew-Hermitian Splitting Method for Certain Two-by-Two Block Matrices
Optimal Parameter in Hermitian and Skew-Hermitian Splitting Method for Certain Two-by-Two Block Matrices Chi-Kwong Li Department of Mathematics The College of William and Mary Williamsburg, Virginia 23187-8795
Introduction to the ML Estimation of ARMA processes
Introduction to the ML Estimation of ARMA processes Eduardo Rossi University of Pavia October 2013 Rossi ARMA Estimation Financial Econometrics - 2013 1 / 1 We consider the AR(p) model: Y t = c + φ 1 Y
Section 9.2 Polar Equations and Graphs
180 Section 9. Polar Equations and Graphs In this section, we will be graphing polar equations on a polar grid. In the first few examples, we will write the polar equation in rectangular form to help identify
Lecture 13 - Root Space Decomposition II
Lecture 13 - Root Space Decomposition II October 18, 2012 1 Review First let us recall the situation. Let g be a simple algebra, with maximal toral subalgebra h (which we are calling a CSA, or Cartan Subalgebra).
ω ω ω ω ω ω+2 ω ω+2 + ω ω ω ω+2 + ω ω+1 ω ω+2 2 ω ω ω ω ω ω ω ω+1 ω ω2 ω ω2 + ω ω ω2 + ω ω ω ω2 + ω ω+1 ω ω2 + ω ω+1 + ω ω ω ω2 + ω
0 1 2 3 4 5 6 ω ω + 1 ω + 2 ω + 3 ω + 4 ω2 ω2 + 1 ω2 + 2 ω2 + 3 ω3 ω3 + 1 ω3 + 2 ω4 ω4 + 1 ω5 ω 2 ω 2 + 1 ω 2 + 2 ω 2 + ω ω 2 + ω + 1 ω 2 + ω2 ω 2 2 ω 2 2 + 1 ω 2 2 + ω ω 2 3 ω 3 ω 3 + 1 ω 3 + ω ω 3 +
Notes on the Open Economy
Notes on the Open Econom Ben J. Heijdra Universit of Groningen April 24 Introduction In this note we stud the two-countr model of Table.4 in more detail. restated here for convenience. The model is Table.4.
A Bonus-Malus System as a Markov Set-Chain. Małgorzata Niemiec Warsaw School of Economics Institute of Econometrics
A Bonus-Malus System as a Markov Set-Chain Małgorzata Niemiec Warsaw School of Economics Institute of Econometrics Contents 1. Markov set-chain 2. Model of bonus-malus system 3. Example 4. Conclusions
557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING
557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING A statistical hypothesis test is a decision rule that takes as an input observed sample data and returns an action relating to two mutually exclusive
ENGR 691/692 Section 66 (Fall 06): Machine Learning Assigned: August 30 Homework 1: Bayesian Decision Theory (solutions) Due: September 13
ENGR 69/69 Section 66 (Fall 06): Machine Learning Assigned: August 30 Homework : Bayesian Decision Theory (solutions) Due: Septemer 3 Prolem : ( pts) Let the conditional densities for a two-category one-dimensional