Fitting mixtures of linear regressions

Μέγεθος: px
Εμφάνιση ξεκινά από τη σελίδα:

Download "Fitting mixtures of linear regressions"

Transcript

1 Journal of Statistical Computation and Simulation Vol. 80, No. 2, February 2010, Fitting mixtures of linear regressions Susana Faria a * and Gilda Soromenho b a Department of Mathematics for Science and Technology, Research CMAT, University of Minho, Guimarães, Portugal; b Faculty of Psychology and Sciences of Education, Research LEAD, University of Lisbon, Lisbon, Portugal (Received 1 April 2008; final version received 30 October 2008 ) In most applications, the parameters of a mixture of linear regression models are estimated by maximum likelihood using the expectation maximization (EM) algorithm. In this article, we propose the comparison of three algorithms to compute maximum likelihood estimates of the parameters of these models: the EM algorithm, the classification EM algorithm and the stochastic EM algorithm. The comparison of the three procedures was done through a simulation study of the performance (computational effort, statistical properties of estimators and goodness of fit) of these approaches on simulated data sets. Simulation results show that the choice of the approach depends essentially on the configuration of the true regression lines and the initialization of the algorithms. Keywords: mixtures of linear regressions; maximum likelihood estimation; EM algorithm; classification EM algorithm; simulation study 1. Introduction Finite mixture models have provided a mathematically based approach to the statistical modelling of a wide variety of random phenomena. Applications of mixture distributions can be found in various fields of statistical applications such as agriculture, biology, economics, medicine and genetics; see e.g. [1 3] for a review. Within the family of mixture models, mixtures of linear regressions have also been studied extensively, especially when no information about membership of the points assigned to each line was available. Mixtures of linear regression models were introduced by Quandt and Ramsey [4] as a very general form of switching regression. They used a technique based on a moment-generating function to estimate the parameters. However, it has mainly been studied from a likelihood point of view. De Veaux [5] developed an EM approach to fit the two regression situations. Jones and McLachlan [6] applied mixtures of regressions in a data analysis and used the EM algorithm to fit these models. Turner [7] fitted a two-component mixture of one variable linear regression to *Corresponding author. sfaria@mct.uminho.pt ISSN print/issn online 2010 Taylor & Francis DOI: /

2 202 S. Faria and G. Soromenho a data set using the EM algorithm. Hawkins et al. [8] studied the problem of determining the number of components in a mixture of linear regression models using methods derived from the likelihood equation. More recently, Zhu and Zhang [9] established asymptotic theory for maximum likelihood estimators in mixture regression models. In this article, we study the procedure for fitting mixtures of linear regressions by means of maximum likelihood. We apply three maximization algorithms to obtain the maximum likelihood estimates: the expectation maximization (EM) algorithm (see [10]), the classification EM (CEM) algorithm (see [11]) and the stochastic EM (SEM) algorithm (see [12]). The comparison of EM and CEM approaches in a cluster analysis is well known in the mixture models literature. Under the Gaussian mixture, Ganesalingam [13] has performed numerical experiments to compare the two approaches in practical situations. An extension of this study was performed by Celeux and Govaert [14], in order to specify the influence of the sample sizes and the dependence of the used algorithms over their initial values. Considering the case of binary data, Govaert and Nadif [15] presented an extension of the comparisons to Bernoulli models. Some comparisons of EM and SEM approaches in a mixture of distributions are also available. Celeux et al. [16] have investigated the practical behaviour of these algorithms through intensive Monte Carlo numerical simulations and a real data study. Dias and Wedel [17] have compared EM and SEM algorithms to estimate the parameters of Gaussian mixture model. Our goal is to compare the performance of these three approaches on mixtures of linear regressions. A simulation study is designed to investigate this problem. The article is organized as follows: in Section 2, we present the mixture of linear regression model and the three maximization algorithms to obtain the maximum likelihood estimates. Section 3 provides a simulation study investigating the performance of the algorithms for fitting two- and three-component mixtures of linear regression models. In Section 4, the conclusions of our study are drawn and additional comments are given. 2. Fitting mixtures of linear regressions The mixture of linear regression model is given as follows: xi T β 1 + ɛ i1 with probability π 1, xi T y i = β 2 + ɛ i2 with probability π 2,. xi Tβ J + ɛ ij with probability π J (1) where y i is the value of the response variable in the ith observation; xi T (i = 1,...,n) denotes the transpose of the (p + 1)-dimensional vector of independent variables for the ith observation, β ( = 1,...,J) denotes the (p+1)-dimensional vector of regressor variables for the th component, π are the mixing probabilities (0 <π < 1, for all = 1,...,J and π = 1). Finally, ɛ i are the random errors; under the assumption of normality, we have ɛ i N(0,σ 2)(i = 1,...,n; = 1,...,J). Given a set of independent observations y 1,y 2,...,y n, corresponding to values x 1,x 2,...,x n of the predictor x, the complete parameter set of the mixture model, θ = (π 1,...,π J,β 1,...,β J,σ1 2,...,σ2 J ), can be estimated by maximizing the log-likelihood n J L(θ x 1,...,x n,y 1,...,y n ) = log π φ (y i x i ), (2) i=1 =1

3 Journal of Statistical Computation and Simulation 203 where φ (y i x i ) denotes the density of an univariate Gaussian distribution with mean x T i β and variance σ The EM algorithm The standard tool for finding the maximum likelihood solution is the EM algorithm (see [10] and [3]). The EM algorithm is a broadly applicable approach to the iterative computation of maximum likelihood estimates when the observations can be viewed as incomplete data. The idea here is to think of the data as consisting of triples (x i,y i,z i ), where z i is the unobserved indicator that specifies the mixture component from which the observation y i is drawn. The EM algorithm is easy to program and proceeds iteratively in two steps, E (for expectation) and M (for maximization). Let θ (r) be the estimate of the parameters after the rth iteration. On the (r + 1)th iteration, the E-step of the EM algorithm involves the calculation of the Q-function, which is the expectation of the complete-data log-likelihood conditional on the current parameter estimates and the observed data, where Q(θ, θ (r) ) = n w (r) π (r) φ (y i x i ) i = J =1 π (r) φ (y i x i ) J i=1 =1 w (r) i φ (y i x i ), (3) (i = 1,...,n; = 1,...,J) (4) is the estimate of the posterior probability that the ith observation belongs to the th component of the mixture after the rth iteration. The M-step updates the estimate θ (r+1) that maximizes the Q-function with respect to θ. Itis equivalent to computing the sample proportion and the weighted least-squares estimates when performing a weighted regression of y 1,...,y n on x 1,...,x n with weights w 1,...,w n ( = 1,...,J). It follows that on the M-step of the (r + 1)th iteration, the current fit for the parameters is given explicitly by ˆπ (r+1) = n i=1 w(r) i n ( = 1,...,J) (5) ˆβ (r+1) = (X T W X) 1 X T W Y ( = 1,...,J), (6) where X is a n (p + 1) matrix of predictors, W is a n n diagonal matrix with diagonal entries and Y is a n 1 vector of response variable; and w (r) i ˆσ 2(r+1) = n i=1 w(r) i (y i x T i n i=1 w(r) i ˆβ (r+1) ) 2 ( = 1,...,J). (7) The E- and M- steps are alternated repeatedly until some specified convergence criterion is achieved. The characteristics of the EM algorithm are well documented (see for instance [3]).

4 204 S. Faria and G. Soromenho 2.2. The CEM algorithm To fit mixtures of linear regressions, we also make use of a classification version of the EM algorithm, the so-called CEM algorithm. The CEM algorithm maximizes in θ and z 1,...,z n the complete data classification log-likelihood, where the missing component label z i of each sample observation is included in the data set: CL(θ z 1,...,z n,x 1,...,x n,y 1,...,y n ) = J =1 {i z i =} log(π φ (y i x i )), (8) where {i z i = } is the set of observations arising from the th mixture component. The CEM algorithm incorporates a classification step (C-step) between the E- and M-steps of EM. This classification step involves assigning each observation to one of the J components that provides the largest posterior probability w i. Thus, an iteration of CEM algorithm consists of three steps. The E-step of the CEM algorithm is identical to the E-step of the EM algorithm. On the C-step of the (r + 1)th iteration, a partition P (r+1) = (P (r+1) 1,...,P (r+1) J ) of (x 1,y 1 ),...,(x n,y n ) is designed by assigning each observation to the component for which w (r) i is largest (if the maximum posterior probability is not unique, we choose the component with the smallest index). We have, P (r+1) ={(x i,y i ) : w (r) i = arg h max w (r) ih } (9) if w (r) i = w (r) ih and <hthen (x i,y i ) P (r+1) ( = 1,...,J). If one of the P (r+1) is empty or has only one observation, it must be considered that the mixture has J 1 components instead of J and the estimation process begins with J 1 components. The M-step updates the estimate θ (r+1) using the sub-samples P (r+1). It follows that on the M-step of the (r + 1)th iteration, the current fit for the parameters is given explicitly by ˆπ (r+1) = n n where n is the total number of observations arising from component ; ( = 1,...,J), (10) ˆβ (r+1) = (X T W X ) 1 X T W Y ( = 1,...,J), (11) where X is a n (p + 1) matrix of predictors for the th component, W is a n n diagonal matrix with diagonal entries w (r) i and Y is a n 1 vector of response variable for the th component; and ˆσ 2(r+1) = n i=1 w(r) i (y i x T i β(r+1) n i=1 w(r) i ) 2. (12) The E-, C- and M-steps are alternated repeatedly until some specified convergence criterion is achieved. CEM algorithm is a K-means-like algorithm and contrary to EM, it converges in a finite number of iterations The SEM algorithm We also apply a procedure for fitting mixtures of linear regressions using a stochastic version of the EM algorithm, the so-called SEM algorithm. The SEM algorithm incorporates a stochastic

5 Journal of Statistical Computation and Simulation 205 step (S-step) between the E- and M-steps of EM. This stochastic step simulates a realization of the unobserved indicator z i,i = 1,...,nby drawing them at random from their current conditional distribution. Thus, an iteration of SEM algorithm consists of three steps. The E-step of the SEM algorithm is identical to the E-step of the EM algorithm. On the S-step of the (r + 1)th iteration, a partition P (r+1) = (P (r+1) 1,...,P (r+1) J ) of (x 1,y 1 ),...,(x n,y n ) is designed by assigning each observation at random to one of the mixture components according to the multinomial distribution with parameter w (r) i, given by Equation (4). The M-step of the SEM algorithm is identical to the M-step of the CEM algorithm. SEM does not converge pointwise. It generates a Markov chain whose stationary distribution is more or less concentrated around the maximum likelihood parameter estimate. 3. Simulation study of algorithm performance A simulation study was performed to assess the performance of the maximum likelihood estimates obtained via the EM algorithm, the CEM algorithm and the SEM algorithm. Data were simulated under a two to three component mixture of linear regressions. We used the freeware R (see [18]) to develop the simulation program Design of the study Initial conditions. In our simulation study, two different strategies of choosing initial values were considered. In the first strategy, the true values were used as the starting values. In the other strategy we ran the algorithm 20 times from random initial position and selected the solution out of 20 runs which provided the best value of the optimized criterion (see [14]). Stopping rules. A rather strict stopping criterion for the EM and the CEM algorithms was used: iterations were stopped when the relative change in log-likelihood between two successive iterations were less than The stopping rule for the SEM algorithm was the total number of iterations required for convergence by the EM algorithm. We do not use the same stopping criteria because the slow convergence of the SEM algorithm makes such criteria hazardous. Number of samples. For each type of simulated data set, 200 samples of size n were simulated. Configurations of the true regression lines. We considered two typical configurations of the true regression lines: parallel and concurrent. These configurations are expected to affect the performance of the proposed algorithms. Data set. Each datum (x i,y i ) was generated by the following scheme. First, a uniform [0, 1] random number c i was generated and its value was used to select a particular component from mixture of regression models. Next, x i was randomly generated from a uniform [x L,x U ] distribution and a normal random variate ɛ i with mean 0 and variance σ 2 was calculated. Finally, the value y i was assigned using x i, ɛ i and the appropriate model parameters (see [19]). We have chosen x L = 1and x U = 3 and in previous simulation studies we have obtained the same simulation results when these values were changed. Measures of algorithm performance: In order to examine the performance of two algorithms, the following criteria were used: the mean number of iterations required for convergence (which gives an indication about the computing time needed), the statistical properties of the estimators θ:

6 206 S. Faria and G. Soromenho (i) bias of the parameter estimates over the 200 replications: BIAS( ˆθ ) = m=1 ˆθ (m) θ, (13) where θ = (π,β,σ 2) and ˆθ (m) = ( ˆπ (m), ˆβ (m), ˆσ 2(m) ), = 1,...,J of the mth replication with m = 1,...,200. (ii) the mean square error (MSE) of the parameter estimates over the 200 replications: MSE( ˆθ ) = ( ˆθ (m) θ ) 2 (14) m=1 the root mean-squared error of prediction (MRSEP): MRSEP = RMSEP (m), (15) m=1 where RMSEP (m) is the root mean-squared error of prediction of the mth replication based on K-fold cross-validation, which is given by RMSEP (m) = 1 n (y i ŷ (m) i ) n 2 (16) with ŷ (m) i corresponding to the fitted value of the observation i of the mth replication. For the K-fold cross-validation, we have chosen K = 5 and K = 10 (see [20], Chapter 7). i=1 The simulation process consists of the following steps: (1) Create a data set of size n. (2) Fit a mixture of linear regression models to the data using the EM, the CEM and the SEM algorithms. Save the number of iterations required for convergence and the estimated parameters ˆθ = ( ˆπ 1,..., ˆπ J, ˆβ 1,..., ˆβ J, ˆσ 2 1,..., ˆσ 2 J ). (3) Split the data into K roughly equal-sized parts. For the kth part, fit the model to the other K 1 parts of the data using the EM, the CEM and the SEM algorithms, and calculate the prediction error of the fitted model when predicting the kth part of the data. Do this for k = 1,...,K, combine the K estimates of prediction error and compute the corresponding value for RMSEP. (4) Repeat steps 1 3, for a total of 200 trials. Compute the mean number of iterations required for convergence, the bias of the parameter estimates (BIAS( ˆθ )), the mean square error (MSE( ˆθ )) of the parameter estimates and the root mean-squared error of prediction (MRSEP) Simulation results: two component mixtures of linear regressions For two component models (J = 2), samples of three different sizes n(n= 50, 100, 500) were generated for each set of true parameter values (β, σ ) shown on Table 1 and the mixing proportion π 1 lying from 0.1 to 0.9. For instance, we present in Figure 1 typical scatter plots for samples with size 100 and mixing proportion 0.5. Tables 2 and 3 provide the mean number of iterations required for convergence using the EM and the CEM algorithms for fitting two-component mixtures of linear regression models. In all

7 Journal of Statistical Computation and Simulation 207 Table 1. True parameter values for the essays with a two-component mixture of linear regressions. Configuration β 10 β 20 β 11 β 21 σ 2 1 σ 2 2 Parallel Concurrent Figure 1. Scatter plot of samples from two-component models with n = 100 and π = (0.5; 0.5). the cases, the mean number of iterations for convergence is smaller using the CEM algorithm rather than using the EM algorithm. Tables 4 7 provide the MSE and the bias of the parameter estimates over the 200 replications of the two-component mixtures of linear regression models, when the mixing proportion π 1 is chosen to be 0.2, 0.5 and 0.7. When the true regression lines are parallel and the algorithms are initiated with the true parameter values, Table 4 shows that CEM estimates have smaller MSE than EM and SEM estimates. It is evident that the estimates obtained by the three algorithms have relatively small bias and MSE tends to decrease as the sample size increases. Table 2. The mean number of iterations required for convergence using the EM and CEM algorithms for two-component mixtures of linear regressions when the true values were used as the starting values. Parallel Concurrent n = 50 n = 100 n = 500 n = 50 n = 100 n = 500 π 1 EM CEM EM CEM EM CEM EM CEM EM CEM EM CEM

8 208 S. Faria and G. Soromenho Table 3. The mean number of iterations required for convergence using the EM and CEM algorithms for two-component mixtures of linear regressions when the algorithms were initiated by random numbers (second strategy). Parallel Concurrent n = 50 n = 100 n = 500 n = 50 n = 100 n = 500 π 1 EM CEM EM CEM EM CEM EM CEM EM CEM EM CEM When the true regression lines are parallel but the initialization of the algorithms is made by random numbers (second strategy), Table 5 shows that SEM performs better than EM and CEM. In this case, however the performance of all three algorithms decrease, producing estimates that have higher MSE and bias. When the true regression lines are concurrent and the algorithms are initiated with the true parameter values, Table 6 shows that the three algorithms have practically the same behaviour. Also, the MSE of the parameter estimates decreases whenever the sample size increases. When the true regression lines are concurrent but the algorithms are initiated by random numbers (second strategy), Table 7 shows that CEM estimates of the parameters have higher MSE and bias than EM and SEM estimates. In generality, EM outperforms SEM by producing estimates of the parameters that have smaller MSE and bias. The resulting values of MRSEP based on 10-fold cross-validation when the true values were used as the starting values for each of the configurations of the true regression lines are plotted in Figure 2. Similar results were obtained calculating MRSEP based on 5-fold cross-validation. When the true regression lines are parallel, the CEM algorithm performs generally better, however, for a sample size of 500 and when the mixing proportions are equal, it seems that the three algorithms have practically the same behaviour. When the true regression lines are concurrent, it seems that the three algorithms have practically the same behaviour. However, for a sample size of 50 and 100, the performances of the CEM and SEM algorithms decrease when the mixing proportion is smaller and the EM algorithm performs better in those cases. The resulting values of MRSEP based on 10-fold cross-validation when the second strategy was used as the starting values for each of the configurations of the true regression lines are plotted in Figure 3. Similar results were obtained calculating MRSEP based on 5-fold cross-validation. When the true regression lines are parallel, the SEM algorithm performs generally better; however, for a sample size of 500, it seems that the EM algorithm performs better. Figure 3 suggests that the EM algorithm performs always better in fitting a two-component mixture of linear regressions when the true regression lines are concurrent and the second strategy was used as the starting value Simulation results: three-component mixtures of linear regressions For three component models (J = 3), samples of size n = 100 and n = 500 were generated for the two sets of parameter values (β, σ ) shown in Table 8. For illustration we show scatter plots of random samples of 100 points in Figure 4.

9 Journal of Statistical Computation and Simulation 209 Table 4. Mean square error and bias of estimates based on 200 replications of the two-component mixtures of linear regression models when the true regression lines are parallel and the true values are used as the starting values. n π 1 Algorithm β 10 β 11 β 20 β 21 σ1 2 σ2 2 π 1 π EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE

10 210 S. Faria and G. Soromenho Table 5. Mean square error and bias of estimates based on 200 replications of the two-component mixtures of linear regression models when the true regression lines are parallel and the algorithms are initiated by random numbers (second strategy). n π 1 Algorithm β 10 β 11 β 20 β 21 σ 2 1 σ 2 2 π 1 π EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE

11 Journal of Statistical Computation and Simulation 211 Table 6. Mean square error and bias of estimates based on 200 replications of the two-component mixtures of linear regression models when the true regression lines are concurrent and the true values are used as the starting values. n π 1 Algorithm β 10 β 11 β 20 β 21 σ1 2 σ2 2 π 1 π EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE

12 212 S. Faria and G. Soromenho Table 7. Mean square error and bias of estimates based on 200 replications of the two-component mixtures of linear regression models when the true regression lines are concurrent and the algorithms are initiated by random numbers (second strategy). n π 1 Algorithm β 10 β 11 β 20 β 21 σ 2 1 σ 2 2 π 1 π EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE EM BIAS MSE CEM BIAS MSE SEM BIAS MSE

13 Journal of Statistical Computation and Simulation 213 Figure 2. values. MRSEP by 10-fold cross-validation for two-component models when the true values were used as the starting Figure 3. MRSEP by 10-fold cross-validation for two-component models when the algorithms were initiated by random numbers (second strategy). Table 8. True parameter values for the essays with a three-component mixture of linear regressions. Configuration β 01 β 02 β 03 β 11 β 12 β 13 σ 2 1 σ 2 2 σ 2 3 Parallel Concurrent Tables 9 and 10 report the mean number of iterations required for convergence using the EM and CEM algorithms for fitting three-component mixtures of linear regression models. Also in all cases, the mean number of iterations for convergence is smaller using the CEM algorithm rather than using the EM algorithm.

14 214 S. Faria and G. Soromenho Figure 4. Scatter plot of samples from three-component models with n = 100 and π = (0.4; 0.3; 0.3). Table 9. The mean number of iterations required for convergence using the EM and CEM algorithms for three-component mixtures of linear regressions when the true values were used as the starting values. Parallel Concurrent n = 100 n = 500 n = 100 n = 500 (π 1 ; π 2 ) EM CEM EM CEM EM CEM EM CEM (0.2; 0.2) (0.2; 0.3) (0.2; 0.4) (0.2; 0.5) (0.2; 0.6) (0.3; 0.2) (0.3; 0.3) (0.3; 0.4) (0.3; 0.5) (0.4; 0.2) (0.4; 0.3) (0.4; 0.4) (0.5; 0.2) (0.5; 0.3) (0.6; 0.2) Tables provide the MSE and the bias of the parameter estimates over the 200 replications of the three-component mixtures of linear regression models, when the mixing proportion (π 1 ; π 2 ; π 3 ) is chosen to be (0.2;0.2;0.6), (0.2;0.4;0.4), (0.3;0.3;0.4) and (0.5;0.3;0.2). When the true regression lines are parallel and the algorithms are initiated with the true parameter values, Table 11 shows that the CEM parameter estimates have smaller MSE than the EM and SEM estimates and EM and SEM have practically the same behaviour. However, for samples of size 500, the estimates MSE are identical. Also, as in the previous cases, the MSE decreases when the sample size increases.

15 Journal of Statistical Computation and Simulation 215 Table 10. The mean number of iterations required for convergence using the EM and CEM algorithms for three-component mixtures of linear regressions when the algorithms were initiated by random numbers (second strategy). Parallel Concurrent n = 100 n = 500 n = 100 n = 500 (π 1 ; π 2 ) EM CEM EM CEM EM CEM EM CEM (0.2; 0.2) (0.2; 0.3) (0.2; 0.4) (0.2; 0.5) (0.2; 0.6) (0.3; 0.2) (0.3; 0.3) (0.3; 0.4) (0.3; 0.5) (0.4; 0.2) (0.4; 0.3) (0.4; 0.4) (0.5; 0.2) (0.5; 0.3) (0.6; 0.2) When the true regression lines are parallel and the algorithms are initiated by random numbers (second strategy), Table 12 shows that the SEM algorithm performs better than the other two. Using this strategy as the starting value, the performance of all algorithms decreases by producing estimates of the parameters that have higher MSE and bias. When the true regression lines are concurrent and the algorithms are initiated with the true parameter values, Table 13 shows that EM outperforms CEM and SEM by producing estimates of the parameters that have lower bias and smaller MSE. It seems that the MSE of CEM estimates of the regression coefficients and the variances are smaller than the MSE of SEM estimates. The MSE of the parameter estimates tends to approach zero as sample size increases. When the true regression lines are concurrent but the algorithms are initiated by random numbers (second strategy), Table 14 shows that CEM estimates of the parameters have higher MSE than EM and SEM estimates. In generality, SEM outperforms EM by producing estimates of the parameters that have smaller MSE. The resulting values of MRSEP based on 10-fold cross-validation when the true values were used as the starting value for each of the configurations of the true regression lines are plotted in Figure 5. Similar results were obtained calculating MRSEP based on 5-fold cross-validation. When the true regression lines are parallel, the CEM algorithm performs better, although in samples of size 500, the three algorithms have practically the same behaviour. Figure 5 suggests that the CEM algorithm performs always better in fitting a three-component mixture of linear regressions when the true regression lines are concurrent and the algorithms are initiated with the true parameter values. The resulting values of MRSEP based on 10-fold cross-validation when the second strategy was used as the starting values for each of the configurations of the true regression lines are plotted in Figure 6. Similar results were obtained calculating MRSEP based on 5-fold cross-validation. In both configurations of the true regression lines, Figure 6 suggests that the SEM algorithm performs always better in fitting a three-component mixture of linear regressions.

16 Table 11. Mean square error and bias of estimates based on 200 replications of the three-component mixtures of linear regression models when the true regression lines are parallel and the true values are used as the starting values. n (π 1 ; π 2 ) Algorithm β 10 β 11 β 20 β 21 β 30 β 31 σ 1 σ 2 σ 3 π 1 π 2 π (0.2;0.2) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.2;0.4) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.3;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.5;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE S. Faria and G. Soromenho

17 500 (0.2;0.2) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.2;0.4) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.3;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.5;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE Journal of Statistical Computation and Simulation 217

18 Table 12. Mean square error (MSE) and bias of estimates based on 200 replications of the three component mixtures of linear regression models when the true regression lines are parallel and the algorithms were initiated by random numbers (second strategy). n (π 1 ; π 2 ) Algorithm β 10 β 11 β 20 β 21 β 30 β 31 σ 1 σ 2 σ 3 π 1 π 2 π (0.2;0.2) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.2;0.4) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.3;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE (0.5;0.3) EM BIAS MSE CEM BIAS MSE SEM BIAS MSE S. Faria and G. Soromenho

6.3 Forecasting ARMA processes

6.3 Forecasting ARMA processes 122 CHAPTER 6. ARMA MODELS 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss a linear

Διαβάστε περισσότερα

Bayesian statistics. DS GA 1002 Probability and Statistics for Data Science.

Bayesian statistics. DS GA 1002 Probability and Statistics for Data Science. Bayesian statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Frequentist vs Bayesian statistics In frequentist

Διαβάστε περισσότερα

Other Test Constructions: Likelihood Ratio & Bayes Tests

Other Test Constructions: Likelihood Ratio & Bayes Tests Other Test Constructions: Likelihood Ratio & Bayes Tests Side-Note: So far we have seen a few approaches for creating tests such as Neyman-Pearson Lemma ( most powerful tests of H 0 : θ = θ 0 vs H 1 :

Διαβάστε περισσότερα

HOMEWORK 4 = G. In order to plot the stress versus the stretch we define a normalized stretch:

HOMEWORK 4 = G. In order to plot the stress versus the stretch we define a normalized stretch: HOMEWORK 4 Problem a For the fast loading case, we want to derive the relationship between P zz and λ z. We know that the nominal stress is expressed as: P zz = ψ λ z where λ z = λ λ z. Therefore, applying

Διαβάστε περισσότερα

Statistical Inference I Locally most powerful tests

Statistical Inference I Locally most powerful tests Statistical Inference I Locally most powerful tests Shirsendu Mukherjee Department of Statistics, Asutosh College, Kolkata, India. shirsendu st@yahoo.co.in So far we have treated the testing of one-sided

Διαβάστε περισσότερα

Solution Series 9. i=1 x i and i=1 x i.

Solution Series 9. i=1 x i and i=1 x i. Lecturer: Prof. Dr. Mete SONER Coordinator: Yilin WANG Solution Series 9 Q1. Let α, β >, the p.d.f. of a beta distribution with parameters α and β is { Γ(α+β) Γ(α)Γ(β) f(x α, β) xα 1 (1 x) β 1 for < x

Διαβάστε περισσότερα

: Monte Carlo EM 313, Louis (1982) EM, EM Newton-Raphson, /. EM, 2 Monte Carlo EM Newton-Raphson, Monte Carlo EM, Monte Carlo EM, /. 3, Monte Carlo EM

: Monte Carlo EM 313, Louis (1982) EM, EM Newton-Raphson, /. EM, 2 Monte Carlo EM Newton-Raphson, Monte Carlo EM, Monte Carlo EM, /. 3, Monte Carlo EM 2008 6 Chinese Journal of Applied Probability and Statistics Vol.24 No.3 Jun. 2008 Monte Carlo EM 1,2 ( 1,, 200241; 2,, 310018) EM, E,,. Monte Carlo EM, EM E Monte Carlo,. EM, Monte Carlo EM,,,,. Newton-Raphson.

Διαβάστε περισσότερα

EE512: Error Control Coding

EE512: Error Control Coding EE512: Error Control Coding Solution for Assignment on Finite Fields February 16, 2007 1. (a) Addition and Multiplication tables for GF (5) and GF (7) are shown in Tables 1 and 2. + 0 1 2 3 4 0 0 1 2 3

Διαβάστε περισσότερα

5.4 The Poisson Distribution.

5.4 The Poisson Distribution. The worst thing you can do about a situation is nothing. Sr. O Shea Jackson 5.4 The Poisson Distribution. Description of the Poisson Distribution Discrete probability distribution. The random variable

Διαβάστε περισσότερα

Homework 3 Solutions

Homework 3 Solutions Homework 3 Solutions Igor Yanovsky (Math 151A TA) Problem 1: Compute the absolute error and relative error in approximations of p by p. (Use calculator!) a) p π, p 22/7; b) p π, p 3.141. Solution: For

Διαβάστε περισσότερα

Problem Set 3: Solutions

Problem Set 3: Solutions CMPSCI 69GG Applied Information Theory Fall 006 Problem Set 3: Solutions. [Cover and Thomas 7.] a Define the following notation, C I p xx; Y max X; Y C I p xx; Ỹ max I X; Ỹ We would like to show that C

Διαβάστε περισσότερα

ST5224: Advanced Statistical Theory II

ST5224: Advanced Statistical Theory II ST5224: Advanced Statistical Theory II 2014/2015: Semester II Tutorial 7 1. Let X be a sample from a population P and consider testing hypotheses H 0 : P = P 0 versus H 1 : P = P 1, where P j is a known

Διαβάστε περισσότερα

CHAPTER 25 SOLVING EQUATIONS BY ITERATIVE METHODS

CHAPTER 25 SOLVING EQUATIONS BY ITERATIVE METHODS CHAPTER 5 SOLVING EQUATIONS BY ITERATIVE METHODS EXERCISE 104 Page 8 1. Find the positive root of the equation x + 3x 5 = 0, correct to 3 significant figures, using the method of bisection. Let f(x) =

Διαβάστε περισσότερα

Estimation for ARMA Processes with Stable Noise. Matt Calder & Richard A. Davis Colorado State University

Estimation for ARMA Processes with Stable Noise. Matt Calder & Richard A. Davis Colorado State University Estimation for ARMA Processes with Stable Noise Matt Calder & Richard A. Davis Colorado State University rdavis@stat.colostate.edu 1 ARMA processes with stable noise Review of M-estimation Examples of

Διαβάστε περισσότερα

Statistics 104: Quantitative Methods for Economics Formula and Theorem Review

Statistics 104: Quantitative Methods for Economics Formula and Theorem Review Harvard College Statistics 104: Quantitative Methods for Economics Formula and Theorem Review Tommy MacWilliam, 13 tmacwilliam@college.harvard.edu March 10, 2011 Contents 1 Introduction to Data 5 1.1 Sample

Διαβάστε περισσότερα

Concrete Mathematics Exercises from 30 September 2016

Concrete Mathematics Exercises from 30 September 2016 Concrete Mathematics Exercises from 30 September 2016 Silvio Capobianco Exercise 1.7 Let H(n) = J(n + 1) J(n). Equation (1.8) tells us that H(2n) = 2, and H(2n+1) = J(2n+2) J(2n+1) = (2J(n+1) 1) (2J(n)+1)

Διαβάστε περισσότερα

4.6 Autoregressive Moving Average Model ARMA(1,1)

4.6 Autoregressive Moving Average Model ARMA(1,1) 84 CHAPTER 4. STATIONARY TS MODELS 4.6 Autoregressive Moving Average Model ARMA(,) This section is an introduction to a wide class of models ARMA(p,q) which we will consider in more detail later in this

Διαβάστε περισσότερα

HW 3 Solutions 1. a) I use the auto.arima R function to search over models using AIC and decide on an ARMA(3,1)

HW 3 Solutions 1. a) I use the auto.arima R function to search over models using AIC and decide on an ARMA(3,1) HW 3 Solutions a) I use the autoarima R function to search over models using AIC and decide on an ARMA3,) b) I compare the ARMA3,) to ARMA,0) ARMA3,) does better in all three criteria c) The plot of the

Διαβάστε περισσότερα

Μηχανική Μάθηση Hypothesis Testing

Μηχανική Μάθηση Hypothesis Testing ΕΛΛΗΝΙΚΗ ΔΗΜΟΚΡΑΤΙΑ ΠΑΝΕΠΙΣΤΗΜΙΟ ΚΡΗΤΗΣ Μηχανική Μάθηση Hypothesis Testing Γιώργος Μπορμπουδάκης Τμήμα Επιστήμης Υπολογιστών Procedure 1. Form the null (H 0 ) and alternative (H 1 ) hypothesis 2. Consider

Διαβάστε περισσότερα

Approximation of distance between locations on earth given by latitude and longitude

Approximation of distance between locations on earth given by latitude and longitude Approximation of distance between locations on earth given by latitude and longitude Jan Behrens 2012-12-31 In this paper we shall provide a method to approximate distances between two points on earth

Διαβάστε περισσότερα

The Simply Typed Lambda Calculus

The Simply Typed Lambda Calculus Type Inference Instead of writing type annotations, can we use an algorithm to infer what the type annotations should be? That depends on the type system. For simple type systems the answer is yes, and

Διαβάστε περισσότερα

Congruence Classes of Invertible Matrices of Order 3 over F 2

Congruence Classes of Invertible Matrices of Order 3 over F 2 International Journal of Algebra, Vol. 8, 24, no. 5, 239-246 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.2988/ija.24.422 Congruence Classes of Invertible Matrices of Order 3 over F 2 Ligong An and

Διαβάστε περισσότερα

ES440/ES911: CFD. Chapter 5. Solution of Linear Equation Systems

ES440/ES911: CFD. Chapter 5. Solution of Linear Equation Systems ES440/ES911: CFD Chapter 5. Solution of Linear Equation Systems Dr Yongmann M. Chung http://www.eng.warwick.ac.uk/staff/ymc/es440.html Y.M.Chung@warwick.ac.uk School of Engineering & Centre for Scientific

Διαβάστε περισσότερα

Figure A.2: MPC and MPCP Age Profiles (estimating ρ, ρ = 2, φ = 0.03)..

Figure A.2: MPC and MPCP Age Profiles (estimating ρ, ρ = 2, φ = 0.03).. Supplemental Material (not for publication) Persistent vs. Permanent Income Shocks in the Buffer-Stock Model Jeppe Druedahl Thomas H. Jørgensen May, A Additional Figures and Tables Figure A.: Wealth and

Διαβάστε περισσότερα

Απόκριση σε Μοναδιαία Ωστική Δύναμη (Unit Impulse) Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο. Απόστολος Σ.

Απόκριση σε Μοναδιαία Ωστική Δύναμη (Unit Impulse) Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο. Απόστολος Σ. Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο The time integral of a force is referred to as impulse, is determined by and is obtained from: Newton s 2 nd Law of motion states that the action

Διαβάστε περισσότερα

Numerical Analysis FMN011

Numerical Analysis FMN011 Numerical Analysis FMN011 Carmen Arévalo Lund University carmen@maths.lth.se Lecture 12 Periodic data A function g has period P if g(x + P ) = g(x) Model: Trigonometric polynomial of order M T M (x) =

Διαβάστε περισσότερα

6.1. Dirac Equation. Hamiltonian. Dirac Eq.

6.1. Dirac Equation. Hamiltonian. Dirac Eq. 6.1. Dirac Equation Ref: M.Kaku, Quantum Field Theory, Oxford Univ Press (1993) η μν = η μν = diag(1, -1, -1, -1) p 0 = p 0 p = p i = -p i p μ p μ = p 0 p 0 + p i p i = E c 2 - p 2 = (m c) 2 H = c p 2

Διαβάστε περισσότερα

Section 8.3 Trigonometric Equations

Section 8.3 Trigonometric Equations 99 Section 8. Trigonometric Equations Objective 1: Solve Equations Involving One Trigonometric Function. In this section and the next, we will exple how to solving equations involving trigonometric functions.

Διαβάστε περισσότερα

FORMULAS FOR STATISTICS 1

FORMULAS FOR STATISTICS 1 FORMULAS FOR STATISTICS 1 X = 1 n Sample statistics X i or x = 1 n x i (sample mean) S 2 = 1 n 1 s 2 = 1 n 1 (X i X) 2 = 1 n 1 (x i x) 2 = 1 n 1 Xi 2 n n 1 X 2 x 2 i n n 1 x 2 or (sample variance) E(X)

Διαβάστε περισσότερα

Probability and Random Processes (Part II)

Probability and Random Processes (Part II) Probability and Random Processes (Part II) 1. If the variance σ x of d(n) = x(n) x(n 1) is one-tenth the variance σ x of a stationary zero-mean discrete-time signal x(n), then the normalized autocorrelation

Διαβάστε περισσότερα

Math 6 SL Probability Distributions Practice Test Mark Scheme

Math 6 SL Probability Distributions Practice Test Mark Scheme Math 6 SL Probability Distributions Practice Test Mark Scheme. (a) Note: Award A for vertical line to right of mean, A for shading to right of their vertical line. AA N (b) evidence of recognizing symmetry

Διαβάστε περισσότερα

Introduction to the ML Estimation of ARMA processes

Introduction to the ML Estimation of ARMA processes Introduction to the ML Estimation of ARMA processes Eduardo Rossi University of Pavia October 2013 Rossi ARMA Estimation Financial Econometrics - 2013 1 / 1 We consider the AR(p) model: Y t = c + φ 1 Y

Διαβάστε περισσότερα

Queensland University of Technology Transport Data Analysis and Modeling Methodologies

Queensland University of Technology Transport Data Analysis and Modeling Methodologies Queensland University of Technology Transport Data Analysis and Modeling Methodologies Lab Session #7 Example 5.2 (with 3SLS Extensions) Seemingly Unrelated Regression Estimation and 3SLS A survey of 206

Διαβάστε περισσότερα

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 19/5/2007

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 19/5/2007 Οδηγίες: Να απαντηθούν όλες οι ερωτήσεις. Αν κάπου κάνετε κάποιες υποθέσεις να αναφερθούν στη σχετική ερώτηση. Όλα τα αρχεία που αναφέρονται στα προβλήματα βρίσκονται στον ίδιο φάκελο με το εκτελέσιμο

Διαβάστε περισσότερα

ΕΙΣΑΓΩΓΗ ΣΤΗ ΣΤΑΤΙΣΤΙΚΗ ΑΝΑΛΥΣΗ

ΕΙΣΑΓΩΓΗ ΣΤΗ ΣΤΑΤΙΣΤΙΚΗ ΑΝΑΛΥΣΗ ΕΙΣΑΓΩΓΗ ΣΤΗ ΣΤΑΤΙΣΤΙΚΗ ΑΝΑΛΥΣΗ ΕΛΕΝΑ ΦΛΟΚΑ Επίκουρος Καθηγήτρια Τµήµα Φυσικής, Τοµέας Φυσικής Περιβάλλοντος- Μετεωρολογίας ΓΕΝΙΚΟΙ ΟΡΙΣΜΟΙ Πληθυσµός Σύνολο ατόµων ή αντικειµένων στα οποία αναφέρονται

Διαβάστε περισσότερα

2 Composition. Invertible Mappings

2 Composition. Invertible Mappings Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan Composition. Invertible Mappings In this section we discuss two procedures for creating new mappings from old ones, namely,

Διαβάστε περισσότερα

Instruction Execution Times

Instruction Execution Times 1 C Execution Times InThisAppendix... Introduction DL330 Execution Times DL330P Execution Times DL340 Execution Times C-2 Execution Times Introduction Data Registers This appendix contains several tables

Διαβάστε περισσότερα

Phys460.nb Solution for the t-dependent Schrodinger s equation How did we find the solution? (not required)

Phys460.nb Solution for the t-dependent Schrodinger s equation How did we find the solution? (not required) Phys460.nb 81 ψ n (t) is still the (same) eigenstate of H But for tdependent H. The answer is NO. 5.5.5. Solution for the tdependent Schrodinger s equation If we assume that at time t 0, the electron starts

Διαβάστε περισσότερα

Lecture 34 Bootstrap confidence intervals

Lecture 34 Bootstrap confidence intervals Lecture 34 Bootstrap confidence intervals Confidence Intervals θ: an unknown parameter of interest We want to find limits θ and θ such that Gt = P nˆθ θ t If G 1 1 α is known, then P θ θ = P θ θ = 1 α

Διαβάστε περισσότερα

Jesse Maassen and Mark Lundstrom Purdue University November 25, 2013

Jesse Maassen and Mark Lundstrom Purdue University November 25, 2013 Notes on Average Scattering imes and Hall Factors Jesse Maassen and Mar Lundstrom Purdue University November 5, 13 I. Introduction 1 II. Solution of the BE 1 III. Exercises: Woring out average scattering

Διαβάστε περισσότερα

Aquinas College. Edexcel Mathematical formulae and statistics tables DO NOT WRITE ON THIS BOOKLET

Aquinas College. Edexcel Mathematical formulae and statistics tables DO NOT WRITE ON THIS BOOKLET Aquinas College Edexcel Mathematical formulae and statistics tables DO NOT WRITE ON THIS BOOKLET Pearson Edexcel Level 3 Advanced Subsidiary and Advanced GCE in Mathematics and Further Mathematics Mathematical

Διαβάστε περισσότερα

Asymptotic distribution of MLE

Asymptotic distribution of MLE Asymptotic distribution of MLE Theorem Let {X t } be a causal and invertible ARMA(p,q) process satisfying Φ(B)X = Θ(B)Z, {Z t } IID(0, σ 2 ). Let ( ˆφ, ˆϑ) the values that minimize LL n (φ, ϑ) among those

Διαβάστε περισσότερα

Partial Differential Equations in Biology The boundary element method. March 26, 2013

Partial Differential Equations in Biology The boundary element method. March 26, 2013 The boundary element method March 26, 203 Introduction and notation The problem: u = f in D R d u = ϕ in Γ D u n = g on Γ N, where D = Γ D Γ N, Γ D Γ N = (possibly, Γ D = [Neumann problem] or Γ N = [Dirichlet

Διαβάστε περισσότερα

C.S. 430 Assignment 6, Sample Solutions

C.S. 430 Assignment 6, Sample Solutions C.S. 430 Assignment 6, Sample Solutions Paul Liu November 15, 2007 Note that these are sample solutions only; in many cases there were many acceptable answers. 1 Reynolds Problem 10.1 1.1 Normal-order

Διαβάστε περισσότερα

Second Order RLC Filters

Second Order RLC Filters ECEN 60 Circuits/Electronics Spring 007-0-07 P. Mathys Second Order RLC Filters RLC Lowpass Filter A passive RLC lowpass filter (LPF) circuit is shown in the following schematic. R L C v O (t) Using phasor

Διαβάστε περισσότερα

SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM

SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question 1 a) The cumulative distribution function of T conditional on N n is Pr T t N n) Pr max X 1,..., X N ) t N n) Pr max

Διαβάστε περισσότερα

Lecture 12: Pseudo likelihood approach

Lecture 12: Pseudo likelihood approach Lecture 12: Pseudo likelihood approach Pseudo MLE Let X 1,...,X n be a random sample from a pdf in a family indexed by two parameters θ and π with likelihood l(θ,π). The method of pseudo MLE may be viewed

Διαβάστε περισσότερα

Tridiagonal matrices. Gérard MEURANT. October, 2008

Tridiagonal matrices. Gérard MEURANT. October, 2008 Tridiagonal matrices Gérard MEURANT October, 2008 1 Similarity 2 Cholesy factorizations 3 Eigenvalues 4 Inverse Similarity Let α 1 ω 1 β 1 α 2 ω 2 T =......... β 2 α 1 ω 1 β 1 α and β i ω i, i = 1,...,

Διαβάστε περισσότερα

Econ 2110: Fall 2008 Suggested Solutions to Problem Set 8 questions or comments to Dan Fetter 1

Econ 2110: Fall 2008 Suggested Solutions to Problem Set 8  questions or comments to Dan Fetter 1 Eon : Fall 8 Suggested Solutions to Problem Set 8 Email questions or omments to Dan Fetter Problem. Let X be a salar with density f(x, θ) (θx + θ) [ x ] with θ. (a) Find the most powerful level α test

Διαβάστε περισσότερα

On a four-dimensional hyperbolic manifold with finite volume

On a four-dimensional hyperbolic manifold with finite volume BULETINUL ACADEMIEI DE ŞTIINŢE A REPUBLICII MOLDOVA. MATEMATICA Numbers 2(72) 3(73), 2013, Pages 80 89 ISSN 1024 7696 On a four-dimensional hyperbolic manifold with finite volume I.S.Gutsul Abstract. In

Διαβάστε περισσότερα

derivation of the Laplacian from rectangular to spherical coordinates

derivation of the Laplacian from rectangular to spherical coordinates derivation of the Laplacian from rectangular to spherical coordinates swapnizzle 03-03- :5:43 We begin by recognizing the familiar conversion from rectangular to spherical coordinates (note that φ is used

Διαβάστε περισσότερα

[1] P Q. Fig. 3.1

[1] P Q. Fig. 3.1 1 (a) Define resistance....... [1] (b) The smallest conductor within a computer processing chip can be represented as a rectangular block that is one atom high, four atoms wide and twenty atoms long. One

Διαβάστε περισσότερα

SCHOOL OF MATHEMATICAL SCIENCES G11LMA Linear Mathematics Examination Solutions

SCHOOL OF MATHEMATICAL SCIENCES G11LMA Linear Mathematics Examination Solutions SCHOOL OF MATHEMATICAL SCIENCES GLMA Linear Mathematics 00- Examination Solutions. (a) i. ( + 5i)( i) = (6 + 5) + (5 )i = + i. Real part is, imaginary part is. (b) ii. + 5i i ( + 5i)( + i) = ( i)( + i)

Διαβάστε περισσότερα

Exercises 10. Find a fundamental matrix of the given system of equations. Also find the fundamental matrix Φ(t) satisfying Φ(0) = I. 1.

Exercises 10. Find a fundamental matrix of the given system of equations. Also find the fundamental matrix Φ(t) satisfying Φ(0) = I. 1. Exercises 0 More exercises are available in Elementary Differential Equations. If you have a problem to solve any of them, feel free to come to office hour. Problem Find a fundamental matrix of the given

Διαβάστε περισσότερα

Section 7.6 Double and Half Angle Formulas

Section 7.6 Double and Half Angle Formulas 09 Section 7. Double and Half Angle Fmulas To derive the double-angles fmulas, we will use the sum of two angles fmulas that we developed in the last section. We will let α θ and β θ: cos(θ) cos(θ + θ)

Διαβάστε περισσότερα

Solutions to Exercise Sheet 5

Solutions to Exercise Sheet 5 Solutions to Eercise Sheet 5 jacques@ucsd.edu. Let X and Y be random variables with joint pdf f(, y) = 3y( + y) where and y. Determine each of the following probabilities. Solutions. a. P (X ). b. P (X

Διαβάστε περισσότερα

SCITECH Volume 13, Issue 2 RESEARCH ORGANISATION Published online: March 29, 2018

SCITECH Volume 13, Issue 2 RESEARCH ORGANISATION Published online: March 29, 2018 Journal of rogressive Research in Mathematics(JRM) ISSN: 2395-028 SCITECH Volume 3, Issue 2 RESEARCH ORGANISATION ublished online: March 29, 208 Journal of rogressive Research in Mathematics www.scitecresearch.com/journals

Διαβάστε περισσότερα

Block Ciphers Modes. Ramki Thurimella

Block Ciphers Modes. Ramki Thurimella Block Ciphers Modes Ramki Thurimella Only Encryption I.e. messages could be modified Should not assume that nonsensical messages do no harm Always must be combined with authentication 2 Padding Must be

Διαβάστε περισσότερα

DESIGN OF MACHINERY SOLUTION MANUAL h in h 4 0.

DESIGN OF MACHINERY SOLUTION MANUAL h in h 4 0. DESIGN OF MACHINERY SOLUTION MANUAL -7-1! PROBLEM -7 Statement: Design a double-dwell cam to move a follower from to 25 6, dwell for 12, fall 25 and dwell for the remader The total cycle must take 4 sec

Διαβάστε περισσότερα

Areas and Lengths in Polar Coordinates

Areas and Lengths in Polar Coordinates Kiryl Tsishchanka Areas and Lengths in Polar Coordinates In this section we develop the formula for the area of a region whose boundary is given by a polar equation. We need to use the formula for the

Διαβάστε περισσότερα

Homework for 1/27 Due 2/5

Homework for 1/27 Due 2/5 Name: ID: Homework for /7 Due /5. [ 8-3] I Example D of Sectio 8.4, the pdf of the populatio distributio is + αx x f(x α) =, α, otherwise ad the method of momets estimate was foud to be ˆα = 3X (where

Διαβάστε περισσότερα

Srednicki Chapter 55

Srednicki Chapter 55 Srednicki Chapter 55 QFT Problems & Solutions A. George August 3, 03 Srednicki 55.. Use equations 55.3-55.0 and A i, A j ] = Π i, Π j ] = 0 (at equal times) to verify equations 55.-55.3. This is our third

Διαβάστε περισσότερα

CE 530 Molecular Simulation

CE 530 Molecular Simulation C 53 olecular Siulation Lecture Histogra Reweighting ethods David. Kofke Departent of Cheical ngineering SUNY uffalo kofke@eng.buffalo.edu Histogra Reweighting ethod to cobine results taken at different

Διαβάστε περισσότερα

Lecture 2. Soundness and completeness of propositional logic

Lecture 2. Soundness and completeness of propositional logic Lecture 2 Soundness and completeness of propositional logic February 9, 2004 1 Overview Review of natural deduction. Soundness and completeness. Semantics of propositional formulas. Soundness proof. Completeness

Διαβάστε περισσότερα

3.4 SUM AND DIFFERENCE FORMULAS. NOTE: cos(α+β) cos α + cos β cos(α-β) cos α -cos β

3.4 SUM AND DIFFERENCE FORMULAS. NOTE: cos(α+β) cos α + cos β cos(α-β) cos α -cos β 3.4 SUM AND DIFFERENCE FORMULAS Page Theorem cos(αβ cos α cos β -sin α cos(α-β cos α cos β sin α NOTE: cos(αβ cos α cos β cos(α-β cos α -cos β Proof of cos(α-β cos α cos β sin α Let s use a unit circle

Διαβάστε περισσότερα

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 24/3/2007

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 24/3/2007 Οδηγίες: Να απαντηθούν όλες οι ερωτήσεις. Όλοι οι αριθμοί που αναφέρονται σε όλα τα ερωτήματα μικρότεροι του 10000 εκτός αν ορίζεται διαφορετικά στη διατύπωση του προβλήματος. Αν κάπου κάνετε κάποιες υποθέσεις

Διαβάστε περισσότερα

An Introduction to Signal Detection and Estimation - Second Edition Chapter II: Selected Solutions

An Introduction to Signal Detection and Estimation - Second Edition Chapter II: Selected Solutions An Introduction to Signal Detection Estimation - Second Edition Chapter II: Selected Solutions H V Poor Princeton University March 16, 5 Exercise : The likelihood ratio is given by L(y) (y +1), y 1 a With

Διαβάστε περισσότερα

Example Sheet 3 Solutions

Example Sheet 3 Solutions Example Sheet 3 Solutions. i Regular Sturm-Liouville. ii Singular Sturm-Liouville mixed boundary conditions. iii Not Sturm-Liouville ODE is not in Sturm-Liouville form. iv Regular Sturm-Liouville note

Διαβάστε περισσότερα

These derivations are not part of the official forthcoming version of Vasilaky and Leonard

These derivations are not part of the official forthcoming version of Vasilaky and Leonard Target Input Model with Learning, Derivations Kathryn N Vasilaky These derivations are not part of the official forthcoming version of Vasilaky and Leonard 06 in Economic Development and Cultural Change.

Διαβάστε περισσότερα

A Bonus-Malus System as a Markov Set-Chain. Małgorzata Niemiec Warsaw School of Economics Institute of Econometrics

A Bonus-Malus System as a Markov Set-Chain. Małgorzata Niemiec Warsaw School of Economics Institute of Econometrics A Bonus-Malus System as a Markov Set-Chain Małgorzata Niemiec Warsaw School of Economics Institute of Econometrics Contents 1. Markov set-chain 2. Model of bonus-malus system 3. Example 4. Conclusions

Διαβάστε περισσότερα

b. Use the parametrization from (a) to compute the area of S a as S a ds. Be sure to substitute for ds!

b. Use the parametrization from (a) to compute the area of S a as S a ds. Be sure to substitute for ds! MTH U341 urface Integrals, tokes theorem, the divergence theorem To be turned in Wed., Dec. 1. 1. Let be the sphere of radius a, x 2 + y 2 + z 2 a 2. a. Use spherical coordinates (with ρ a) to parametrize.

Διαβάστε περισσότερα

Exercises to Statistics of Material Fatigue No. 5

Exercises to Statistics of Material Fatigue No. 5 Prof. Dr. Christine Müller Dipl.-Math. Christoph Kustosz Eercises to Statistics of Material Fatigue No. 5 E. 9 (5 a Show, that a Fisher information matri for a two dimensional parameter θ (θ,θ 2 R 2, can

Διαβάστε περισσότερα

Description of the PX-HC algorithm

Description of the PX-HC algorithm A Description of the PX-HC algorithm Let N = C c= N c and write C Nc K c= i= k= as, the Gibbs sampling algorithm at iteration m for continuous outcomes: Step A: For =,, J, draw θ m in the following steps:

Διαβάστε περισσότερα

Second Order Partial Differential Equations

Second Order Partial Differential Equations Chapter 7 Second Order Partial Differential Equations 7.1 Introduction A second order linear PDE in two independent variables (x, y Ω can be written as A(x, y u x + B(x, y u xy + C(x, y u u u + D(x, y

Διαβάστε περισσότερα

Appendix to On the stability of a compressible axisymmetric rotating flow in a pipe. By Z. Rusak & J. H. Lee

Appendix to On the stability of a compressible axisymmetric rotating flow in a pipe. By Z. Rusak & J. H. Lee Appendi to On the stability of a compressible aisymmetric rotating flow in a pipe By Z. Rusak & J. H. Lee Journal of Fluid Mechanics, vol. 5 4, pp. 5 4 This material has not been copy-edited or typeset

Διαβάστε περισσότερα

Εργαστήριο Ανάπτυξης Εφαρμογών Βάσεων Δεδομένων. Εξάμηνο 7 ο

Εργαστήριο Ανάπτυξης Εφαρμογών Βάσεων Δεδομένων. Εξάμηνο 7 ο Εργαστήριο Ανάπτυξης Εφαρμογών Βάσεων Δεδομένων Εξάμηνο 7 ο Procedures and Functions Stored procedures and functions are named blocks of code that enable you to group and organize a series of SQL and PL/SQL

Διαβάστε περισσότερα

Math221: HW# 1 solutions

Math221: HW# 1 solutions Math: HW# solutions Andy Royston October, 5 7.5.7, 3 rd Ed. We have a n = b n = a = fxdx = xdx =, x cos nxdx = x sin nx n sin nxdx n = cos nx n = n n, x sin nxdx = x cos nx n + cos nxdx n cos n = + sin

Διαβάστε περισσότερα

Areas and Lengths in Polar Coordinates

Areas and Lengths in Polar Coordinates Kiryl Tsishchanka Areas and Lengths in Polar Coordinates In this section we develop the formula for the area of a region whose boundary is given by a polar equation. We need to use the formula for the

Διαβάστε περισσότερα

Matrices and Determinants

Matrices and Determinants Matrices and Determinants SUBJECTIVE PROBLEMS: Q 1. For what value of k do the following system of equations possess a non-trivial (i.e., not all zero) solution over the set of rationals Q? x + ky + 3z

Διαβάστε περισσότερα

SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM

SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM SOLUTIONS TO MATH38181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question 1 a) The cumulative distribution function of T conditional on N n is Pr (T t N n) Pr (max (X 1,..., X N ) t N n) Pr (max

Διαβάστε περισσότερα

«ΑΝΑΠΣΤΞΖ ΓΠ ΚΑΗ ΥΩΡΗΚΖ ΑΝΑΛΤΖ ΜΔΣΔΩΡΟΛΟΓΗΚΩΝ ΓΔΓΟΜΔΝΩΝ ΣΟΝ ΔΛΛΑΓΗΚΟ ΥΩΡΟ»

«ΑΝΑΠΣΤΞΖ ΓΠ ΚΑΗ ΥΩΡΗΚΖ ΑΝΑΛΤΖ ΜΔΣΔΩΡΟΛΟΓΗΚΩΝ ΓΔΓΟΜΔΝΩΝ ΣΟΝ ΔΛΛΑΓΗΚΟ ΥΩΡΟ» ΓΔΩΠΟΝΗΚΟ ΠΑΝΔΠΗΣΖΜΗΟ ΑΘΖΝΩΝ ΣΜΗΜΑ ΑΞΙΟΠΟΙΗΗ ΦΤΙΚΩΝ ΠΟΡΩΝ & ΓΕΩΡΓΙΚΗ ΜΗΥΑΝΙΚΗ ΣΟΜΕΑ ΕΔΑΦΟΛΟΓΙΑ ΚΑΙ ΓΕΩΡΓΙΚΗ ΥΗΜΕΙΑ ΕΙΔΙΚΕΤΗ: ΕΦΑΡΜΟΓΕ ΣΗ ΓΕΩΠΛΗΡΟΦΟΡΙΚΗ ΣΟΤ ΦΤΙΚΟΤ ΠΟΡΟΤ «ΑΝΑΠΣΤΞΖ ΓΠ ΚΑΗ ΥΩΡΗΚΖ ΑΝΑΛΤΖ ΜΔΣΔΩΡΟΛΟΓΗΚΩΝ

Διαβάστε περισσότερα

Section 9.2 Polar Equations and Graphs

Section 9.2 Polar Equations and Graphs 180 Section 9. Polar Equations and Graphs In this section, we will be graphing polar equations on a polar grid. In the first few examples, we will write the polar equation in rectangular form to help identify

Διαβάστε περισσότερα

Homework 8 Model Solution Section

Homework 8 Model Solution Section MATH 004 Homework Solution Homework 8 Model Solution Section 14.5 14.6. 14.5. Use the Chain Rule to find dz where z cosx + 4y), x 5t 4, y 1 t. dz dx + dy y sinx + 4y)0t + 4) sinx + 4y) 1t ) 0t + 4t ) sinx

Διαβάστε περισσότερα

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 6/5/2006

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 6/5/2006 Οδηγίες: Να απαντηθούν όλες οι ερωτήσεις. Ολοι οι αριθμοί που αναφέρονται σε όλα τα ερωτήματα είναι μικρότεροι το 1000 εκτός αν ορίζεται διαφορετικά στη διατύπωση του προβλήματος. Διάρκεια: 3,5 ώρες Καλή

Διαβάστε περισσότερα

New bounds for spherical two-distance sets and equiangular lines

New bounds for spherical two-distance sets and equiangular lines New bounds for spherical two-distance sets and equiangular lines Michigan State University Oct 8-31, 016 Anhui University Definition If X = {x 1, x,, x N } S n 1 (unit sphere in R n ) and x i, x j = a

Διαβάστε περισσότερα

Overview. Transition Semantics. Configurations and the transition relation. Executions and computation

Overview. Transition Semantics. Configurations and the transition relation. Executions and computation Overview Transition Semantics Configurations and the transition relation Executions and computation Inference rules for small-step structural operational semantics for the simple imperative language Transition

Διαβάστε περισσότερα

Web-based supplementary materials for Bayesian Quantile Regression for Ordinal Longitudinal Data

Web-based supplementary materials for Bayesian Quantile Regression for Ordinal Longitudinal Data Web-based supplementary materials for Bayesian Quantile Regression for Ordinal Longitudinal Data Rahim Alhamzawi, Haithem Taha Mohammad Ali Department of Statistics, College of Administration and Economics,

Διαβάστε περισσότερα

ΑΠΟΔΟΤΙΚΗ ΑΠΟΤΙΜΗΣΗ ΕΡΩΤΗΣΕΩΝ OLAP Η ΜΕΤΑΠΤΥΧΙΑΚΗ ΕΡΓΑΣΙΑ ΕΞΕΙΔΙΚΕΥΣΗΣ. Υποβάλλεται στην

ΑΠΟΔΟΤΙΚΗ ΑΠΟΤΙΜΗΣΗ ΕΡΩΤΗΣΕΩΝ OLAP Η ΜΕΤΑΠΤΥΧΙΑΚΗ ΕΡΓΑΣΙΑ ΕΞΕΙΔΙΚΕΥΣΗΣ. Υποβάλλεται στην ΑΠΟΔΟΤΙΚΗ ΑΠΟΤΙΜΗΣΗ ΕΡΩΤΗΣΕΩΝ OLAP Η ΜΕΤΑΠΤΥΧΙΑΚΗ ΕΡΓΑΣΙΑ ΕΞΕΙΔΙΚΕΥΣΗΣ Υποβάλλεται στην ορισθείσα από την Γενική Συνέλευση Ειδικής Σύνθεσης του Τμήματος Πληροφορικής Εξεταστική Επιτροπή από την Χαρά Παπαγεωργίου

Διαβάστε περισσότερα

Chapter 6: Systems of Linear Differential. be continuous functions on the interval

Chapter 6: Systems of Linear Differential. be continuous functions on the interval Chapter 6: Systems of Linear Differential Equations Let a (t), a 2 (t),..., a nn (t), b (t), b 2 (t),..., b n (t) be continuous functions on the interval I. The system of n first-order differential equations

Διαβάστε περισσότερα

Optimal Parameter in Hermitian and Skew-Hermitian Splitting Method for Certain Two-by-Two Block Matrices

Optimal Parameter in Hermitian and Skew-Hermitian Splitting Method for Certain Two-by-Two Block Matrices Optimal Parameter in Hermitian and Skew-Hermitian Splitting Method for Certain Two-by-Two Block Matrices Chi-Kwong Li Department of Mathematics The College of William and Mary Williamsburg, Virginia 23187-8795

Διαβάστε περισσότερα

Biostatistics for Health Sciences Review Sheet

Biostatistics for Health Sciences Review Sheet Biostatistics for Health Sciences Review Sheet http://mathvault.ca June 1, 2017 Contents 1 Descriptive Statistics 2 1.1 Variables.............................................. 2 1.1.1 Qualitative........................................

Διαβάστε περισσότερα

Διπλωματική Εργασία. Μελέτη των μηχανικών ιδιοτήτων των stents που χρησιμοποιούνται στην Ιατρική. Αντωνίου Φάνης

Διπλωματική Εργασία. Μελέτη των μηχανικών ιδιοτήτων των stents που χρησιμοποιούνται στην Ιατρική. Αντωνίου Φάνης Διπλωματική Εργασία Μελέτη των μηχανικών ιδιοτήτων των stents που χρησιμοποιούνται στην Ιατρική Αντωνίου Φάνης Επιβλέπουσες: Θεοδώρα Παπαδοπούλου, Ομότιμη Καθηγήτρια ΕΜΠ Ζάννη-Βλαστού Ρόζα, Καθηγήτρια

Διαβάστε περισσότερα

Strain gauge and rosettes

Strain gauge and rosettes Strain gauge and rosettes Introduction A strain gauge is a device which is used to measure strain (deformation) on an object subjected to forces. Strain can be measured using various types of devices classified

Διαβάστε περισσότερα

TMA4115 Matematikk 3

TMA4115 Matematikk 3 TMA4115 Matematikk 3 Andrew Stacey Norges Teknisk-Naturvitenskapelige Universitet Trondheim Spring 2010 Lecture 12: Mathematics Marvellous Matrices Andrew Stacey Norges Teknisk-Naturvitenskapelige Universitet

Διαβάστε περισσότερα

Every set of first-order formulas is equivalent to an independent set

Every set of first-order formulas is equivalent to an independent set Every set of first-order formulas is equivalent to an independent set May 6, 2008 Abstract A set of first-order formulas, whatever the cardinality of the set of symbols, is equivalent to an independent

Διαβάστε περισσότερα

Forced Pendulum Numerical approach

Forced Pendulum Numerical approach Numerical approach UiO April 8, 2014 Physical problem and equation We have a pendulum of length l, with mass m. The pendulum is subject to gravitation as well as both a forcing and linear resistance force.

Διαβάστε περισσότερα

Physical DB Design. B-Trees Index files can become quite large for large main files Indices on index files are possible.

Physical DB Design. B-Trees Index files can become quite large for large main files Indices on index files are possible. B-Trees Index files can become quite large for large main files Indices on index files are possible 3 rd -level index 2 nd -level index 1 st -level index Main file 1 The 1 st -level index consists of pairs

Διαβάστε περισσότερα

Πρόβλημα 1: Αναζήτηση Ελάχιστης/Μέγιστης Τιμής

Πρόβλημα 1: Αναζήτηση Ελάχιστης/Μέγιστης Τιμής Πρόβλημα 1: Αναζήτηση Ελάχιστης/Μέγιστης Τιμής Να γραφεί πρόγραμμα το οποίο δέχεται ως είσοδο μια ακολουθία S από n (n 40) ακέραιους αριθμούς και επιστρέφει ως έξοδο δύο ακολουθίες από θετικούς ακέραιους

Διαβάστε περισσότερα

Fourier Series. MATH 211, Calculus II. J. Robert Buchanan. Spring Department of Mathematics

Fourier Series. MATH 211, Calculus II. J. Robert Buchanan. Spring Department of Mathematics Fourier Series MATH 211, Calculus II J. Robert Buchanan Department of Mathematics Spring 2018 Introduction Not all functions can be represented by Taylor series. f (k) (c) A Taylor series f (x) = (x c)

Διαβάστε περισσότερα

Example of the Baum-Welch Algorithm

Example of the Baum-Welch Algorithm Example of the Baum-Welch Algorithm Larry Moss Q520, Spring 2008 1 Our corpus c We start with a very simple corpus. We take the set Y of unanalyzed words to be {ABBA, BAB}, and c to be given by c(abba)

Διαβάστε περισσότερα