Multivariate probability distributions

Similar documents
Today s outline: pp

14.30 Introduction to Statistical Methods in Economics Spring 2009

Recursive Estimation

Probability Model for 2 RV s

Exam 2 is Tue Nov 21. Bring a pencil and a calculator. Discuss similarity to exam1. HW3 is due Tue Dec 5.

Joint probability distributions

Will Monroe July 21, with materials by Mehran Sahami and Chris Piech. Joint Distributions

Missing Data Analysis for the Employee Dataset

Lecture 8: Jointly distributed random variables

Topic 5 - Joint distributions and the CLT

This is a good time to refresh your memory on double-integration. We will be using this skill in the upcoming lectures.

Regression III: Advanced Methods

LINEAR PROGRAMMING: A GEOMETRIC APPROACH. Copyright Cengage Learning. All rights reserved.

STATISTICAL LABORATORY, April 30th, 2010 BIVARIATE PROBABILITY DISTRIBUTIONS

Section 4.3. Graphing Exponential Functions

Chapter 1. Linear Equations and Straight Lines. 2 of 71. Copyright 2014, 2010, 2007 Pearson Education, Inc.

Page 129 Exercise 5: Suppose that the joint p.d.f. of two random variables X and Y is as follows: { c(x. 0 otherwise. ( 1 = c. = c

SPSS Basics for Probability Distributions

B. Examples Set up the integral(s) needed to find the area of the region bounded by

= f (a, b) + (hf x + kf y ) (a,b) +

This is a good time to refresh your memory on double-integration. We will be using this skill in the upcoming lectures.

BUSINESS DECISION MAKING. Topic 1 Introduction to Statistical Thinking and Business Decision Making Process; Data Collection and Presentation

Chapter 6 Normal Probability Distributions

Chapter 2 Modeling Distributions of Data

Quick Start Guide Jacob Stolk PhD Simone Stolk MPH November 2018

Probability Models.S4 Simulating Random Variables

GRAPHING WORKSHOP. A graph of an equation is an illustration of a set of points whose coordinates satisfy the equation.

STAT STATISTICAL METHODS. Statistics: The science of using data to make decisions and draw conclusions

Chapter Two: Descriptive Methods 1/50

Courtesy :

Graphs of Exponential

Probability and Statistics for Final Year Engineering Students

To complete the computer assignments, you ll use the EViews software installed on the lab PCs in WMC 2502 and WMC 2506.

Polynomial and Rational Functions. Copyright Cengage Learning. All rights reserved.

JMP Book Descriptions

Direction Fields; Euler s Method

NOTES TO CONSIDER BEFORE ATTEMPTING EX 1A TYPES OF DATA

Numerical Integration

Intro. To Graphing Linear Equations

Topic 6: Calculus Integration Volume of Revolution Paper 2

Missing Data Analysis for the Employee Dataset

Linear and quadratic Taylor polynomials for functions of several variables.

CS 112: Computer System Modeling Fundamentals. Prof. Jenn Wortman Vaughan April 21, 2011 Lecture 8

Lecture 09: Continuous RV. Lisa Yan July 16, 2018

8. MINITAB COMMANDS WEEK-BY-WEEK

September 18, B Math Test Chapter 1 Name: x can be expressed as: {y y 0, y R}.

Chapter 2: The Normal Distribution

Subject. Creating a diagram. Dataset. Importing the data file. Descriptive statistics with TANAGRA.

September 08, Graph y 2 =x. How? Is it a function? Function?

Relations and Functions 2.1

Statool User s Manual

Computer Graphics: Graphics Output Primitives Line Drawing Algorithms

Distributions of random variables

Lesson 18-1 Lesson Lesson 18-1 Lesson Lesson 18-2 Lesson 18-2

Slide 1 / 220. Linear Relations and Functions

Factoring Practice. 1. x x 2 10x x x x x 3-8. (x 4)(x + 4) (x 6)(x 4) (x + 3)(x 2-3x + 9)

Round each observation to the nearest tenth of a cent and draw a stem and leaf plot.

State the domain and range of the relation. EX: {(-1,1), (1,5), (0,3)} 1 P a g e Province Mathematics Southwest TN Community College

Biostatistics & SAS programming. Kevin Zhang

Derivatives and Graphs of Functions

MCMC Methods for data modeling

Probability and Statistics. Copyright Cengage Learning. All rights reserved.

5/3/2010Z:\ jeh\self\notes.doc\7 Chapter 7 Graphical models and belief propagation Graphical models and belief propagation

MULTI-DIMENSIONAL MONTE CARLO INTEGRATION

Chapter 2: Looking at Multivariate Data

10.2 Basic Concepts of Limits

Chapter 5. Normal. Normal Curve. the Normal. Curve Examples. Standard Units Standard Units Examples. for Data

MATH 021 UNIT 2 HOMEWORK ASSIGNMENTS

Elementary Statistics

Sec 4.1 Coordinates and Scatter Plots. Coordinate Plane: Formed by two real number lines that intersect at a right angle.

Intro to Probability Instructor: Alexandre Bouchard

Module 12 Rational Functions and Rational Equations

LAB 1 INSTRUCTIONS DESCRIBING AND DISPLAYING DATA

WELCOME! Lecture 3 Thommy Perlinger

MAC What is a Rational Function? Module 12. Rational Functions and Rational Equations. Learning Objective

Pairs of a random variable

Quantitative Biology II!

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Chapter 2 - Graphical Summaries of Data

06: Logistic Regression

Distributions of Continuous Data

COPULA MODELS FOR BIG DATA USING DATA SHUFFLING

CS113: Lecture 5. Topics: Pointers. Pointers and Activation Records

Sampling and Monte-Carlo Integration

Bivariate Linear Regression James M. Murray, Ph.D. University of Wisconsin - La Crosse Updated: October 04, 2017

Lecture 4. Digital Image Enhancement. 1. Principle of image enhancement 2. Spatial domain transformation. Histogram processing

troduction to Algebra Section 8.1

x 2 + 3, r 4(x) = x2 1

Sharp EL-9900 Graphing Calculator

Objective 1: To simulate the rolling of a die 100 times and to build a probability distribution.

Estimating survival from Gray s flexible model. Outline. I. Introduction. I. Introduction. I. Introduction

Multivariable Calculus

CHAPTER 2 Modeling Distributions of Data

Getting Started With Excel

Chapter 2: Statistical Models for Distributions

Stat 342 Exam 3 Fall 2014

What We ll Do... Random

Conic Sections. College Algebra

Continued =5.28

CSCI567 Machine Learning (Fall 2014)

Transcription:

Multivariate probability distributions September, 07 STAT 0 Class Slide

Outline of Topics Background Discrete bivariate distribution 3 Continuous bivariate distribution STAT 0 Class Slide

Multivariate analysis When one measurement is made on each observation in a dataset, univariate analysis is used, e.g., survival time of patients If more than one measurement is made on each observation, a multivariate analysis is used, e.g., survival time, age, cancer subtype, size of cancer, etc. We focus on bivariate analysis, where exactly two measurements are made on each observation The two measurements will be called X and Y. Since X and Y are obtained for each observation, the data for one observation is the pair (X, Y ) STAT 0 Class Slide 3

Bivariate data Bivariate data can be represented as: Observation X Y X Y X Y 3 X 3 Y 3 4 X 4 Y 4... n X n Y n Each observation is a pair of values, e.g., (X 4, Y 4 ) is the 4-th observation X and Y can be both discrete, both continuous, or one discrete and one continuous. We focus on the first two cases Some examples: X (survived > year) and Y (cancer subtype) of each patient in a sample X (length of job training) and Y (time to find a job) for each unemployed individual in a job training program X (income) and Y (happiness) for each individual in a survey STAT 0 Class Slide 4

Bivariate distributions We can study X and Y separately, i.e., we can analyse X, X,..., X n and Y, Y,..., Y n separately using probability distribution function, probability density function or cumulative distribution function. These are examples of univariate analyses. When X and Y are studied separately, their distribution and probability are called marginal When X and Y are considered together, many interesting questions can be answered, e.g., Is subtype I cancer (X ) associated with a higher chance of survival beyond year (Y )? Does longer job training (X ) result in shorter time to find a job (Y )? Do people with higher income (X ) lead a happier life (Y )? The joint behavior of X and Y is summarized in a bivariate probability distribution. A bivariate distribution is an example of a joint distribution STAT 0 Class Slide 5

Review of a discrete distribution: Drawing a marble from an urn 4 3 5 Probability 3 5 5 Probability distribution tells us the long run frequency for is higher than STAT 0 Class Slide

Discrete bivariate distribution - Drawing marbles with replacement Draw (X ) Draw (Y ) 4 3 5 9 5 5 5 4 5 P(X = and Y = ) = P(, ) = ( ( 3 3 ) 5) 5 = 9 5 P(X = and Y = ) = P(, ) = ( ( 3 ) 5) 5 = 5 P(, ) + P(, ) + P(, ) + P(, ) = 9 5 + 5 + 5 + 4 5 = STAT 0 Class Slide 7

Discrete bivariate distribution- Drawing marbles without replacement Draw (X ) Draw (Y ) 4 3 5 0 0 0 0 P(X = and Y = ) = P(, ) = ( ( 3 ) 4) 5 = 0 P(X = and Y = ) = P(, ) = ( ( 3 ) 4) 5 = 0 P(, ) + P(, ) + P(, ) + P(, ) = 0 + 0 + 0 + 0 = STAT 0 Class Slide 8

Discrete bivariate distributions A discrete bivariate distribution is used to model the joint behavior of two variables, X and Y, both of which are discrete. X and Y are discrete random variables if there is a countable number of possible values for X : a, a,..., a k and for Y : b, b,..., b l. (X, Y ) is the unknown outcome if we randomly draw an observation from the population. P(X = a i, Y = b j ) is the joint probability distribution function of observing X = a i, Y = b j. A valid joint probability distribution function must satisfy the following rules: P(X = a i, Y = b j ) must be between 0 and We are certain that one of the values will appear, therefore: P[(X = a, Y = b ) or (X = a, Y = b ) or... or (X = a k, Y = b l )] = P(X = a, Y = b ) + P(X = a, Y = b ) +... + P(X = a k, Y = b l ) = STAT 0 Class Slide 9

Discrete joint distribution: Example P(X = a, Y = b) = a+b, if a, b = 0,,, 3 Y X 0 3 P(X = a) 0 0 3 3 P(Y = b) 3 4 0 3 4 5 4 3 4 5 0 4 P(X, Y ) 3 + 4 + 5 + Y 0 X 3 0 3 P(X = a, Y = b) are the joint probabilities P(X = a), P(Y = b) are called marginal probabilities. P(X = a) gives us information about X ignoring Y and P(Y = b) gives us information about Y ignoring X We can always find marginal probabilities from joint probabilities (as in Example ) but not the other way around unless X and Y are independent (see next slide) STAT 0 Class Slide 0

Discrete joint distribution- Independence X and Y are independent if, for all X = a, Y = b: P(X = a Y = b) = P(X = a) P(Y = b X = a) = P(Y = b) P(X = a, Y = b) = P(X = a)p(y = b) P(Y = b X = a) and P(X = a Y = b) are conditional probabilities If X and Y are independent, then we can easily (a) calculate P(X = a, Y = b) by P(X = a)p(y = b) (b) write P(X = a Y = b) as P(X = a) (c) write P(Y = b X = a) as P(Y = b) STAT 0 Class Slide

Discrete joint distribution- Example (cont d) P(X, Y ) Y Try, 0 X 3 0 3 P(Y = X = 3) = Alternatively, try P(Y =, X = 3) P(X = 3) P(X = 3, Y = ) = 4 = 4/ / = 4 0 P(Y = ) =. P(X = 3)P(Y = ) = 0 Either way is sufficient to show X and Y are not independent. Furthermore, we can try any combination of X = a, Y = b to disprove independence. STAT 0 Class Slide

Discrete joint distribution: Example P(X = a, Y = b) = ab, if a =,, 3; b =, Y X P(X = a) P(X, Y ) 3 3 P(Y = b) 4 3 9 3 + X Y 3 P(X =, Y = ) = P(X = 3, Y = ) = = P(X = )P(Y = ) = 3 = =. = P(X = 3)P(Y = ) = 9 = 08 = To show independence, we must show P(X = a, Y = b) = P(X = a)p(y = b) for all combinations of X = a, Y = b STAT 0 Class Slide 3

Probability under a univariate probability density function (PDF) PDF P(X ) can be found by a method called integration (not examinable in STAT0): PDF P(X ) f (x) =.5e.5x X P(X ) = 0 f (x)dx = e.5 0.77 It turns out, for any x > 0, P(X x) = e.5x f (x) dx we often write P(X x) as F (x) and call F (x) the cumulative distribution function (CDF) STAT 0 Class Slide 4 X F () is a probability but f () (a point on f (x)) is not a probability

The univariate cumulative distribution function (CDF) PDF P(X x) F (x) F (x) can be used to find P(X x) for any x, e.g., F () = P(X ) A plot of F (x) is a convenient way for finding probabilities. CDF F (x) x X Probability is found by drawing a line ( ) from the horizontal axis until it meets the CDF and then drawing a horizontal line until it meets the vertical axis STAT 0 Class Slide 5 x X All CDF plots have an asymptote at ( ): F (x) because it is a probability

Continuous joint distribution A continuous bivariate distribution is used to model the joint behavior of two variables, X and Y, both of which are continuous. X and Y are continuous random variables if the possible values of X fall in a range (a, b) (, ) and those for Y are in a range (c, d) (, ). The bivariate distribution of (X, Y ) is defined by a joint PDF, f (x, y), with the characteristics: f (x, y) 0 if (x, y) (a, b) (c, d), f (x, y) = 0 otherwise f (x, y) P(X = x, Y = y) = 0 for all values of x, y P(a X b, c Y d) = since X must be in (a, b) and Y must be in (c, d) F (x, y) = P(X x, Y y) is the joint CDF STAT 0 Class Slide

The joint CDF Probabilities are often found by manipulating the joint CDF F (x, y). Example Suppose the joint behaviour of X, Y are given by PDF f (x, y) = e y x, x, y > 0, which is the blue function. Using integration (not examinable in STAT0), f (x, y) the red shaded figure is F (x, y) = P(X x, Y y) = [ e x ][ e y ], x, y > 0 y Y Suppose we wish to find P(X, Y ): P(X, Y ) F (, ) = [ e ][ e ] 0.547 x X STAT 0 Class Slide 7

The joint CDF () f (x, y) Example The joint PDF is f (x, y) =, 0 < x < y, 0 < y <, which is the blue function. The red shaded figure is y Y F (x, y) = P(X x, Y y) = yx x, 0 < x < y, 0 < y < Suppose we wish to find P ( X, Y 3 4) : ( P X, Y 3 ) 4 = ( F, 3 ) 4 ( ) ( 3 4 ) ( ) y Y x X = STAT 0 Class Slide x X

Marginal PDF and CDF As in the discrete case (cf. slide 0), the marginal PDFs f (x) and f (y) for X and Y can be obtained from the joint PDF f (x, y) (Derivation not examinable in STAT0) To find marginal probabilities of X and Y, we need the marginal CDFs F (x) and F (y) Two ways to obtain marginal CDFs F (x) can be obtained from f (x); similarly for F (y) (cf. slide 5) Using joint CDF F (x, y): Reason: STAT 0 Class Slide 9 F (x) = F (x, ), F (y) = F (, y) F (x) P(X x) = P(X x, don t care about Y ) = P(X x, Y ) Same reasoning for F (y) = F (, y) = P(X x, Y ) F (x, )

Marginal CDF: Example f (x, y) =, F (x, y) = yx x, 0 < x < y, 0 < y <, F (x) = F (x, ) = F (x, ), since y < = ()x x = x x, 0 < x < F (y) = F (, y) = F (y, y), since x < y = y(y) (y) = y, 0 < y < STAT 0 Class Slide 0

Conditional probabilities P(X x, Y y) F (x, y) P(X x Y y) = = P(Y y) F (y) P(X x, Y y) F (x, y) P(Y y X x) = = P(X x) F (x) Example F (x, y) = yx f (x, x y), F (y) = y, 0 < x < y, 0 < y < f (x, y) P(Y 0.8) Y P(X x Y y) = yx x (0.8)(0.4) (0.4) P(X 0.4 Y 0.8) = 0.8 f (x, y) = 0.75 y Y X Y STAT 0 Class Slide X 0.8 P(X 0.4 Y 0.8)

Independence A and B are independent if and only if P(A and B) = P(A)P(B) Analogously, if X and Y are independent, we can simplify P(X x, Y y) as P(X x)p(y y). Two ways of proving independence Method P(X x, Y y) = P(X x)p(y y) F (x, y) = F (x)f (y) f (x, y) = f (x)f (y) Method Method requires us to know the marginal PDFs and CDFs. A simpler way to establish independence is to show that f (x, y) can be factorized as f (x, y) = g(x)h(y) STAT 0 Class Slide for any positive functions g, h

Independence Example f (x, y) = e y x,, F (x, y) = [ e x ][ e y ], x, y > 0 Method f (x, y) = e x y = e x }{{} f (x) }{{} e y f (y) ; F (x, y) = [ e x ] [ e y ] }{{}}{{} F (x) F (y) where f (x), x > 0, f (y), y > 0 and F (x), x > 0, F (y), y > 0 are the marginal PDFs and CDFs Method f (x, y) can also be factorized as f (x, y) = g(x)h(y), where { g(x) = e x, 0 < x < h(y) = e y, 0 < y < Using either method, we show X and Y are independent. STAT 0 Class Slide 3

Independence (cont d) Example f (x, y) =, F (x, y) = yx x, 0 < x < y, 0 < y < Method Let I be the indicator function such that I (A) = if A is true, and 0 otherwise, then f (x, y) = I (0 < x < y < ) f (x)f (y) = [( x)][y]. Similarly, F (x, y) = [yx x ]I (0 < x < y < ) F (x)f (y) = [x x ][y ]. Method I (0 < x < y < ) cannot be factorized into two functions g(x) and h(y) Using either method, X and Y are not independent STAT 0 Class Slide 4