01/30/07 Conditional probability, conditional expectation Reference: Grigoriu Sec. 2.11, Bertsekas Sec. 3.5 Office hours: Wednesdays and Fridays 4-5 PM For random vectors X and Y which have a joint probability density, one can use a simpler formula for the conditional probability density: Conditional expectation: Simple way to handle this: This is an example of a general principle that you can apply the usual (unconditional) rules for probability and expectation also in the conditional case provided that all probability distributions and expectations in the formula are conditioned on the same event. Note also the above formula works when B is an event involving another random vector Y, for example, There is a more abstract formulation of conditional expectation which you can find in mathematical sdenotes013007b Page 1 There is a more abstract formulation of conditional expectation which you can find in mathematical texts, and we'll come back to it later. Some useful formulas for calculations involving multiple random variables: computing Marginal probability density function from joint probability density function: More generally: Sums of random variables: sdenotes013007b Page 2 There is no assumption here that Y and X are independent. The point is that when we condition on the random variable X, it behaves deterministically with respect to that conditional expectation. Important class of jointly distributed random variables is the jointly Gaussian (jointly normal) distribution. Joint Gaussian (normal) probability density: One important property of jointly Gaussian distributed random variables is that if their sdenotes013007b Page 3 covariance is 0, then they are independent (and vice versa). Lots of known formulas for Gaussian distributions: conditional probability, conditional expectation, … Another important property of jointly Gaussian random variables is that: zero covariance implies independence (and vice versa) linear combinations of Gaussian random vectors are Gaussian random vectors conditional distributions and marginal distributions resulting from jointly Gaussian distributions are also Gaussian Simple calculation example: We will derive the formula for the variance of a sum of (general) random variables from the formula for expectation of linear combination of r.v.s sdenotes013007b Page 4 Simulating Random Variables numerically Online readings will be posted Inverse Transform Method (always works but can be rather impractical to program and/or expensive to compute) Given a one-dimensional random variable X on a state space S which is a subset of the real line. sdenotes013007b Page 5 What does it mean for the computer to simulate a "pseudorandom" number? Pseudorandom numbers are generated by deterministic algorithms that are designed to behave as much as possible like random numbers do. If you want to be really careful about pseudorandom number generators, then L'Ecuyer has up-to-date high quality algorithms. sdenotes013007b Page 6 sdenotes013007b Page 7 What about Gaussian random variables? Inverse transform method is not very efficient because the CDF for a Gaussian is erf, need to invert that. So instead, one uses typically one of two clever algorithms that are specially designed for Gaussian random variables: Box Muller Method Polar-Marsagla Method sdenotes013007b Page 8 In Cartesian coordinates: When I change variables (such as going to polar coordinates) or more generally, map one set of random variables to another, then how do we relate the joint probability density of the original set of random variables to the pdf for the new set of variables? sdenotes013007b Page 9 Change of because the mapping is one-to-one (except for singularity at origin, but this has zero measures so no problem.) More generally, if the mapping has full rank but may have multiple inverse images, the above argument must be modified by starting with: and then repeating the argument, we arrive at the more general formula: See Grigoriu Sec. 2.11 sdenotes013007b Page 10
© Copyright 2025 Paperzz