@@ -111,19 +111,32 @@ To appreciate how statisticians connect probabilities to data, the key is to und
111
111
112
112
**Scalar example**
113
113
114
+ Let $X$ be a scalar random variable that takes on the $I$ possible values
115
+ $0, 1, 2, \ldots, I-1$ with probabilities
114
116
115
- Consider the following discrete distribution
117
+ $$
118
+ {\rm Prob}(X = i) = f_i, \quad
119
+ $$
120
+ where
121
+
122
+ $$
123
+ f_i \geqslant 0, \quad \sum_i f_i = 1 .
124
+ $$
125
+
126
+ We sometimes write
116
127
117
128
$$
118
- X \sim \{ {f_i}\} _ {i=0}^{I-1},\quad f_i \geqslant 0, \quad \sum_i f_i = 1
129
+ X \sim \{ {f_i}\} _ {i=0}^{I-1}
119
130
$$
120
131
121
- Draw a sample $x_0, x_1, \dots , x_{N-1}$, $N$ draws of $X$ from $\{f_i\}^I_{i=1}$.
132
+ as a short-hand way of saying that the random variable $X$ is described by the probability distribution $ \{{f_i}\}_{i=0}^{I-1}$.
133
+
134
+ Consider drawing a sample $x_0, x_1, \dots , x_{N-1}$ of $N$ independent and identically distributoed draws of $X$.
122
135
123
136
What do the "identical" and "independent" mean in IID or iid ("identically and independently distributed)?
124
137
125
138
- "identical" means that each draw is from the same distribution.
126
- - "independent" means that the joint distribution equal tthe product of marginal distributions, i.e.,
139
+ - "independent" means that joint distribution equal products of marginal distributions, i.e.,
127
140
128
141
$$
129
142
\begin{aligned}
132
145
\end{aligned}
133
146
$$
134
147
135
- Consider the **empirical distribution**:
148
+ We define an e **empirical distribution** as follows.
149
+
150
+ For each $i = 0,\dots,I-1$, let
136
151
137
152
$$
138
153
\begin{aligned}
139
- i & = 0,\dots,I-1,\\
140
154
N_i & = \text{number of times} \ X = i,\\
141
155
N & = \sum^{I-1}_ {i=0} N_i \quad \text{total number of draws},\\
142
156
\tilde {f_i} & = \frac{N_i}{N} \sim \ \text{frequency of draws for which}\ X=i
@@ -425,7 +439,7 @@ Conditional distributions are
425
439
426
440
$$
427
441
\begin{aligned}
428
- \textrm{Prob}\{ X=i|Y=j\} & =\frac{f_ig_j}{\sum_ {i}f_ig_j}=\frac{f_ig_j}{g_i }=f_i \\
442
+ \textrm{Prob}\{ X=i|Y=j\} & =\frac{f_ig_j}{\sum_ {i}f_ig_j}=\frac{f_ig_j}{g_j }=f_i \\
429
443
\textrm{Prob}\{ Y=j|X=i\} & =\frac{f_ig_j}{\sum_ {j}f_ig_j}=\frac{f_ig_j}{f_i}=g_j
430
444
\end{aligned}
431
445
$$
609
623
\begin{aligned}
610
624
\tilde{U} & =F(X)=1-\lambda^{x+1}\\
611
625
1-\tilde{U} & =\lambda^{x+1}\\
612
- log(1-\tilde{U})& =(x+1)\log\lambda\\
626
+ \ log(1-\tilde{U})& =(x+1)\log\lambda\\
613
627
\frac{\log(1-\tilde{U})}{\log\lambda}& =x+1\\
614
628
\frac{\log(1-\tilde{U})}{\log\lambda}-1 &=x
615
629
\end{aligned}
@@ -1561,7 +1575,7 @@ Now we'll try to go in a reverse direction.
1561
1575
1562
1576
We'll find that from two marginal distributions, can we usually construct more than one joint distribution that verifies these marginals.
1563
1577
1564
- Each of these joint distributions is called a ** coupling** of the two martingal distributions.
1578
+ Each of these joint distributions is called a ** coupling** of the two marginal distributions.
1565
1579
1566
1580
Let's start with marginal distributions
1567
1581
0 commit comments