(5 intermediate revisions by the same user not shown)
Line 22: Line 22:
 
----
 
----
 
==Question==
 
==Question==
 +
 +
'''1. (25 Points)'''
  
 
X and Y are iid random variable with
 
X and Y are iid random variable with
Line 56: Line 58:
  
 
:'''Click [[ECE_PhD_QE_CNSIP_2007_Problem1.2|here]] to view student [[ECE_PhD_QE_CNSIP_2007_Problem1.2|answers and discussions]]''
 
:'''Click [[ECE_PhD_QE_CNSIP_2007_Problem1.2|here]] to view student [[ECE_PhD_QE_CNSIP_2007_Problem1.2|answers and discussions]]''
 +
----
 +
'''3. (25 Points)'''
 +
 +
Let <math class="inline">\mathbf{X}\left(t\right)</math>  be a real Gaussian random process with mean function <math class="inline">\mu\left(t\right)</math>  and autocovariance function <math class="inline">C_{\mathbf{XX}}\left(t_{1},t_{2}\right)</math> .
 +
 +
'''(a)'''
 +
 +
Write the expression for the <math class="inline">n</math> -th order characteristic function of <math class="inline">\mathbf{X}\left(t\right)</math>  in terms of <math class="inline">\mu\left(t\right)</math>  and <math class="inline">C_{\mathbf{XX}}\left(t_{1},t_{2}\right)</math> .
 +
 +
ref.
 +
 +
There are the note about the [[ECE 600 General Concepts of Stochastic Processes Definitions|n-th order characteristic function of Gaussians random process]] . The only difference between the note and this problem is that this problem use the <math class="inline">\mu\left(t\right)</math>  rather than <math class="inline">\eta_{\mathbf{X}}\left(t\right)=E\left[\mathbf{X}\left(t\right)\right]</math> .
 +
 +
'''Solution'''
 +
 +
<math class="inline">\Phi_{\mathbf{X}\left(t_{1}\right)\cdots\mathbf{X}\left(t_{n}\right)}\left(\omega_{1},\cdots,\omega_{n}\right)=\exp\left\{ i\sum_{k=1}^{n}\mu_{\mathbf{X}}\left(t_{k}\right)\omega_{k}-\frac{1}{2}\sum_{j=1}^{n}\sum_{k=1}^{n}C_{\mathbf{XX}}\left(t_{j},t_{k}\right)\omega_{j}\omega_{k}\right\}</math> .
 +
 +
'''(b)'''
 +
 +
Show that the probabilistic description of <math class="inline">\mathbf{X}\left(t\right)</math>  is completely characterized by <math class="inline">\mu\left(t\right)</math>  and autocovariance function <math class="inline">C_{\mathbf{XX}}\left(t_{1},t_{2}\right)</math> .
 +
 +
 +
:'''Click [[ECE_PhD_QE_CNSIP_2007_Problem1.3|here]] to view student [[ECE_PhD_QE_CNSIP_2007_Problem1.3|answers and discussions]]''
 +
----
 +
 +
'''4. (25 Points)'''
 +
 +
Let <math class="inline">\mathbf{X}_{1},\mathbf{X}_{2},\mathbf{X}_{3},\cdots</math>  be a sequence of independent, identically distributed random variables, each having Cauchy pdf <math class="inline">f\left(x\right)=\frac{1}{\pi\left(1+x^{2}\right)}\;,\qquad-\infty<x<\infty. Let \mathbf{Y}_{n}=\frac{1}{n}\sum_{i=1}^{n}\mathbf{X}_{i}.</math> Find the pdf of <math class="inline">\mathbf{Y}_{n}</math> . Describe how the pdf of <math class="inline">\mathbf{Y}_{n}</math>  depends on <math class="inline">n</math> . Does the sequence <math class="inline">\mathbf{Y}_{1},\mathbf{Y}_{2},\mathbf{Y}_{3},\cdots</math>  converge in distribution? If yes, what is the distribution of the random variable it converges to?
 +
 +
 +
:'''Click [[ECE_PhD_QE_CNSIP_2007_Problem1.4|here]] to view student [[ECE_PhD_QE_CNSIP_2007_Problem1.4|answers and discussions]]''
 +
----
 
----
 
----
 
[[ECE_PhD_Qualifying_Exams|Back to ECE Qualifying Exams (QE) page]]
 
[[ECE_PhD_Qualifying_Exams|Back to ECE Qualifying Exams (QE) page]]

Latest revision as of 09:57, 10 March 2015


ECE Ph.D. Qualifying Exam

Communication, Networking, Signal and Image Processing (CS)

Question 1: Probability and Random Processes

August 2007



Question

1. (25 Points)

X and Y are iid random variable with

$ P(X=i) = P(Y=i) = \frac {1}{2^i}\ ,i = 1,2,3,... $

a) Find $ P(min(X,Y)=k)\ $.

b) Find $ P(X=Y)\ $.

c) Find $ P(Y>X)\ $.

d) Find $ P(Y=kX)\ $.


Click here to view student answers and discussions

2. (25 Points)

Let $ \left\{ \mathbf{X}_{n}\right\} _{n\geq1} $ be a sequence of binomially distributed random variables, with the $ n $ -th random variable $ \mathbf{X}_{n} $ having pmf $ p_{\mathbf{X}_{n}}\left(k\right)=P\left(\left\{ \mathbf{X}_{n}=k\right\} \right)=\left(\begin{array}{c} n\\ k \end{array}\right)p_{n}^{k}\left(1-p_{n}\right)^{n-k}\;,\qquad k=0,\cdots,n,\quad p_{n}\in\left(0,1\right). $

Show that, if the $ p_{n} $ have the property that $ np_{n}\rightarrow\lambda $ as $ n\rightarrow\infty $ , where $ \lambda $ is a positive constant, then the sequence $ \left\{ \mathbf{X}_{n}\right\} _{n\geq1} $ converges in distribution to a Poisson random variable $ \mathbf{X} $ with mean $ \lambda $ .

Hint:

You may find the following fact useful:

$ \lim_{n\rightarrow\infty}\left(1+\frac{x}{n}\right)^{n}=e^{x}. $


'Click here to view student answers and discussions

3. (25 Points)

Let $ \mathbf{X}\left(t\right) $ be a real Gaussian random process with mean function $ \mu\left(t\right) $ and autocovariance function $ C_{\mathbf{XX}}\left(t_{1},t_{2}\right) $ .

(a)

Write the expression for the $ n $ -th order characteristic function of $ \mathbf{X}\left(t\right) $ in terms of $ \mu\left(t\right) $ and $ C_{\mathbf{XX}}\left(t_{1},t_{2}\right) $ .

ref.

There are the note about the n-th order characteristic function of Gaussians random process . The only difference between the note and this problem is that this problem use the $ \mu\left(t\right) $ rather than $ \eta_{\mathbf{X}}\left(t\right)=E\left[\mathbf{X}\left(t\right)\right] $ .

Solution

$ \Phi_{\mathbf{X}\left(t_{1}\right)\cdots\mathbf{X}\left(t_{n}\right)}\left(\omega_{1},\cdots,\omega_{n}\right)=\exp\left\{ i\sum_{k=1}^{n}\mu_{\mathbf{X}}\left(t_{k}\right)\omega_{k}-\frac{1}{2}\sum_{j=1}^{n}\sum_{k=1}^{n}C_{\mathbf{XX}}\left(t_{j},t_{k}\right)\omega_{j}\omega_{k}\right\} $ .

(b)

Show that the probabilistic description of $ \mathbf{X}\left(t\right) $ is completely characterized by $ \mu\left(t\right) $ and autocovariance function $ C_{\mathbf{XX}}\left(t_{1},t_{2}\right) $ .


'Click here to view student answers and discussions

4. (25 Points)

Let $ \mathbf{X}_{1},\mathbf{X}_{2},\mathbf{X}_{3},\cdots $ be a sequence of independent, identically distributed random variables, each having Cauchy pdf $ f\left(x\right)=\frac{1}{\pi\left(1+x^{2}\right)}\;,\qquad-\infty<x<\infty. Let \mathbf{Y}_{n}=\frac{1}{n}\sum_{i=1}^{n}\mathbf{X}_{i}. $ Find the pdf of $ \mathbf{Y}_{n} $ . Describe how the pdf of $ \mathbf{Y}_{n} $ depends on $ n $ . Does the sequence $ \mathbf{Y}_{1},\mathbf{Y}_{2},\mathbf{Y}_{3},\cdots $ converge in distribution? If yes, what is the distribution of the random variable it converges to?


'Click here to view student answers and discussions


Back to ECE Qualifying Exams (QE) page

Alumni Liaison

Abstract algebra continues the conceptual developments of linear algebra, on an even grander scale.

Dr. Paul Garrett