Aug 28, 2013 #1 D David83 Advanced Member level 1 Joined Jan 21, 2011 Messages 410 Helped 45 Reputation 92 Reaction score 45 Trophy points 1,308 Activity points 3,639 Hello, I am trying to simulate an OFDM system where the transmitted symbols are BPSK. The system goes like this: Generate binary stream of N bits Convert it to bipolar stream Take the oversampled IFFT (multipled by K) with oversampling ratio Ns (K=N*Ns) + CP D/A Conversion and baseband transmission Add noise A/D conversion & Remove CP & FFT (divided by K) operation Take the first N symbols Do ML detection My question is how to generate the noise such that the SNR of subcarrier k is |Hk|^2*SNR? I do it like this: \[n=\sqrt{\frac{N\,Ns}{SNR}}(n_R+j\,n_I)\] where n_R and n_I are both normalized Gaussian random variables. Is this correct?
Hello, I am trying to simulate an OFDM system where the transmitted symbols are BPSK. The system goes like this: Generate binary stream of N bits Convert it to bipolar stream Take the oversampled IFFT (multipled by K) with oversampling ratio Ns (K=N*Ns) + CP D/A Conversion and baseband transmission Add noise A/D conversion & Remove CP & FFT (divided by K) operation Take the first N symbols Do ML detection My question is how to generate the noise such that the SNR of subcarrier k is |Hk|^2*SNR? I do it like this: \[n=\sqrt{\frac{N\,Ns}{SNR}}(n_R+j\,n_I)\] where n_R and n_I are both normalized Gaussian random variables. Is this correct?