Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 125
DIGITAL COMMUNICATIONS
Part I: Source Encoding
2000 Bijan Mobasseri 2 Why digital? Ease of signal generation Regenerative repeating capability Increased noise immunity Lower hardware cost Ease of computer/communication integration 2000 Bijan Mobasseri 3 Basic block diagram Info source Source encoder Channel encoder Digital modulation Output transducer Source decoder Channel decoder Digital demod CH channel 2000 Bijan Mobasseri 4 Some definitions Information source Raw data:voice, audio Source encoder:converts analog info to a binary bitstream Channel encoder:map bitstream to a pulse pattern Digital modulator: RF carrier modulation of bits or bauds 2000 Bijan Mobasseri 5 A bit of history Foundation of digital communication is the work of Nyquist(1924) Problem:how to telegraph fastest on a channel of bandwidth W? Ironically, the original model for communications was digital! (Morse code) First telegraph link was established between Baltimore and Washington in 1844 2000 Bijan Mobasseri 6 Nyquist theorem Nyquist theorem, still standing today, says that over a channel of bandwidth W, we can signal fastest with no interference at a rate no more than 2W Any faster and we will get intersymbol interference He further proved that the pulse shape that achieves this rate is a sinc 2000 Bijan Mobasseri 7 Signaling too fast Here is what might happen when signaling exceeds Nyquists rate Transmittted bitstream Received bitstream Pulse smearing could have been avoided if pulses had more separation, I.e. bitrate reduced 2000 Bijan Mobasseri 8 Shannon channel capacity Claude Shannon, a Bell Labs Mathematician, proved in 1948 that a communication channel is fundamentally speed-limited. This limit is given by C=Wlog 2 (1+P/N o W) bits/sec Where W is channels bandwidth, P signal power and N o is noise spectral density
2000 Bijan Mobasseri 9 Implications of channel capacity If data rate is kept below channel capacity, R<C, then t is theoretically possible to achieve error-free transmission If data rate exceeds channel capacity, error- free transmission is no longer possible 2000 Bijan Mobasseri 10 First step toward digital comm: sampling theorem Main question: can a finite number of samples of a continuous wave be enough to represent the information? OR Can you tell what the original signal was below? 2000 Bijan Mobasseri 11 How to fill in the blanks? Could you have guessed this? Is there a unique signal connecting the samples? 2000 Bijan Mobasseri 12 Sampling schemes There are at least 3 sampling schemes Ideal Flat-top Sample and hold 2000 Bijan Mobasseri 13 Ideal sampling Ideal sampling refers to the type of samples taken. Here, we are talking about impulse like(zero width) samples. T s
2000 Bijan Mobasseri 14 Ideal sampler Multiply the continuous signal g(t) with a train of impulses g(t) Eo(t-nT s ) g o (t)=Eg(nT s ) o(t-nT s ) T s
2000 Bijan Mobasseri 15 Key question What is the proper sampling rate to allow for a perfect reconstruction of the signal from its samples? To answer this question, we need to know how g(t) and g o (t) are related? 2000 Bijan Mobasseri 16 Spectrum of g o (t) g o (t) is given by the following product g o (t)=g(t)Eo(t-nT s ) Taking Fourier transform G o (f)= G(f)*{f s Eo(f-nf s ) Graphical rendition of this convolution follows next 2000 Bijan Mobasseri 17 Expanding the convolution We can exchange convolution and summation G o (f)=G(f)*{f s Eo(f-nf s )= f s E {G(f)* o (f-nf s )} Each convolution shifts G(f) to f= nf s
G(f)* o (f-nf s )} nf s G(f) 2000 Bijan Mobasseri 18 G o (f):final result Spectrum of the sampled signal is then given by
This is simply the replication of the original continuous signal at multiples of sampling rate G o (f)=f s E {G(f-nf s ) 2000 Bijan Mobasseri 19 Showing the spectrum of g o (t) Each term of the convolution is the original spectrum shifted to a multiple of sampling frequency G(f) G o (f) f s 2f s
f s
2000 Bijan Mobasseri 20 Recovering the original signal It is possible to recover the original spectrum by lowpass filtering the sampled signal
G o (f) f s 2f s
f s
LPF W W -W 2000 Bijan Mobasseri 21 Nyquist sampling rate In order to cleanly extract baseband (original) spectrum, we need sufficient separation with the adjacent sidebands Min. separation can be found as follows G o (f) f s
W f s
f s -w>W f s >2W 2000 Bijan Mobasseri 22 Sampling below Nyquist: aliasing If signal is sampled below its Nyquist rate, spectral folding, or aliasing, occurs. f s <2W Lowpass filtering will not recover the baseband spectrum intact as a result of spectral folding 2000 Bijan Mobasseri 23 Sample-and-hold A practical way of sampling a signal is sample-and-hold operation. Here is the idea:signal is sampled and its value held until the next sample 2000 Bijan Mobasseri 24 Issues Here are the questions we need to answer: What is the sampling rate now? Can the message be recovered? What price do we pay for going with a practical approach? 2000 Bijan Mobasseri 25 Modeling sample-and-hold The result of sample-and-hold can be simulated by writing the sampled signal as s(t)=Em(nT s )h(t-nT s ) Where h(t) is a basic square pulse and m(t) is the baseband message
This is a square pulse h(t) scaled by signal Sample at that point, ie m(nT s )h(t-nT s ) T s
2000 Bijan Mobasseri 26 A systems view It is possible to come up with a system that does sample-and-hold. X h(t) T s
T s
Each impulse generates a square pulse,h(t), at the output. Outputs are also spaced by Ts this we have a sample-and- hold signal h(t) Ideal sampling 2000 Bijan Mobasseri 27 Message reconstruction Key question: can we go back to the original signal after sample-and-hold ? This question can be answered in the frequency domain 2000 Bijan Mobasseri 28 Spectrum of the sample-and-hold signal Sample-and-hold signal is generated by passing an ideally sampled signal, m o (t), through a filter h(t). Therefore, we can write s(t)= m o (t)*h(t) or S(f)= M o (f)H(f) what we have available Contains message M(f) Known( it is a sinc) 2000 Bijan Mobasseri 29 Is message recoverable? Lets look at the individual components of S(f). From ideal sampling results M o (f)=f s EM(f-kf s ) M o (f) 2000 Bijan Mobasseri 30 Problems with message recovery The problem here is we dont have access to M o (f). If we did, it would be like ideal sampling What we do have access to is S(f) S(f)= M o (f)H(f) We therefore have a distorted version of an ideally sampled signal 2000 Bijan Mobasseri 31 Example message Lets show what is happening. Assume a message spectrum that is flat as follows W -W M(f) M o (f) fs 2fs 2000 Bijan Mobasseri 32 Sample-and-hold spectrum We dont see M o (f). We see M o (f)H(f). Since h(t) was a square pulse of width T s , H(f) is sinc(fT s ) . M o (f). H(f) f f 1/Ts=fs W 2000 Bijan Mobasseri 33 Distortion potential The original analog message is in the lowpass term of M o (f) H(f) through the product M o (f)H(f) causes a distortion of this term. Lowpass filtering of the sample-and-hold signal will only recover a distorted message 2000 Bijan Mobasseri 34 Illustrating distortion H(f) f 2fs 1/Ts=fs W M o (f) Sample and hold signal. If lowpass filtered, the original Message is not recovered want to recover this What is actually recovered fs 2000 Bijan Mobasseri 35 How to control distortion? In order to minimize the effect of H(f) on reconstruction, we must make H(f) as flat as possible in the message bandwidth(-W,W) What does it mean? It means move the first zero crossing to the right by increasing the sampling rate, or decreasing pulse width 2000 Bijan Mobasseri 36 Does it make sense?
The narrower the pulse, hence higher sampling rate, the more accurate you can capture signal variations 2000 Bijan Mobasseri 37 Variation on sample-and-hold Contrast the two following arrangements sample period and pulse width are not the same t Ts 2000 Bijan Mobasseri 38 How does this affect reconstruction? The only thing that will change is h(t) and hence H(f) H(f) f f 1/t W M o (f) Sample and hold signal. If lowpass filtered, the original Message is not recovered want to recover this What is actually recovered different zero crossing 2000 Bijan Mobasseri 39 How to improve reconstruction? Again, we need to flatten out H(f) within (- W,W). and the way to do it is to use narrower pulses (smaller t) 2000 Bijan Mobasseri 40 Sample-and-hold converges to ideal sampling If reducing the pulse width of h(t) is a good idea, why not take it to the limit and make them zero? We can do that in which case sample-and- hold collapses to ideal sampling(impulses are zero width pulses) Pulse Code Modulation Filtering, Sampling, Quantization and Encoding 2000 Bijan Mobasseri 42 Elements of PCM Transmitter Encoder consists of 5 pieces
Transmission path Continuous message LPF Sampler Quantizer Encoder Regenerative repeater Regenerative repeater 2000 Bijan Mobasseri 43 Quantization Quantization is the process of taking continuous samples and converting them to a finite set of discrete levels 1.2 1.52 .86 -0.41 ? 2000 Bijan Mobasseri 44 Defining a quantizer Quantizer is defined by its input/output characteristics; continuous values in, discrete values out in out in out Midtread type Midrise type Output remains constant Even as input varies over a range 2000 Bijan Mobasseri 45 Quantization noise/error Quantizer clearly discards some information. Question is how much error is committed? q(m) Message(m) Quantized message (v) Error=q=m-v 2000 Bijan Mobasseri 46 Illustrating quantization error
Sampled quantized Quantization error v1 v2 v3 A A quantizer step size 2000 Bijan Mobasseri 47 More on A A Controls how fine samples are quantized. Equivalently, A controls quantization error. To determine A we need to know two parameters Number of quantization levels Dynamic range of the signal 2000 Bijan Mobasseri 48 A for a uniform quantizer Let sample values lie in the range ( -m max , +m max ). We also want to have exactly L levels at the output of the quantizer. Simple math tells us max min L levels A=2m max /L 2000 Bijan Mobasseri 49 Quantization error bounds Quantization error is bounded by half the step size Level 2 Level 1 Error q A |q|<A/2 Error q 2000 Bijan Mobasseri 50 Statistics of q Quantization error is random. It can be positive or negative with equal probability. This is an example of a uniformly distributed random variable. q Density function f(q) A/2 -A/2 1/A 2000 Bijan Mobasseri 51 Quantization noise power Any uniformly distributed random variable in the range (-a/2 to a/2) has an average power(variance) given by a 2 /12. Here, quantization noise range is A, therefore o 2 q = A 2 /12 2000 Bijan Mobasseri 52 Signal-to-quantization noise Leaving aside random noise, there is always a finite quantization noise. Let the original continuous signal have power P=<m 2 (t)> and quantization noise variance(power) o 2 q (SNR) q =P/ o 2 q =12P/ A 2 2000 Bijan Mobasseri 53 Substituting for A We have related step size to signal dynamic range and number of quantization levels
Therefore, signal to quantization noise(sqnr) sqnr=(SNR)q=[3P/m 2 max ]L 2
A=2m max /L 2000 Bijan Mobasseri 54 Example Let m(t)=cos(2tf m t). What is the signal to quantization noise ratio(sqnr) for a 256- level quantizer Average message power P is 0.5, therefore sqnr=(3x0.5/1)256 2 =98304~50dB 2000 Bijan Mobasseri 55 Nonuniform quantizer Uniform quantization is a fantasy. Reason is that signal amplitude is not equally spread out. It occupies mostly low amplitude levels 2000 Bijan Mobasseri 56 Solution:nonuniform intervals Quantize fine where amplitudes spend most of their time 2000 Bijan Mobasseri 57 Implementing nonuniform quantization:companding Signal is first processed through a nonlinear device that stretches low amplitudes and compresses large amplitudes input output Low amplitudes stretched Large amplitudes pressed 2000 Bijan Mobasseri 58 A-law and -law There are two companding curves, A-law and -law. Both are very similar Each has an adjustment parameter that controls the degree of companding (slope of the curve) Following companding, a uniform quantization is used 2000 Bijan Mobasseri 59 Encoder Quantizer outputs are merely levels. We need to convert them to a bitstream to finish the A/D operation There are many ways of doing this Natural coding Gray coding 2000 Bijan Mobasseri 60 Natural coding How many bits does it take to represent L- levels? The answer is n=log 2 L bits/sample Natural coding is a simple decimal to binary conversion 0000 1001 2010 3011 . 7111 Encoder output(3 bits per sample Quantizer levels(8) 2000 Bijan Mobasseri 61 Gray coding Here is the problem with natural coding: if levels 2(010) and 1(001) are mistaken, then we suffer two bit errors We want an encoding scheme that assigns code words to adjacent levels that differ in at most one bit location 2000 Bijan Mobasseri 62 Gray coding example Take a 4-bit quantizer (16 levels). Adjacent levels differ by juts one bit
0 0 0 0 1 1 0 0 0 0 2 0 1 0 0 3 0 1 0 1 4 1 1 0 1 . 2000 Bijan Mobasseri 63 Quantizer word size Knowing n, we can refer to n-bit quantizers For example, if L=256 with n=8bits/sample We are then looking at an 8-bit quantizer 2000 Bijan Mobasseri 64 Interaction between sqnr and bit/sample Converting sqnr to dB provides a different insight. Take 10log 10 (sqnr) sqnr=kL 2 where k=[3P/m 2 max ] In dB (sqnr) dB =o+20logL= o+20log2 n (sqnr) dB = o+6n dB 2000 Bijan Mobasseri 65 sqnr varies linearly with bits/sample What we just saw says higher sqnr is achieved by increasing n(bits/sample). Question then is, what keeps us from doing that for ever thus getting arbitrarily large sqnrs? 2000 Bijan Mobasseri 66 Cost factor We can increase number of bits/sample hence quantization levels but at a cost The cost is in increased bandwidth but why? One clue is that as we go to finer quantization, levels become tightly packed and difficult to discern at the receiver hence higher error rates. There is also a bandwidth cost
2000 Bijan Mobasseri 67 Basis for finding PCM bandwidth Nyquist said in a channel with transmission bandwidth B T , we can transmit at most 2B T
pulses per second: R(pulses/second)<2B T (Hz) Or B T (Hz)>R/2(pulses/second) 2000 Bijan Mobasseri 68 Transmission over phone lines Analog phone lines are limited to 4KHz in bandwidth, what is the fastest pulse rate possible? R<2BT=2x4000=8000 pulses/sec Thats it? Modems do a bit faster than this! One way to raise this rate is to stuff each pulse with multiple bits. More on that later 2000 Bijan Mobasseri 69 Accomodating a digital source A source is generating a million bits/sec. What is the minimum required transmission bandwidth. B T >R/2=10 6 /2=500 KHz 2000 Bijan Mobasseri 70 PCM bit rate The bit rate at the output of encoder is simply the following product R(bits/sec)=n(bits/sample)xf s (samples/sec) R=nf s bits/sec
1 0 1 1 0 1 quantized Encoded at 5 bits/sample 2000 Bijan Mobasseri 71 PCM bandwidth But we know sampling frequency is 2W. Substituting f s =2W in R=n f s
R=2nW (bits/sec) We also had B T >R/2. Replacing R we get B T >nW
2000 Bijan Mobasseri 72 Comments on PCM bandwidth We have established a lower bound(min) on the required bandwidth. The cost of doing PCM is the large required bandwidth. The way we can measure it is Bandwidth expansion quantified by B T /W>n (bits/sample) 2000 Bijan Mobasseri 73 Bandwidth expansion factor Similar to FM, there is a bandwidth expansion factor relative to baseband, i.e. |=B T /W>n Lets say we have 8 bits/sample meaning it takes , at a minimum, 8 times more than baseband bandwidth to do PCM 2000 Bijan Mobasseri 74 PCM bandwidth example Want to transmit voice (~4KHz ) using an 8-bit PCM. How much bandwidth is needed? We know W=4KHz, fs=8 KHz and n=8. B T >nW=8x4000=32KHz This is the minimum PCM bandwidth under ideal conditions. Ideal has to do with pulse shape used 2000 Bijan Mobasseri 75 Bandwidth-power exchange We said using finer quantization (more bits/sample) enhances sqnr because (sqnr) dB = o+6n dB At the same time we showed bandwidth increases linearly with n. So we have a trade-off
2000 Bijan Mobasseri 76 sqnr improvement Lets say we increase n by 1 from 8 to 9 bits/sample. As result, sqnr increases by 6 dB sqnr= o+6x8= o+48 sqnr= o+6x9= o+54
+6dB 2000 Bijan Mobasseri 77 Bandwidth increase Going from n= 8 bits/sample, to 9 bits/sample, min. bandwidth rises from 8W to 9W. If message bandwidth is 4 KHz, then B T =32 KHz for n=8 B T =36 KHz for n=9 +4 KHz or 12.5% increase 2000 Bijan Mobasseri 78 Is it worth it? Lets look at the trade-off: Cost in increased bandwidth:12.5% Benefit in increased sqnr: 6dB Every 3 dB means a doubling of the sqnr ratio. So we have quadrupled sqnr by paying 12.5% more in bandwidth 2000 Bijan Mobasseri 79 Another way to look at the exchange We provided 12.5% more bandwidth and ended up with 6 dB more sqnr. If we are satisfied with the sqnr we have, we can dial back transmitted power by 6 dB and suffer no loss in sqnr In other words, we have exchanged bandwidth for lower power 2000 Bijan Mobasseri 80 Similarity with FM PCM and FM are examples of wideband modulation. All such modulations provide bandwidth-power exchange but at different rates. Recall |=B T /W FM.SNR~| 2 PCM..SNR~2 2|
Much more sensitive to beta, Better exchnage 2000 Bijan Mobasseri 81 Complete PCM system design Want to transmit voice with average power of 1/2 watt and peak amplitude 1 volt using 256 level quantizer. Find sqnr Bit rate PCM bandwidth 2000 Bijan Mobasseri 82 Signal to quantization noise We had sqnr=[3P/m 2 max ]L 2 We have L=256, P=1/2 and m max =1.
sqnr=98304~50 dB
2000 Bijan Mobasseri 83 PCM bitrate Bit rate is given by R=2nW (bits/sec)=2x8x4000=64 Kb/sec This rate is a standard PCM voice channel This is why we can have 56K transmission over the digital portion of telephone network which can accomodating 64 Kb/sec. 2000 Bijan Mobasseri 84 PCM bandwidth We can really talk about minimum bandwidth given by B T | min =nW=8x4000=32 KHz In other words, we need a minimum of 32 KHz bandwidth to transmit 64 KB/sec of data. 2000 Bijan Mobasseri 85 Realistic PCM bandwidth Rule of thumb to find the required bandwidth for digital data is that bandwidth=bit rate B T =R So for 64 KB/sec we need 64 KHz of bandwidth One hertz per bit 2000 Bijan Mobasseri 86 Differential PCM Concept of differential encoding is of great importance in communications The underlying idea is not to look at samples individually but to look at past values as well. Often, samples change very little thus a substantial compression can be achieved 2000 Bijan Mobasseri 87 Why differential? Lets say we have a DC signal and blindly go about PCM-encoding it. Is it smart?
Clearly not. What we have failed to realize is that samples dont change. We can send the first sample and tell the receiver that the rest are the same 2000 Bijan Mobasseri 88 Definition of differential encoding We can therefore say that in differential encoding, what is recorded and ultimately transmitted is the change in sample amplitudes not their absolute values We should send only what is NEW.
2000 Bijan Mobasseri 89 Where is the saving? Consider the following two situations
The right samples are adjacent sample differences with much smaller dynamic range requiring fewer quantization levels 2 1.6 0.8 1.6 2 2 1.6 2 2 -0.8 -0.4 0 0.4 -0.4 0.4 0.8 2000 Bijan Mobasseri 90 Implementation of DPCM:prediction At the heart of DPCM is the idea of prediction Based on n-1 previous samples, encoder generates an estimate of the nth sample. Since the nth sample is known, prediction error can be found. This error is then transmitted 2000 Bijan Mobasseri 91 Illustrating prediction Here is what is happening at the transmitter Past samples(already sent) To be trasmited Prediction of the Current sample Prediction error Only Prediction error is sent 2000 Bijan Mobasseri 92 What does the receiver do? Receiver has the identical prediction algorithm available to it. It has also received all previous samples so it can make a prediction of its own Transmitter helps out by supplying the prediction error which is then used by the receiver to update the predicted value 2000 Bijan Mobasseri 93 Interesting speculation What if our power of prediction was perfect? In other words, what if we could predict the next sample with no error?. What kind of communication system would be looking at? 2000 Bijan Mobasseri 94 Prediction error Let m(t) be the message and Ts sample interval, then prediction error is given
e(nT s ) = m(nT s ) m nT s ( ) Prediction error 2000 Bijan Mobasseri 95 Prediction filter Prediction is normally done using a weighted sum of N previous samples
The quality of prediction depends on the good choice of weights w i
m nT s ( ) = w i m n i ( )T s ( ) i =1 N
2000 Bijan Mobasseri 96
Finding the optimum filter How do you find the best weights? Obviously, we need to minimize the prediction error. This is done statistically
Choose a set of weights that gives the lowest (on average) prediction error over w Min e 2 nT s ( ) { } 2000 Bijan Mobasseri 97 Prediction gain Prediction provides an SNR improvement by a factor called prediction gain G p = o M 2 o e 2 = message power prediction error power 2000 Bijan Mobasseri 98 How much gain? On average, this gain is about 4-11 dB. Recall that 6 dB of SNR gain can be exchanged for 1 bit per sample At 8000 samples/sec(for speech) we can save 1 to 2 bits per sample thus saving 8-16 Kb/sec. 2000 Bijan Mobasseri 99 DPCM encoder
Prediction error is used to correct the estimate in time for the next round of prediction + quantizer encoder + N-tap prediction Prediction error Prediction - + Prediction error Updated prediction Input sample 2000 Bijan Mobasseri 100 Delta modulation (DM) DM is actually a very simplified form of DPCM In DM, prediction of the next sample is simply the previous sample Estimate of Prediction error 2000 Bijan Mobasseri 101 DM encoder-diagram + 1-bit quantizer + Delay Ts Prediction error(A) Prediction - + Prediction error Updated prediction Input sample in out A -A 2000 Bijan Mobasseri 102 DM encoder operation Prediction error generates A at the output of quantizer If error is positive, it means prediction is below sample value in which case the estimate is updated by + A for the next step 2000 Bijan Mobasseri 103 Slope overload effect Signal rises faster than prediction: A too small samples Ts initial estimate A predictions 2000 Bijan Mobasseri 104 Steady state: granular noise Prediction can track the signal; prediction error small Two drops to reach the signal A 2000 Bijan Mobasseri 105 Shortcomings of DM It is clearly the prediction stage that is lacking Samples must be closely taken to insure that previous-sample prediction algorithm is reasonably accurate This means higher sample rates 2000 Bijan Mobasseri 106 Multiplexing Concurrent communications calls for some form of multiplexing. There are 3 categories FDMA(frequency division multiple access) TDMA(time division multiple access) CDMA(code division multiple access) All 3 enjoy a healthy presence in the communications market 2000 Bijan Mobasseri 107 FDMA In FDM, multiple users can be on at the same time by placing them in orthogonal frequency bands guardband user 1 user 2 user N TOTAL BANDWIDTH 2000 Bijan Mobasseri 108 FDMA example:AMPS AMPS, wireless analog standard, is a good example Reverse link(mobile-to-base): 824-849MHz Forward link: 869-894 MHz channel bandwidth:30 KHz total # channels: 833 Modulation: FM, peak deviation 12.5 KHz
2000 Bijan Mobasseri 109 TDMA Where FDMA is primarily an analog standard, TDMA and CDMA are for digital communication In TDMA, each user is assigned a time slot, as opposed to a frequency slot in FDMA
2000 Bijan Mobasseri 110 Basic idea behind TDMA Take the following 3 digital lines
frame 2000 Bijan Mobasseri 111 TDM-PCM quantizer and encoder quantizer and encoder channel decoder TDM-PAM TDM-PCM(bits) lpf lpf 2000 Bijan Mobasseri 112 Parameters of TDM-PCM A TDM-PCM line multiplexing M users is characterized by the following parameters data rate(bit or pulse rate) bandwidth 2000 Bijan Mobasseri 113 TDM-PCM Data rate Here is what we have M users Each sampled at Nyquist rate Each sample PCMd into n bit words Total bit rate then is R=M(users)xf s (samples /sec/user)xn(bits/sec) =nMf s bits sec 2000 Bijan Mobasseri 114 TDM-PCM bandwidth Recall Nyquist bandwidth. Given R pulses per second, we need at least R/2 Hz. In reality we need more (depending on the pulse shape) so B T =R=nMf s Hz 2000 Bijan Mobasseri 115 T1 line Best known of all TDM schemes is AT&Ts T1 line T1 line multiplexes 24 voice channels(4KHz) into one single bitstream running at the rate of 1.544 Mb/sec. Lets see how 2000 Bijan Mobasseri 116 T1 line facts Each of the 24 voice lines are sampled at 8 KHz Each sample is then encoded into 8 bits A frame consists of 24 samples, one from each line Some data bits are preempted for control and supervisory signaling 2000 Bijan Mobasseri 117 T1 line structure: all frames except 1,7,13,19...
1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 channel 1 channel 2 channel 24 FRAME(repeats) information bits (8-bits per sample) 2000 Bijan Mobasseri 118 Inserting non-data bits In addition to data, we need slots for signaling bits (on-hook/off hook, charging) Every 6th frame (1,7,13,19..) is selected and the least significant bit per channel is replaced by a signaling bit 1 2 3 4 5 6 7 1 2 3 4 5 6 7 1 2 3 4 5 6 7 channel 1 channel 2 channel 24 2000 Bijan Mobasseri 119 Framing bit Timing is of utmost significance in T1. We MUST be able to know where the beginning of each frame is At the end of each frame a single bit is added to help with frame identification 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 channel 1 channel 2 channel 24 information bits (8-bits per sample) F 2000 Bijan Mobasseri 120 T1 frame length How long is one frame?One revolution generates frame 24 sampled at 8KHz rotates at 8000 revs/sec. frame length=1/8000= 125 microseconds 2000 Bijan Mobasseri 121 T1 bit rate per frame Data rate 8x24=192 bits per frame Framing bit rate 1 bit per frame Total per frame 193 bits/frame
2000 Bijan Mobasseri 122 Total T1 bit rate We know there are 8000 frames a sec. and there are 193 bits per frame. Therefore
T1 rate=193x8000=1.544 Mb/sec 2000 Bijan Mobasseri 123 Signaling rate component Not all 1.544 Mb/sec is data. In every 6th frame, we replace 24 data bits by signaling bits. Therefore signaling rate= (8000 frames/sec)(1/6)(24 bits)=32 Kbits/sec 2000 Bijan Mobasseri 124 TDM hierarchy It is possible to build upon T1 as follows 1st level multiplexer 24 64 kb/sec 2nd level multiplexer 3rd level multiplexer DS-0 DS-1 DS-2 DS-3 DS-1: 1.544 MB/sec DS-2: 6.312 Mb/sec DS-3: 44.736 Mb/sec 7 lines 2000 Bijan Mobasseri 125 Recommended problems 6.2 6.15 6.17