>> Welcome to module one of Digital Signal Processing. In this module we are going to see what signals actually are. We are going to go through a history, see the earliest example of these discrete-time signals. Actually it goes back to Egyptian times. Then through this history see how digital signals, for example with the telegraph signals, became important in communications. And today, how signals are pervasive in many applications, in every day life objects. For this we're going to see what the signal is, what a continuous time analog signal is, what a discrete-time, continuous-amplitude signal is and how these signals relate to each other and are used in communication devices. We are not going to have any math in this first module. It is more illustrative and the mathematics will come later in this class. This is an introduction to what digital signal processing is all about. Before getting going, let's give some background material. There is a textbook called Signal Processing for Communications by Paolo Prandoni and myself. You can have a paper version or you can get the free PDF or HTML version on the website here indicated on the slide. There will be quizzes, there will be homework sets, and there will be occasional complementary lectures. What is actually a signal? We talk about digital signal processing, so we need to define what the signal is. Typically, it's a description of the evolution over physical phenomenon. Quite simply, if I speak here, there is sound pressure waves going through the air that's a typical signal. When you listen to the speech, there is a loud speaker creating a sound pressure waves that reaches your ear. And that's another signal. However, in between is the world of digital signal processing because after the microphone it gets transformed in a set of members. It is processed in the computer. It is being transferred through the internet. Finally it is decoded to create the sound pressure wave to reach your ears. Other example are the temperature evolution over time, the magnetic deviation for example, L P recording , is a grey level on paper for a black and white photograph,some flickering colors on TV screen. Here we have a thermometer recording temperature over time. So you see the evolution And there are discrete ticks and you see how it changes over time. So what are the characteristics of digital signals. There are two key ingredients. First there is discrete time. As we have seen in the previous slide on the horizontal axis there are discrete Evenly spaced ticks and that corresponds to discretisation in time. There is also discrete amplitude because the numbers that are measured will be represented in a computer and cannot have some infinite precision. So what amount more sophisticated things, functions, derivative, and integrals. The question of discreet versus continuous, or analog versus discreet, goes probably back to the earliest time of science, for example, the school of Athens. There was a lot of debate between philosophers and mathematicians about the idea of continuum, or the difference between countable things and uncountable things. So in this picture, you see green are famous philosophers like Plato, in red, famous mathematicians like Pythagoras, somebody that we are going to meet again in this class, and there is a famous paradox which is called Zeno's paradox. So if you should narrow will it ever arrive in destination? We know that physics Allows us to verify this but mathematics have the problem with this and we can see this graphically. So you want to go from A to B, you cover half of the distance that is C, center quarters that's D and also eighth that's E etc will you ever get there and of course we know you gets there because the sum from 1 to infinity of 1 over 2 to the n is equal to 1, a beautiful formula that we'll see several times reappearing in this. Unfortunately during the middle ages in Europe, things were a bit lost. As you can see, people had other worries. In the 17th century things picked up again. Here we have a physicist and astronomer Galileo, and the philosopher Rene Descartes, and both contributed to the advancement of mathematics at that time. Descartes' idea was simple but powerful. Start with a point, put it into a co-ordinate system Then put more sophisticated things like lines, and you can use algebra. This led to the idea of calculus, which allowed to mathematically describe physical phenomenon. For example Galileo was able to describe the trajectory of a bullet, using infinite decimal variations in both horizontal and vertical direction. Calculus itself was formalized by Newton and Leibniz, and is one of the great advances of mathematics in the 17th and 18th century. It is time to do some very simple continuous time signal processing. We have a function in blue here, between a and b, and we would like to compute it's average. As it is well known, this well be the integral of the function, divided by the length's of the interval, and it is shown here in red dots. What would be the equivalent in this discreet time symbol processing. We have a set of samples between say, 0 and capital N minus 1. The average is simply 1 over n, the sum Was the antidote terms x[n] between 0 and N minus 1. Again, it is shown in the red dotted line. In this case, because the signal is very smooth, the continuous time average and the discrete time average Are essentially the same. This was nice and easy but what if the signal is too fast, and we don't know exactly how to compute either the continuous time operations or an equivalent operation on samples. Enters Joseph Fourier, one of the greatest mathematicians of the nineteenth century. And the inventor of Fourier series, Fourier analysis which are essentially the ground tools of signal processing. We show simply a picture to give the idea of Fourier analysis. It is a local Fourier spectrum as you would see for example on an equalizer table in a disco. And it shows the distribution of power across frequencies, something we are going to understand in detail in this class. But to do this quick time processing of continuous time signals we need some further results. And these were derived by Harry Niquist and Claude Shannon, two researchers at Bell Labs. They derived the so-called sampling theorem, first appearing in 1920's and formalized in 1948. If the function X of T is sufficiently slow then there is a simple interpolation formula for X of T, it's the sum of the samples Xn, Interpolating with the function that is called sync function. It looks a little but complicated now, but it's something we're going to study in great detail because it's 1 of the fundamental formulas linking this discrete time and continuous time signal processing. Let us look at this sampling in action. So we have the blue curve, we take samples, the red dots from the samples. We use the same interpolation. We put one blue curve, second one, third one, fourth one, etc. When we sum them all together, we get back the original blue curve. It is magic. This interaction of continuous time and discrete time processing is summarized in these two pictures. On the left you have a picture of the analog world. On the right you have a picture of the discrete or digital world, as you would see in a Digital camera for example, and this is because the world is analog. It has continuous time continuous space, and the computer is digital. It is discreet time discreet temperature. When you look at an image taken with a digital camera, you may wonder what the resolution is. And here we have a picture of a bird. This bird happens to have very high visual acuity, probably much better than mine. Still, if you zoom into the digital picture, after a while, around the eye here, you see little squares appearing, showing indeed that the picture is digital Because discrete values over the domain of the image and it also has actually discrete amplitude which we cannot quite see here at this level of resolution. As we said the key ingredients are discrete time and discrete amplitude for digital signals. So, let us look at x of t here. It's a sinusoid, and investigate discrete time first. We see this with xn and discrete amplitude. We see this with these levels of the amplitudes which are also discrete ties. And so this signal looks very different from the original continuous time signal x of t. It has discrete values on the time axes and discrete values on the vertical amplitude axis. So why do we need digital amplitude? Well, because storage is digital, because processing is digital, and because transmission is digital. And you are going to see all of these in sequence. So data storage, which is of course very important, used to be purely analog. You had paper. You had wax cylinders. You had vinyl. You had compact cassettes, VHS, etcetera. In imagery you had Kodachrome, slides, Super 8, film etc. Very complicated, a whole biodiversity of analog storages. In digital, much simpler. There is only zeros and ones, so all digital storage, to some extent, looks the same. The storage medium might look very different, so here we have a collection of storage from the last 25 years. However, fundamentally there are only 0's and 1's on these storage devices. So in that sense, they are all compatible with each other. Processing also moved from analog to digital. On the left side, you have a few examples of analog processing devices, an analog watch, an analog amplifier. On the right side you have a piece of code. Now this piece of code could run on many different digital computers. It would be compatible with all these digital platforms. The analog processing devices Are essentially incompatible with each other. Data transmission has also gone from analog to digital. So lets look at the very simple model here, you've on the left side of the transmitter, you have a channel on the right side you have a receiver. What happens to analog signals when they are send over a channel. So x of t goes through the channel, its first multiplied by 1 over G because there is path loss and then there is noise added indicated here with the sigma of t. The output here is x hat of t. Let's start with some analog signal x of t. Multiply it by 1 over g, and add some noise. How do we recover a good reproduction of x of t? Well, we can compensate for the path loss, so we multiply by g, to get xhat 1 of t. But the problem is that x1 hat of t, is x of t. That's the good news plus g times sigma of t so the noise has been amplified. Let's see this in action. We start with x of t, we scale by G, we add some noise, we multiply by G. And indeed now, we have a very noisy signal. This was the idea behind trans-Atlantic cables which were laid in the 19th century and were essentially analog devices until telegraph signals were properly encoded as digital signals. As can be seen in this picture, this was quite an adventure to lay a cable across the Atlantic and then to try to transmit analog signals across these very long distances. For a long channel because the path loss is so big, you need to put repeaters. So the process we have just seen, would be repeated capital N times. Each time the paths loss would be compensated, but the noise will be amplified by a factor of n. Let us see this in action, so start with x of t, paths loss by g, added noise, amplification by G with the amplification the amplification of the noise, and the signal. For the second segment we have the pass loss again, so X hat 1 is divided by G. And added noise, then we amplify to get x hat 2 of t, which now has twice an amount of noise, 2 g times signal of t. So, if we do this n times, you can see that the analog signal, after repeated amplification. Is mostly noise. And that becomes problematic to transmit information. In digital communication, the physics do not change. We have the same path loss, we have added noise. However, two things change. One is that we don't send arbitrary signals but, for example, only signals that[INAUDIBLE]. Take values plus 1 and minus 1, and we do some specific processing to recover these signals. Specifically at the outward of the channel, we multiply by g, and then we take the signa operation. So x1hat, is signa of x of t, plug g times sigma of t. Let us again look at this in action. We start with the signal x of t that is easier, plus 5 or minus 5. 5. It goes through the channel, so it loses amplitude by a factor of g, and their is some noise added. We multiply by g, so we recover x of t plus g times the noise of sigma t. Then we apply the threshold operation. And true enough, we recover a plus 5 minus 5 signal, which is identical to the ones that was sent on the channel. Thanks to digital processing the transmission of information has made tremendous progress. In the mid nineteenth century a transatlantic cable would transmit 8 words per minute. That's about 5 bits per second. A hundred years later a coaxial cable with 48 voice channels. At already 3 megabits per second. In 2005, fiber optic technology allowed 10 terabits per second. A terabit is 10 to the 12 bits per second. And today, in 2012, we have fiber cables with 60 terabits per second. On the voice channel, the one that is used for telephony, in 1950s you could send 1200 bits per second. In the 1990's, that was already 56 kilobits per second. Today, with ADSL technology, we are talking about 24 megabits per second. Please note that the last module in the class will actually explain how ADSL The works using all the tricks in the box that we are learning in this class. It is time to conclude this introductory module. And we conclude with a picture. If you zoom into this picture you see it's the motto of the class, signal is strength.