# Signals and Systems/Time Domain Analysis

There are many tools available to analyze a system in the time domain, although many of these tools are very complicated and involved. Nonetheless, these tools are invaluable for use in the study of linear signals and systems, so they will be covered here.

## Linear Time-Invariant (LTI) SystemsEdit

This page will contain the definition of a LTI system and this will be used to motivate the definition of convolution as the output of a LTI system in the next section. To begin with a system has to be defined and the LTI properties have to be listed. Then, for a given input it can be shown (in this section or the following) that the output of a LTI system is a convolution of the input and the system's impulse response, thus motivating the definition of convolution.

Consider a system for which an input of *x _{i}(t)* results in an output of

*y*respectively for

_{i}(t)*i = 1, 2*.

### LinearityEdit

There are 2 requirements for *linearity*. A function must satisfy both to be called "linear".

**Additivity**: An input of results in an output of .**Homogeneity**: An input of results in an output of

Being linear is also known in the literature as "satisfying the principle of superposition". **Superposition** is a fancy term for saying that the system is additive and homogeneous. The terms linearity and superposition can be used interchangeably, but in this book we will prefer to use the term linearity exclusively.

We can combine the two requirements into a single equation: In a linear system, an input of results in an output of .

### AdditivityEdit

A system is said to be additive if a sum of inputs results in a sum of outputs. To test for additivity, we need to create two arbitrary inputs, *x _{1}(t)* and

*x*. We then use these inputs to produce two respective outputs:

_{2}(t)Now, we need to take a sum of inputs, and prove that the system output is a sum of the previous outputs:

If this final relationship is not satisfied for *all possible inputs*, then the system is not additive.

### HomogeneityEdit

Similar to additivity, a system is homogeneous if a scaled input (multiplied by a constant) results in a scaled output. If we have two inputs to a system:

Where

Where *c* is an arbitrary constant. If this is the case then the system is homogeneous if

for any arbitrary *c*.

### Time InvarianceEdit

If the input signal *x(t)* produces an output *y(t)* then any time shifted input, *x(t + δ)*, results in a time-shifted output *y(t + δ)*.

This property can be satisfied if the transfer function of the system is not a function of time except expressed by the input and output.

### Example: Simple Time InvarianceEdit

To demonstrate how to determine if a system is time-invariant, consider the two systems:

- System A:
- System B:

Since system A explicitly depends on *t* outside of *x(t)* and *y(t)*, it is **time-variant**. System B, however, does not depend explicitly on *t* so it is time-invariant (given *x(t)* is time-invariant).

### Example: Formal ProofEdit

A more formal proof of why systems A & B from above are respectively time varying and time-invariant is now presented. To perform this proof, the second definition of time invariance will be used.

- System A
- Start with a time shift of the input
- Now describe the time shifted output by δ
- Clearly , therefore the system is not time-invariant.

- System B
- Start with a time shift of the input
- Now time shift the output by δ
- Clearly , therefore the system is time-invariant.

## Linear Time Invariant (LTI) SystemsEdit

The system is linear time-invariant (LTI) if it satisfies both the property of linearity and time-invariance. This book will study LTI systems almost exclusively, because they are the easiest systems to work with, and they are ideal to analyze and design.

## Other Function PropertiesEdit

Besides being linear, or time-invariant, there are a number of other properties that we can identify in a function:

### MemoryEdit

A system is said to have memory if the output from the system is dependent on past inputs (or future inputs) to the system. A system is called *memoryless* if the output is only dependent on the current input. Memoryless systems are easier to work with, but systems with memory are more common in digital signal processing applications.

### CausalityEdit

Causality is a property that is very similar to memory. A system is called **causal** if it is only dependent on past or current inputs. A system is called **non-causal** if the output of the system is dependent on future inputs. Most of the practical systems are causal.

### StabilityEdit

Stability is a very important concept in systems, but it is also one of the hardest function properties to prove. There are several different criteria for system stability, but the most common requirement is that the system must produce a finite output when subjected to a finite input. For instance, if we apply 5 volts to the input terminals of a given circuit, we would like it if the circuit output didn't approach infinity, and the circuit itself didn't melt or explode. This type of stability is often known as "Bounded Input, Bounded Output" stability, or BIBO.

Studying BIBO stability is a relatively complicated course of study, and later books on the Electrical Engineering bookshelf will attempt to cover the topic.

## Linear OperatorsEdit

Mathematical operators that satisfy the property of linearity are known as **linear operators**. Here are some common linear operators:

- Derivative
- Integral
- Fourier Transform

### Example: Linear FunctionsEdit

Determine if the following two functions are linear or not:

## Impulse ResponseEdit

**Impulse response** tells us how a system **reacts** when we hit it with an **impulse signal**(also called as **dirac delta function**).This Impulse response is very important term in analyzing the behaviour of systems.

### Zero-Input ResponseEdit

### Zero-State ResponseEdit

zero state response means steady state or forced response. This is the system response y(t) to an input f(t) when the system is in zero state; that is, when all initial conditions are zero.

### Second-Order SolutionEdit

- Example. Finding the total response of a driven RLC circuit.

## ConvolutionEdit

Convolution (folding together) is a complicated operation involving integrating, multiplying, adding, and time-shifting two signals together.

The convolution *a * b* of two functions *a* and *b* is defined as the function:

The Greek letter τ (tau) is used as the integration variable, because the letter *t* is already in use. τ is used as a "dummy variable" because we use it merely to calculate the integral.

In the convolution integral, all references to *t* are replaced with τ, except for the *-t* in the argument to the function *b*. Function *b* is *time inverted* by changing τ to -τ. Graphically, this process moves everything from the right-side of the *y* axis to the left side and vice-versa. Time inversion turns the function into a mirror image of itself.

Next, function *b* is *time-shifted* by the variable *t*. Remember, once we replace everything with τ, we are now computing in the *tau domain*, and not in the time domain like we were previously. Because of this, *t* can be used as a shift parameter.

We multiply the two functions together, time shifting along the way, and we take the area under the resulting curve at each point. Two functions overlap in increasing amounts until some "watershed" after which the two functions overlap less and less. Where the two functions overlap in the *t* domain, there is a value for the convolution. If one (or both) of the functions do not exist over any given range, the value of the convolution operation at that range will be zero.

After the integration, the definite integral plugs the variable *t* back in for remaining references of the variable τ, and we have a function of *t* again. It is important to remember that the resulting function will be a combination of the two input functions, and will share some properties of both.

### Properties of ConvolutionEdit

The convolution function satisfies certain conditions:

- Commutativity

- Associativity

- Distributivity

- Associativity With Scalar Multiplication

for any real (or complex) number *a*.

- Differentiation Rule

### Example 1Edit

Find the convolution, *z(t)*, of the following two signals, *x(t)* and *y(t)*, by using (a) the integral representation of the convolution equation and (b) muliplication in the Laplace domain.

The signal *y(t)* is simply the Heaviside step, *u(t)*.

The signal *x(t)* is given by the following infinite sinusoid, *x _{0}(t)*, and windowing function,

*x*:

_{w}(t)Thus, the convolution we wish to perform is therefore:

From the distributive law:

## CorrelationEdit

Akin to Convolution is a technique called "Correlation" that combines two functions in the time domain into a single resultant function in the time domain. Correlation is not as important to our study as convolution is, but it has a number of properties that will be useful nonetheless.

The correlation of two functions, *g(t)* and *h(t)* is defined as such:

Where the capital *R* is the **Correlation Operator**, and the subscripts to *R* are the arguments to the correlation operation.

We notice immediately that correlation is similar to convolution, except that we don't time-invert the second argument before we shift and integrate. Because of this, we can define correlation in terms of convolution, as such:

### Uses of CorrelationEdit

Correlation is used in many places because it demonstrates one important fact: Correlation determines how much similarity there is between the two argument functions. The more the area under the correlation curve, the more is the similarity between the two signals.

### AutocorrelationEdit

The term "autocorrelation" is the name of the operation when a function is correlated with itself. The autocorrelation is denoted when both of the subscripts to the Correlation operator are the same:

While it might seem ridiculous to correlate a function with itself, there are a number of uses for autocorrelation that will be discussed later. Autocorrelation satisfies several important properties:

- The maximum value of the autocorrelation always occurs at
*t = 0*. The function always decreases, stays constant, or fluctuates (if the signal is periodic) as*t*approaches infinity. - Autocorrelation is symmetric about the
*x*axis.

### CrosscorrelationEdit

Cross correlation is every instance of correlation that is not considered "autocorrelation". In general, crosscorrelation occurs when the function arguments to the correlation are not equal. Crosscorrelation is used to find the similarity between two signals.

### Example: RADAREdit

RADAR (RAdio Detection And Ranging) is a system that uses pulses of electromagnetic waves to determine the position of a distant object. RADAR operates by sending out a signal, and then listening for echos. If there is an object in range, the signal will bounce off that object and return to the RADAR station. The RADAR will then take the cross correlation of two signals, the sent signal and the received signal. A spike in the cross correlation signal indicates that an object is present, and the location of the spike indicates how much time has passed (and therefore how far away the object is).