Linear prediction filter coefficients
finds the coefficients of a
g] = lpc(
pth-order linear predictor, an FIR filter
that predicts the current value of the real-valued time series
on past samples. The function also returns
g, the variance of the
prediction error. If
x is a matrix, the function treats each column as
an independent channel.
Estimate Series Using Forward Predictor
Estimate a data series using a third-order forward predictor. Compare the estimate to the original signal.
First, create the signal data as the output of an autoregressive (AR) process driven by normalized white Gaussian noise. Use the last 4096 samples of the AR process output to avoid startup transients.
noise = randn(50000,1); x = filter(1,[1 1/2 1/3 1/4],noise); x = x(end-4096+1:end);
Compute the predictor coefficients and the estimated signal.
a = lpc(x,3); est_x = filter([0 -a(2:end)],1,x);
Compare the predicted signal to the original signal by plotting the last 100 samples of each.
plot(1:100,x(end-100+1:end),1:100,est_x(end-100+1:end),'--') grid xlabel('Sample Number') ylabel('Amplitude') legend('Original signal','LPC estimate')
Compute the prediction error and the autocorrelation sequence of the prediction error. Plot the autocorrelation. The prediction error is approximately white Gaussian noise, as expected for a third-order AR input process.
e = x-est_x; [acs,lags] = xcorr(e,'coeff'); plot(lags,acs) grid xlabel('Lags') ylabel('Normalized Autocorrelation') ylim([-0.1 1.1])
x — Input array
vector | matrix
Input array, specified as a vector or matrix. If
x is a matrix,
then the function treats each column as an independent channel.
p — Prediction filter polynomial order
length( (default) | positive integer
Prediction filter polynomial order, specified as a positive integer.
p must be less than or equal to the length of
The prediction error, e(n), can be viewed as the output of the prediction error filter A(z), where
H(z) is the optimal linear predictor.
x(n) is the input signal.
is the predicted signal.
lpc determines the coefficients of a forward linear predictor by
minimizing the prediction error in the least squares sense. It has applications in filter
design and speech coding.
lpc uses the autocorrelation method of autoregressive (AR) modeling
to find the filter coefficients. The generated filter might not model the process exactly,
even if the data sequence is truly an AR process of the correct order, because the
autocorrelation method implicitly windows the data. In other words, the method assumes that
signal samples beyond the length of
x are 0.
lpc computes the least-squares solution to Xa = b, where
and m is the length of x. Solving the least-squares problem using the normal equations leads to the Yule-Walker equations
[r(1) r(2) ... r(p+1)
is an autocorrelation estimate for
x computed using
Levinson-Durbin algorithm (see
levinson) solves the Yule-Walker equations in O(p2) flops.
 Jackson, L. B. Digital Filters and Signal Processing. 2nd Edition. Boston: Kluwer Academic Publishers, 1989, pp. 255–257.
Introduced before R2006a