Chebyshev/Orthogonal Polynomial Model

This section provides details on the Chebyshev/Orthogonal Polynomial model approximation technique used in Isight.

Related Topics
Approximation References

Overview

Orthogonal polynomial approximation is a type of regression technique. Orthogonal polynomials minimize the autocorrelation between the response values that exist because of the sampling locations. An advantage of using orthogonal functions as a basis for fitting is that the inputs can be decoupled in the analysis of variance (ANOVA) (Nakajima, 2006).

Chebyshev orthogonal polynomials are a common type of orthogonal polynomials that are particularly useful for equally spaced sample points. They are used when the sampling strategy is an orthogonal array. Isight implements Taguchi’s method (Taguchi, 1987) of fitting Chebyshev polynomials from an orthogonal array.

Isight provides the capability to compute orthogonal polynomial approximations for other kinds of samplings. In such cases the following approximation models are available:

Chebyshev Polynomials

The model is constructed as a linear regression of standard Chebyshev polynomials.

Successive Orthogonal Polynomials

A model consisting of a set of successive orthogonal polynomials that are orthogonal on the sample points. The successive orthogonal polynomial technique generates a series of polynomials that are orthogonal with respect to the data provided. These polynomials are used as basis functions to obtain an approximation for the responses. These orthogonal basis-functions depend only on the sample locations not on the response values.

Chebyshev Polynomial Approximation

Chebyshev polynomials are a set of orthogonal polynomials that are solutions of a special kind of Sturm-Liouville differential equation called a Chebyshev differential equation.

The equation is

(1x2)yxy+n2y=0.

Chebyshev polynomials can be of two kinds. In one dimension these polynomials are defined as follows:

Polynomials of the first kind

T0(x)=1T1(x)=xTn+1(x)=2xTn(x)Tn1(x).

Polynomials of the second kind

U0(x)=1U1(x)=2xUn+1(x)=2xUn(x)Un1(x)

The roots of these polynomials are not equally spaced. Taguchi describes a set of one-dimensional polynomials, which he calls Chebyshev, that have equally spaced roots. When these equally spaced roots are assumed to be the factor levels in an orthogonal array, a quadrature procedure is available for approximating a response using Chebyshev polynomials as individual terms.

In general, the quadrature method of fitting an approximation is more efficient and stable compared to a regression-based approach. However, the quadrature approach dictates that the function being approximated be evaluated at pre-defined locations. For Chebyshev polynomials, these positions correspond exactly to a sample obtained using an orthogonal array.

The following equations show the Chebyshev polynomials with equally spaced roots in one dimension:

T1(x)=(xx)¯T2(x)=(xx¯)2b2T3(x)=(xx¯)3b31(xx¯)T4(x)=(xx¯)4b41(xx¯)2T5(x)=(xx¯)5b51(xx¯)3b52(xx¯)

where x is the average value of the levels. Taguchi generates multivariate polynomials by taking products of Chebyshev polynomials in each variable as listed above. Taguchi also provides tables for computing the coefficients of these terms for an orthogonal array.

For example, suppose we have three variables x1, x2, and x3 to which we want to fit a response g. We can generate the following multivariate polynomial basis:

Linear:{T1(x1);T1(x2);T1(x3)}(3 pure terms)Quadratic:{T2(x1);T2(x2);T2(x3);(3 pure terms)(T1(x1)T1(x2));(T1(x1)T1(x3));(T1(x2)T1(x3))}(3 cross terms)Cubic:{T3(x1);T3(x2);T3(x3);(3 pure terms)(T2(x1)T1(x2));(T1(x1)T2(x2));(T2(x2)T1(x3));(T1(x2)T2(x3));(T2(x2)T1(x3));(T1(x2)T2(x3));(7 cross terms)(T1(x1)T1(x2)T1(x3))}

Therefore, the function g is approximated as

g(x1,x2,x2)g^=a11T1(x1)++a13T1(x3)+a21T2(x1)++a24(T1(x1)T1(x2))++a26(T1(x2)T1(x3))+a31T3(x1)++a34(T2(x1)T1(x2))++a2,10(T1(x1)T1(x2)T1(x3)).

Isight uses Taguchi’s tables to calculate the coefficients aij for orthogonal array sampling and least squares regression to calculate all other sampling techniques. A term-by-term ANOVA can also be computed for a Chebyshev polynomial approximation when orthogonal array sampling is used.

Note: All points in an orthogonal array have to be available to use Taguchi’s method. Therefore, the error value obtained using the cross-validation approach is no longer meaningful for orthogonal arrays. When cross-validation is used, the sample set for fitting the approximation no longer contains all the points. Therefore, Isight uses the regression approach to build the approximations for cross validation, whereas Isight uses Taguchi’s approach for models with all the points.

Successive Orthogonal Polynomial Approximation

Isight provides the capability to use an arbitrary set of orthogonal polynomials to construct an approximation. Taguchi describes a multi-variable method for only the linear case, but the approach has been extended for an arbitrary degree in Isight. For example, suppose we have three variables x1, x2, and x3 to which we want to fit a response g. To simplify the expressions, assume that x1, x2, and x3 have a mean of zero.

We can generate the following sequence of orthogonal functions (sometimes called a contrast):

linear:f1=x1f2=x2b2,1f1,LT(f2)=x2f3=x3b3,1f1b3,2f2,LT(f3)=x3quadratic:f4=x12b4,1f1b4,2f2b4,3f3,LT(f4)=x12f5=x22b5,1f1b5,2f2b5,3f3b5,4f4,LT(f5)=x22f6=x32b6,1f1b6,2f2b6,3f3b6,4f4b6,5f5,LT(f6)=x32f7=x1x2i=16b7,ifi,LT(f7)=x1x2f8=x1x3i=17b8,ifi,LT(f8)=x1x3f9=x2x3i=18b9,ifi,LT(f9)=x2x3

Here LT indicates the Leading Term of that polynomial. If x1, x2, and x3 are permuted, the general form of the polynomial will be different. The coefficients in the orthogonal polynomial {bi,j;i=1n,j=1(i1)} can be computed using the discrete orthogonality condition as follows ( X ¯ k is the kth vector in the input data):

k=1Nfi(X¯k)fj(X¯k)=0k=1Nfj(X¯k)LTi(X¯k)bi,jk=1Nfj2(X¯k)=0bi,j=k=1Nfj(X¯k)LTi(X¯k)k=1Nfj2(X¯k)

After solving for these values, we can obtain the coefficients {ai:i=1n} of the fit

g(x1,x2,x3)i=1naifi

using the following equation:

ai=k=1Nfi(X¯k)g(X¯k)k=1Nfi2(X¯k).

The time taken to generate these successive orthogonal polynomials increases drastically with an increase in the number of input variables of the approximation. Therefore, this approximation is best suited for models with a small number of input variables and a large number of data points (or responses).