**American Journal of Mathematical Analysis**

## Linear Model Analysis of Observational Data in the Sense of Least –Squares Criterion

Department of Statistics College of Science for Girls, King Abdulaziz University, Jeddah, Saudi ArabiaAbstract | |

1. | Introduction |

2. | Observational Data |

3. | Least-squares Approximation |

4. | Mathematica Simulation |

References |

### Abstract

The present paper is devoted for the following goals: To develop an algorithm for model analysis of observational data in the sense of the least –squares criterion with full error analysis.. By this algorithm one computes, all the solutions with their variances, the variance of the fit, the average square distance between the least square solution and the exact solution, and graphical representation between the row and the fitted data. Mathematica module of the algorithm was established, through five points, its purpose –input - output –needed procedures and the list of the module. By this paper we have been tried to produce an error controlled algorithm of the least squares method for observational data.

**Keywords:** statistical data analysis -least squares method, mathematica simulations

Received May 05, 2015; Revised June 05, 2015; Accepted June 24, 2015

**Copyright**© 2015 Science and Education Publishing. All Rights Reserved.

### Cite this article:

- J.A. Basabrain. Linear Model Analysis of Observational Data in the Sense of Least –Squares Criterion.
*American Journal of Mathematical Analysis*. Vol. 3, No. 2, 2015, pp 47-53. http://pubs.sciepub.com/ajma/3/2/5

- Basabrain, J.A.. "Linear Model Analysis of Observational Data in the Sense of Least –Squares Criterion."
*American Journal of Mathematical Analysis*3.2 (2015): 47-53.

- Basabrain, J. (2015). Linear Model Analysis of Observational Data in the Sense of Least –Squares Criterion.
*American Journal of Mathematical Analysis*,*3*(2), 47-53.

- Basabrain, J.A.. "Linear Model Analysis of Observational Data in the Sense of Least –Squares Criterion."
*American Journal of Mathematical Analysis*3, no. 2 (2015): 47-53.

Import into BibTeX | Import into EndNote | Import into RefMan | Import into RefWorks |

### 1. Introduction

One of the hallmarks of modern astrophysics is the extreme rise in the amount of data available and the prevalence of images and other data in electronic form particularly that available to the public via the Internet. The present situation is in marked contrast to the view given by Kolb and Turner ^{[1]} more than 25 years ago who lamented that in cosmology at that time there was a paucity of data. There are now so many people involved with data reduction and analysis that in the last decadal survey there was a call for the formation of a new area of astronomy and astrophysics called *astroinformatics* ^{[3]}. Latest information on this rapidly emerging discipline can be found with listings on the Internet through a Google or similar search under that name.

Astronomers and astrophysicists have always shared data and it is even truer today than ever before. The electronic data revolution for astrophysics and astronomy took real hold in the pre-Hubble spacecraft days with the establishment of the NASA/ADC (Astronomical Data Center) when catalogs listing data about a number of stars and other objects were digitized, recorded on CDs, and distributed at national and international meetings for free.

Although the least-squares method is the most powerful techniques that has been devised for the problems of data analysis, it is at the same time exceedingly critical. This is because the least-squares method suffers from the deficiency that, its estimation procedure does not have detecting and controlling techniques for the sensitivity of the solution to the optimization criterion of the variance is minimum. As a result, there may exist a situation in which there are many significantly different solutions that reduce the variance to an acceptable small value. At this stage we should point out that (1) the accuracy of the estimators and the accuracy of the fitted curve are to distinct problems; and (2) an accurate estimator will always produce small variance, but small variance does not guarantee an accurate estimator. This could be seen from property 3 (Section 3) of the least square solution by noting that the lower bounds for the average square distance between the exact and the least-squares values is which may be large even if is very small, depending on the magnitude of the minimum eigenvalue, , of the coefficient matrix of the least-squares normal equations.

Unless detecting and controlling the above mentioned situation, it is not possible to make a well-defined diction about the analysis of the observational data. Due to this difficulty and the importance of least squares for treating the nowadays incredible numbers of the available data, the present paper is devoted for the following goals :To develop an algorithm for model analysis of observational data in the sense of the least –squares criterion with error analysis.. By this algorithm one computes, all the solutions with their variances, the variance o the fit, the average square distance between the least square solution an the exact solution an graphical representation between the row and the fitted data. Mathematica module of the algorithm is also given.

### 2. Observational Data

Let be a sequences of observational data points assumed free from errors. Corresponding to each we have a number of some function at which generally will be in error. Denote , the true value at , by and define:

(1) |

as the observational errors at different data points.

The problem in the analysis of the observational data is to approximate or fit the data by some function in such a way that contains or represents most (if not all) the information about contained in the data and little (if any) of the errors. This is accomplished in practice by selecting a function

(2) |

which depends on the parameters . (The superscript on denotes the fact that they will (generally depend on ). The function may be linear or non-linear in the parameters.

Since our study is the linear model analysis of the observational data, the function will be selected as linear combinations of the parameters so that it takes the form:

(3) |

The may, for example, be the set of monomials, exponentials, trigonometric polynomials, or indeed any arbitrary set of sufficiently defined functional values, provided only that they are *linearly independent* of the m values of x. Normally, n is small compared with the number, m of data points. In fact, the number n is unknown but for the most appropriate approximation to a given data the number n should be :

**1**-large enough so that the information about in the data can well be represented by a proper choice of the parameters , while at the same time n should be,

**2**-too small to avoid the fitting of the observed data too closely in the sense that the errors in the observed data are retained in the approximation.

Now the problem is to find the *best* estimate of the function i.e, to determine particular values for the parameters to obtain the best approximation:

(4) |

Although there is agreement that the best estimate should give small deviations of the given data from the fitted curve- i.e., that the modulus of the quantity :

(5) |

should be small at each of the m points- there are differences of opinion as to how this to be achieved.

Let **ε** denote an error vector composed of the m deviations . Since and are regarded as fixed, the vector **ε** depends only on the parameters . Each common method for dealing with the competing requirements that are small, corresponds to the selection of a *norm* for the vector **ε**.Some examples of norms in the real m- dimensional space are :

(6) |

(7) |

(8) |

The norm is called *Gershgorin norm*, the *Euclidean norm* and the*maximum norm*or, occasionally, the *uniform norm*. For any choice norm,the best estimate is one satisfying the requirement:

(9) |

Assuming one of the above norms, say the Q norm is chosen, then we can determine (by solving certain system of equations) the *estimators* of the best estimate of that satisfies Equation (9), i.e., for the linear model analysis we have the best approximation:

(10) |

to be the true function over in the sense of the Q norm fitting for the given Our best assumption is that for some unknown value of n, say N, the true function can be expressed as a finite linear combination of the selected set of functions in the sense of the Q–norm fitting ;that is, we assume

(11) |

such that the Q norm of

(12) |

for some nN and for a given . Equation (11),with the condition of Equation (12), are the formulation of the linear model analysis of the observational data in the sense of Q-norm fitting.

Now it remains to decide what is the most efficient norm that should be used in the analysis of observational data. This decision constitutes very serious difficultly because:

1- The norm depends greatly on the distributions of the observational errors .

2- There is a large variation in the effectiveness of various norms and no single norm is good (or even mediocre).

3- The distributions of the observational errors cannot be determined with any precision.

It is undoubtedly true that the Euclidean norm is the most efficient when . are normally distributed, and in this case the estimators are the maximum likelihood *estimators*. It is natural therefore, to consider in our study the assumption that.are normally distributed and analyzed the linear –model representation of the observational data in the sense of the Euclidean norm-i.e., in the *least –squares criterion.*

Before starting the analysis, however, we introduce additional conditions concerning the character of the errors and assume that are unbiased (without systematic error),independent and normally distributed with zero mean and variance . These conditions indicate that:

E(z) being the expectations of *z* and the unknown variance.

In the case of weighted observations, Equation (16) is replaced by:

(13) |

where are the weights of the observations (=. Therefore the problem of this section is the application of the least-squares method for the linear model representation of the observational data.

### 3. Least-squares Approximation

Below is a list of vectors and matrices used in the present section:

The transpose of a vector or a matrix will be indicated by the superscript 'T'.

**3.1. Derivation of the Normal Equations**

In matrix notation, the problem of the weighted least-squares approximation is to determine the estimators such that

(14) |

where the vector of the residuals is defined as

(15) |

From Equations (14) and (15) we have ;

that is

(16) |

Since is scalar then,

that is:

(17) |

since

(18) |

then

(19) |

From Equations (16) and (19) we get ;

(20) |

Equation (20) could be written as:

(21) |

(22) |

The necessary conditions for the minimum are

Consequently, from Equation (21) we get

(23) |

therefore

(24) |

where the general elements are given by

(25) |

and

(26) |

Equations (24) are called the *normal equations* of the weighted least-squares approximations for the linear –model representations of the observational data. These equations represent a set of n *linear equations* in n unknowns.The coefficient matrix **G** is symmetric and is positive definite (i.e., all eigen values are positive) if all the fundamental functionsare linearly independent for the arguments ; that is, if the rows of are linearly independent. It should be noted that the normal Equations (24) represent a relation between two exact quantities in the sense that if the matrix **G** and the right-hand side **b **represent the observational data exactly, then there is a vector which satisfies exactly the least –squares criterion. Of course, in practice such an exact situation no longer exists. In the following section, some important properties of (in the sense described above) will be given.

**3.2. Some Properties of**

According to the conditions of Equations (13)and (17) it was proved ^{[1]} the following properties of :

Also it should be noted that:

If H(u) is a function of a measured quantity u, then the standard error then

(27) |

where is the standard error of u

If the precision is measured by the probable error e, then

(28) |

### 4. Mathematica Simulation

**4.1. Mathematica Module: Linear Model**

● **Purpose**

To fit observational data (x,y)(say) to the linear model in the sense of least-squares criterion with its error analysis.

● **Input**

: List of the independet varibles of the data.

: List of the dependet varibles of the data.

: List of the basic functions used for the linear model representation of the data.

Positive numbers each used to disply color drawing according to the built in function .

ind : A charcter indecating the name of the independent variable, (x for example).

dep: A charcter indecating the name of the dependent variable, (y for example).

● **Output**

The c’s coefficients and their probable errors.

The probable error of the fit.

The average square distance between the exact and the least squares solutions.

Graphical representations

● **Needed ****procedures**

None

● **List of**** ****the Module**

We applied this module for many problems for galactic structure, the results are very satisfactory and will be published soon in astronomical journals

In concluding the present paper, an algorithm for model analysis of observational data was deloped in the sense of the least –squares criterion with full error analysis.. By this algorithm one computes, all the solutions with their variances, the variance of the fit, the average square distance between the least square solution and the exact solution, and graphical representation between the row and the fitted data. Mathematica module of the algorithm was established, through five points, its purpose –input - output - needed procedures and the list of the module. By this paper we have been tried to produce an error controlled algorithm of the least squares method for observational data.

### References

[1] | Kolb, E.W., and Turner, M.S. 1990, The Early Universe. Frontiers in Physics. Redwood City, CA: Addison –Wesley. Advanced Book Program. | ||

In article | |||

[2] | Kopal, Z and Sharaf, M.A. 1980, Linear Analysis of the Light Curves of Eclipsing Variables, Astrophysics and Space Science, 70, 77-101. | ||

In article | View Article | ||

[3] | Borne, J, M., and 91 others: 2009, Astroinformatics: A 21^{st} Century Approach to Astronomy. Tech. rept. Submitted to 2010 Decadal Survey. | ||

In article | |||