automated spectral classification using neural networks · cefet-pr, 13 a 15 de setembro de 1995...
TRANSCRIPT
22 SIMPÓSIO BRASILEIRO DE AUTOMAÇÃO INTELIGENTE CEFET-PR, 13 a 15 de Setembro de 1995 Curitiba Paraná
Automated Spectral Classification Using Neural Networks E. F. Vieira1 and J. D. Ponz2
1 LAEFF 2 ECNOD/ESA Villafranca, P. O. Box 50727, 28080 Madrid, Spain [email protected], [email protected]
Abstract
Along the life of the International Ultraviolet Explorer (IUE) astronomical satellite project, a large archive with spectral data has been generated, requiring automated classification methods to be analyzed in an objective formo Previous automated classification methods used with IUE spectra were based on multivariate statistics. In this paper, we compare three classification methods that can be directly applied to spectra in the archive:
- metric distance, - supervised classification using Artificial Neural Networks (ANN) and - unsupervised classification using Self Organized Maps (SOM).
These methods are used to classify IUE low-dispersion spectra of normal stars with spectral types ranging from 03 to G5. The classification based on supervised artificial neural networks performs better than the metric distance and the current SOM classification, allowing the determination of the spectral classes with an accuracy of 1.1 spectral subclasses. The unsupervised classification can be also use to test the supervised technique.
1. Introduction
This paper explores the application of automated methods to the problem of classification of stellar spectra.
The availability of large spectral archives and the efficiency achieved with modern instrumentation require automated classification methods to improve the classification by visual inspection. These methods are still in exploratory phase; the aims are clear, to devise an objective, repeatable and robust classification scheme providing an estimation of systematic and random errors, and allowing the quantification of spectral resolution and signal to noise ratio on classification errors.
Traditionally, spectra have been classified manually according to star temperatures and luminosity classes. In this preliminary work we consider temperature as parameter to be classified. The spectral or temperature classes are named O, B, A, F, G, K and M where O is for the hottest stars and M is for the coldest ones. Further subdivisions of classes (e.g. AO, AI, A2, etc.) is
Send offprint requests to: J. D. Ponz
based on detailed spectral features (Allen 1976). The estimated error in the classification made by human experts is about 0.5 spectral sub classes (Weaver 1994) (von Hippel, Storrie-Lombardi, & Storrie-Lombardi 1994) .
Automated classifiers of stellar spectra can be divided into metric distance algorithms, multivariate statistics and artificial neural networks (hereafter ANN) . Metric distance methods were originally proposed by K urtz and LaSala (K urtz 1982) (Kurtz 1984) (Lasala & Kurtz 1985) and have been used by Penprase (Penprase 1994) to classify stellar spectra using the digital spectraI atlas of Jacoby et aI. (Jacobi, Hunter, & Christian 1984) as template. Multivariate statistical methods are .linear algorithms used for exploratory data analysis. These methods have been applied to spectral classification by using Principal Component Analysis (PCA) to reduce the dimension of the problem, followed by Cluster Analysis (CA) to discover groups of objects in the parameter space obtained in the previ- . ous step (Murtagh & Heck (Murtagh & Heck
. 1984) and references herein). Stellar classifica-
50
tion with ANN is a new approach that has been used by von Hippel et a!. (von Hippel, StorrieLombardi, & Storrie-Lombardi 1994) (von Hippel et aI. 1994) to confirm the visual classification of the Miehigan Spectral Catalogue on objective prism spectra, determining the temperature classification to better than 1.7 spectral sub classes from B3 to M4. Gulati et a!. (Gulati et ai. 1994) classify the spectral atlas of Jacoby et a!. (Jacobi, Hunter, & Christian 1984) with an accuracy of 2 spectral subclasses, based on selected spectral features.
The availability of the IUE low-dispersion archive (Wamsteker, Driessen, & Munoz 1989) allows the application of pattern recognition methods to explore the ultraviolet domain. The analysis of this archive is especially interesting, due to the homogeneity of the sample. As indicated by Heck (Heck et ai. 1983), it is important to remember at this point that MK spectraI classifications defined from the visible range cannot simply be extrapolated to the ultraviolet spectral range. So far, only multivariate statistieal methods have been used with IUE spectra. Egret and Heck (Egret & Heck 1983) analyze the relative fluxes at 16 selected wavelengths of O and B stars with PCA. Egret et al. (Egret et aI. 1984) analyze low-resolution spectra using PCA on 93 variables computed as median flux values at certain wavelength bands and selected absorption and emission lines. These analyses indieate a high correlation betweeil the first principal component and the temperature. Heck · et a!. (Heck et aI. 1986) classify the IUE LowDispersion Spectra Reference Atlas (Heck et aI. 1983) of normal stars. Weighted intensities of sixty lines together with an asymmetry coefficient describing the continuum shape are used for the classification. The algorithm consisted of PCA folIowed by CA to define different groups that confirmed the manual classification of the Atlas. Imadache and Crézé (Imadache & Crézé 1990) and Imadache (Imadache 1992) extend the sample and generalize the method, using the fulI spectral range instead of pre-selected spectral features.
The present work has been done within the context of the IUE Final Archive project. The aim is to provide an efficient and objective classification procedure to explore the complete IUE database, based on methods that do not require prior knowledge about the object to be classi-
2' SIMPÓSIO BRASILEIRO DE AUTOMAÇAO INTELIGENTE ~
.~
fied. Three methods are compared: a simple metrie distance method, a supervised ANN classifier and a unsupervised Self Organized Map (SOM) classifier. The input sample is described in section 2. The classification method based on metric distance is summarized in section 3. Section 4 explains the supervised classification and section 5 describes the unsupervised method. The results are presented in section 6 and section 7 summarizes the conclusions.
2. The Sample Spectra
The spectra were taken from the IUE LowDispersion Reference Atlas of Normal Stars (Heck et aI. 1983) (hereafter the Atlas), covering the wavelength range from 1150 to 3200 Á. The Atlas contains 229 normal stars distributed from the spectral type 03 to KO. The classification given in the Atlas was carried out following a classieal morphologieal approach (J aschek & Jaschek 1984), based on UV criteria alone. The set of 64 standard stars selected in the Atlas, with spectral types from 03 to G5, was used as a template in the metrie distance classification and was the training sample in ANN classification. The test set contained 163 spectra, excluding the 64 standard stars and two stars with spectral types G8 and KO, outside the spectral types covered by the training set.
The spectra were obtained by merging together data from the two IUE cameras, sampled at a uniform wavelength step of 2 Á, after processing with the standard calibration pipeline. Although the spectra are good in quality, there are two aspects that seriously hinder the automated classification: interstelIar extinction and contamination with geo-coronal Ly-a emission. Some pre-processing was required to eliminate these effects and to normalize the data.
AlI spectra were corrected for interstelIar extinction by using Seaton's (Seaton 1979) extinction law. Due to the properties of the extinction law at À1 = 1600, À2 = 2400 and Àc = 2175 Á thecolor excess E(B - V) was estimated as
The observed fluxes fI, f~ and fg were obtained by filtering the high frequency components in the transformed Fourier space (Lasala &
.'7.
21 SIMPÓSIO BRASILEIRO DE AUTOMAÇAO INTELIGENTE
K urtz 1985). Figure 1 shows a typical 04 spectrum before and after correction for reddening.
Selected range
After de-reddening
1200 1600 2000 2400 2800 3200 Wavelength
Fig. 1. 'Original (top) and de-reddened (bottom) spectra corresponding to a 04 star. The selected range is indicated by the solid lines in the middle
3. Classification Using Metric Distance
Normalized spectra are considered vectors in]RN and a metric is introduced in the vector space. In this classification scheme the metric distance between the object spectrum and each spectrum in the training set is computed and the spectral class of the star in the training set having the minimum distance is assigned to the object.
Let fij = fi(Àj)j j = 1, ... , N be the flux of the i-th star in the catalogue and Skj = Sk(Àj)j j = 1, . .. , N be the flux of the k-th standard star in the training set, after correction for reddening and normalization. The distance, dik,
is defined by
N 2 1 ~ 2
dik = N L.J(!ij - Skj) . j=l
(2)
4. Supervised Classification Using ANN
A supervised classification scheme based on artificial neural networks (ANN) has been used. This techl1ique was originally developed by McCul-10gh and Pitts (McCullogh & Pitts 1943) and has been generalized with an algorithm for training
51
Flux of information
Output Layer
Input Layer
Fig. 2. A typical ANN topology
networks having multiple layers, known as backpropagation (Rumelhart, Hinton, & Williams 1986). In a general form, a neural network represents a function, F, that maps a given input set into a selected output set. Assuming that normalized spectra are elements in a N -dimensional vector space and spectral classes are defined by Mdimensional classification vectors, the network approximates the mapping F : mN --* ]R M ,
in such a way that a standard star, S k, is associated with the vector Ck = F(Sk)' The output vector defin,es the spectral class so that 00 is given by (1, O, O, . .. ,O), 01 is represented by (0,1,0, . .. ,O) and so on. In this form, the network can be regarded as a classifier that maps the input space of normalized fluxes into Iof M, Le., one output unity and all others zero. Such a network, with a squared-error cost function, gives a good estimation of Bayesian probabilities (Richard & Lippmann 1991), so that ·for an input spectrum J, the i-th component of the output vector is the probability P(CiIJ) for class i given the input spectrum. In the Figure 2 is represented a typical ANN topology.
5. Unsupervised Classification Using SOM
In the Self Organized Map (SOM) the net organizes the spectra into clusters based on similarities using a metric to define the distance
52
between two spectra. The algorithm used to perform such clustering was developed by Kohonen (Kohonen 1984).
A schematic Self Organized Map (SOM) is shown in Figure 3. In this representation the input layer is shown with 3 neurons and a bidimensional net with 9 neurons. The output values are represented by the vertical arrows. Each input neuron is full connected to the bidimensional neto
Output
Input Layer
Bi-Dimensional Net
Fig. 3. Representation of a bi-dimensional SOM.
6. RESULTS
6.1. ANN
The resuIts obtained for different architectures compared with the Atlas are shown in Table 1, where r is the correlation coefficient and u is the standard deviation. The distribution of discrepandes between the manual classification and the automated methods are shown in Figure 5 for ANN and Metric Distance. In the Figure 5 is shown the results of the classification for the first configuration in the Table 1. The configuration having 120 nodes and 2 hidden layers was used to compare ANN and metric distance. There is a good agreement between the this two classification methods.
60
40
20
21 SIMPÓSIO BRASILEIRO DE AUTOMAÇAO INTELIGENTE
_ANN D Melric Distance
o~~~~~~~~~~~~~
·5 -4 -3 -2 -1 o 234 5 Error
~21~1
Fig. 4. Distribution for classification discrepancies for ANN and Metric Distance versus Atlas
G r-----------------------------~
GlF ~ üi õA .g ãí ~B
B A Atlas
F G
G r---------------------------~
F
z ~A
B
B A Atlas
F G
G .---------------------------~
F
z ~A
B
B A F G Metric Distance
Fig. 5. Results of classification: Metric Distance versus Atlas (top), ANN versus Atlas (middle) and ANN versus Metric Distance (bottom). The solid lines are drawn just for reference
2' SIMPÓSIO BRASILEIRO DE AUTOMAÇAO INTELIGENTE
Hidden Hidden Nodes Layers r
120 2 0.988 60 2 0.983 30 2 0.946 120 1 0.984 60 1 0.986 30 1 0.982
Table 1. Network Configurations
6.2. SOM
u
1.107 1.350 2.379 1.313 1.221 1.378
In this work it was used a 8x8 bi-dimensional net with 744 neurons in the input layer. The main set of spectra was exactly the same used in the supervised classification. The classifier gives an error of 1.62 sub classes if compared with the Atlas, with a correlation of 0.9844. In addition, 27 stars could not be classified according to the classification criterium used in this experimento The adopted criterium uses standard stars as labels. If a neuron in the bi-dimensional plane is not associated with some of the 64 standard stars, none of the stars assigned to this neuron by the net will are classified.
K
G
~F o (J)A
B
o o B A K Alia
K
G
~F o (J)A
B
o o B A ANN
Fig. 6. Results of classification: SOM versus Atlas (top), SOM versus' ANN (bottom). The solid lines are drawn just for reference
53
7. CONCLUSIONS
- The automated classification can be applied to large databases and the errors are compatible with the ones found in manual classification methods. The error found for supervised algorithm is of 1.10 subclasses and of 1.62 subclasses for the unsupervised method.
- The main advantage of the two methods applied in this work is that both methods use the spectra as observed without previous analysis of spectral features.
- In addition, the unsupervised classification doesn 't need a priori know ledge to perform the classification.
- Furthermore, no personal criteria are used and that the classification can be reproduced, which in general is not possibIe in classification made by human experts.
- The methods are robust enough to be used in case of spectra with missing information and further improvements can be obtained with IUE spectra rejecting bad pixels as indicated in the quality flags associated with the spectraI values.
- This research will be continued to derive physical parameters by using stellar models in connection with observed spectra, and taking into account Iuminosity classes.
References
Allen, C.W. 1976, Astrophysical Quantities(third ed.). (The Athlone Press, University of London) . Page 198
Egret, D., & Heck, A. 1983, ESA SP,SP(202), 59 Egret, D., Heck, A., Nobelis, P.H., .& Thrlot, J.C.
1984, in Future of Ultraviolet Astronomy Based on Six Years of IUE Research, p. 512. NASA CP-2349
Gulati, R.K., Gupta, R., Gothoskar, P., & Khobragade, S. 1994, ApJ, 426, 340
Heck, A., Egret, D., Jaschek, M., & Jaschek, C. 1983, Feb). IUE Low Dispersion Spectra Reference Atlas. SP 1052, ESA. PartI. Normal Stars
Heck, A., Egret, D., Nobelis, Ph., & Thrlot, J.C. 1986, Astrophysics and Space Science, 120, 223-237
Imadache, A. 1992. Techniques de Classification Automatique de Spectres Stellaires Ultraviolets. Ph. D. thesis, Observatoire Astronomique de Strasbourg, Strasbourg
Imadache, A., & Crézé, M. 1990. Evolution in astrophysics. Technical Report SP-310, ESA
54
Jacobi, G.H., Hunter, D.A., & Christian, C.A. 1984, ApJS, 56, 257
Jaschek, M., & Jaschek, C. 1984, in The MK proces and Stellar Classification, p. 290. David Dunlap Observatory
Kohonen, Teuvo 1984, Self-organization and Associative Memory Volume 8 of Springer Series in Information Sciences (Springer Verlag, Nueva York)
Kurtz, M.J. 1982. Automatic spectral classification. Ph. D. thesis, Dartmouth College, New Hamphsire
Kurtz, M.J 1984, in The MK Process and Stellar Classification, Volume 60, p. 136. ASP Conference Series
Lasala, J., & Kurtz, M.J. 1985, July) , PASP,(97), 605-608
McCullogh, W.S., & Pitts, W.H. 1943, Bull. Math. Biophysics, 5, 115
Murtagh, F., & Heck, A. 1984, Multivariate Data Analysis (Reindel, Dordrecht)
Penprase, B.E. 1994, in The MK Process at 50 Years, Volume 60, p. 325. ASP Conference Series
Richard, M.D., & Lippmann, R.P. 1991, NeuraÍ Computing, 3, 461
Rumelhart, D.E., Hinton, G.E., & Williams, R.J. 1986, October), Nature, 323, 533-536
Seaton, M.J. 1979, MNRAS,(187), 73P-76P. Short Communication
Hippel, von T., Storrie-Lombardi, L.J., & StorrieLombardi, M.C. 1994, in The MK Process at 50 Years, Volume 60, p. 289-302. ASP Conference Series
Hippel, von T., Storrie-Lombardi, L.J., StorrieLombardi, M.C., & lrwin, M.J. 1994, MNRAS, 269,97
Wamsteker, W., Driessen, C., & Muiíoz, J.R. et alo 1989, A&AS, 79, 1
Weaver, Wm.B. 1994, in The MK Process at 50 Years, Volume 60, p. 303-311. ASP Conference Series
~21~
2! SIMPÓSIO BRASILEIRO DE ~ AUTOMAÇAO INTELIGENTE