Документ взят из кэша поисковой машины. Адрес оригинального документа : http://www.adass.org/adass/proceedings/adass98/cheselkam/
Дата изменения: Sat Jul 17 00:46:41 1999
Дата индексирования: Tue Oct 2 07:11:19 2012
Кодировка:

Поисковые слова: mdi
Automatic Detection of Linear Features in Astronomical Images Next: Modern CCD Observations of Moving Celestial Objects: Algorithms and Software for Interactive Processing
Up: Data Analysis and Processing Techniques
Previous: Data Structure and Software of the UCAC-S Project
Table of Contents - Subject Index - Author Index - Search - PS reprint - PDF reprint

Cheselka, M. 1999, in ASP Conf. Ser., Vol. 172, Astronomical Data Analysis Software and Systems VIII, eds. D. M. Mehringer, R. L. Plante, & D. A. Roberts (San Francisco: ASP), 349

Automatic Detection of Linear Features in Astronomical Images

Matthew Cheselka
IRAF Group, National Optical Astronomy Observatories, Tucson AZ 85726

Abstract:

A new IRAF task has been developed that automatically identifies linear (line-like) features in an image or set of images. Such features could include moving targets taken over a long exposure (asteroids, meteors, satellites, aircraft), bad rows and columns of CCD arrays, featured caused by CCD ``bleeding'', or diffraction spikes. The linear features can have gaps and still be recognized as a single feature. Identifying a linear feature in the input image requires several steps. First, a list of pixels within a range of data values is found. Second, pixels within this list are examined to determine if they are colinear. Lastly, the task examines the colinear list of pixels and finds pixels which are adjacent within specified parameters. The colinearity detection is done via the Hough Transform (Gonzalez & Wintz 1987). Once features have been identified, information about the features is written to an output file and profile plots are generated. An optional binary mask can also be created in the case where bad CCD rows or columns are being identified, or where particular linear features need to be masked out.

1. Introduction

Computer programs that are able to automatically identify features in images have been shown to be very useful compared with manually attempting the same thing. This is especially true when there are large numbers of images to examine. In addition, identification of linear features in images for the purposes of analysis or masking them out has many applications.

In order to create a function that will do this, we need to understand exactly what a linear feature is. First, a ``feature'' is an object in an image that has a high enough S/N ratio to be detected. Second, the intensity distribution is such that pixels which make up the feature are located adjacent to one another or have some other kind of distinguishable pattern. A feature is ``linear'' if these pixels all fall on a line of some width and length.

With these definitions, the Hough Transform is adequate to search for colinear lists of pixel locations and a simple adjacency algorithm is suitable to check for the proper position distribution.

2. Algorithm Description

In order to automatically detect linear features, the first step is to select pixels in the image (Fig. 1, left panel) that fall within a given data value range. The input image is assumed to already be dark subtracted and flattened. The user has the option of selecting one or more ranges if it is determined, for example, that some features will be bright and some will be faint. Each data value range can either be computed from the minimum and maximum data values, in terms of evenly spaced $\sigma$ ranges above the mean, or in terms of evenly spaced data value ranges. The user must specify the number and the type of ranges desired.

Figure 1: Input image with linear feature (left); resulting image after thresholding (right).
\begin{figure}
\plottwo{cheselkam1.eps}{cheselkam2.eps}
\vskip 0.2cm
\par\end{figure}

Figure 2: Parameter space ``image'' of pixel list (left); the detected colinear feature (right).
\begin{figure}
\plottwo{cheselkam3.eps}{cheselkam4.eps}
\end{figure}

For each data range, a list of pixel locations (x,y coordinates) is created and fed into the Hough transform (Fig. 1, right panel). The Hough transform converts these pixel locations into a new parameter space given by the equation:


\begin{displaymath}
\rho = (x \cos \theta) + (y \sin \theta)
\end{displaymath} (1)

Each pixel location (an x,y pair) is therefore represented as a curve (Fig. 2, left panel) in this new parameter space corresponding to the above equation. In this parameter space, the x-axis is $\theta$ and the y-axis is $\rho$.

Pixels in image space that are aligned with one another will have curves that intersect a one particular point in parameter space. The task creates an ``image'' in parameter space by accumulating these curves and stacking one on top of another. Naturally, points of intersection will have larger values than other points.

Once a parameter space image is created, a simple peak finder will locate the highest peak. The highest value in parameter space corresponds to the maximum number of selected colinear points in image space. The value at that peak corresponds to the number of pixels in image space that are aligned. The location of the peak in parameter space is used to ``backtrack'' to find the image space pixels that correspond to that peak. Typically, a parameter space image will have several peaks. Each peak corresponds to other locations in image space where pixels have lined up. This fact can be exploited to search for features that are of any user-specified length.

Another aspect of extended linear features is that the image space pixels that describe them are in some way adjacent to one another and have a minimum and maximum length. Undoubtedly, there will be pixels that do not correspond to the feature but are still part of a colinear set. Once such a set is found, the list of pixel locations is sent to an adjacency calculator to determine if a suitable number of them are next to one another. This is where the next set of user-specified parameters comes into play. First, the distance between adjacent pixels must be smaller than a given separation. Second, the length of the feature must fall between a minimum and maximum number of pixels. If all conditions are met, a feature has been positively identified (Fig. 2, right panel). Figures 3 and 4 show other examples of this method.

There was a desire for this task to go beyond just identification by actually doing some simple analysis of the features themselves. First, information about the feature (beginning and ending pixel location, length, position angle, mean intensity and standard deviation) is written to an output file. Second, the user can also optionally create profile plots of each feature and either have them appear inside their xgterm window or saved as IRAF graphics metacode files to be examined at a later time. Third, and probably most useful, is that a binary mask (a pixel list file) is created that can then be used by other IRAF tasks as a region identifier for interpolation or other masking purposes.

Figure 3: Asteroid streak (left); detected linear feature (right).
\begin{figure}
\plottwo{cheselkam5.eps}{cheselkam6.eps}
\vskip 0.5cm
\par\end{figure}

Figure 4: NOAO Mosaic image showing bad columns (left); detected linear features (right).
\begin{figure}
\plottwo{cheselkam7.eps}{cheselkam8.eps}
\end{figure}

References

Gonzalez, R. C. & Wintz, P. 1987, Digital Image Processing (2nd ed.), (Reading: Addison-Wesley), 130


© Copyright 1999 Astronomical Society of the Pacific, 390 Ashton Avenue, San Francisco, California 94112, USA
Next: Modern CCD Observations of Moving Celestial Objects: Algorithms and Software for Interactive Processing
Up: Data Analysis and Processing Techniques
Previous: Data Structure and Software of the UCAC-S Project
Table of Contents - Subject Index - Author Index - Search - PS reprint - PDF reprint

adass@ncsa.uiuc.edu