Proceedings Volume 2492

Applications and Science of Artificial Neural Networks

Steven K. Rogers, Dennis W. Ruck
cover
Proceedings Volume 2492

Applications and Science of Artificial Neural Networks

Steven K. Rogers, Dennis W. Ruck
View the digital version of this volume at SPIE Digital Libarary.

Volume Details

Date Published: 6 April 1995
Contents: 27 Sessions, 116 Papers, 0 Presentations
Conference: SPIE's 1995 Symposium on OE/Aerospace Sensing and Dual Use Photonics 1995
Volume Number: 2492

Table of Contents

icon_mobile_dropdown

Table of Contents

All links to SPIE Proceedings will open in the SPIE Digital Library. external link icon
View Session icon_mobile_dropdown
  • Hardware Applications I
  • Applications in Communications
  • Hardware Applications II
  • Applications in Control and Physics
  • Hybrid Architectures II
  • Applications in Control and Physics
  • Applications in Image Processing I
  • Biological Applications
  • Hybrid Architectures I
  • Hybrid Architectures II
  • Applications in Pattern Recognition I
  • Learning IV
  • Learning I
  • Learning II
  • Applications I
  • Learning III
  • Applications in Pattern Recognition II
  • Applications II
  • Biological Applications
  • Applications III
  • Hardware Applications II
  • Learning IV
  • Biological Applications
  • Hybrid Architectures II
  • Learning IV
  • Applications III
  • Applications in Speech
  • Applications in Pattern Recognition III
  • Learning IV
  • Applications in Pattern Recognition III
  • Applications in Image Processing II
  • Applications in Image Processing and ATR
  • Applications in Space Technology
  • General Aspects of Neural Networks
  • Neural Networks for Offline Analysis in High-Energy Physics I
  • Neural Networks for Offline Analysis in High-Energy Physics II
  • Pattern Recognition and Online Analysis
  • Hardware Implementations of Neural Networks
  • Applications in Control and Physics
  • Learning IV
  • Applications in Pattern Recognition I
Hardware Applications I
icon_mobile_dropdown
Realization of a neuronal hardware with digital signal processor and programmable gate arrays
Anke Meyer-Baese, Uwe Meyer-Baese, Henning Scheich
In this paper we describe how the processing speed of a radial basis neural network can be performed by the use of field programmable gate arrays (FPGA). The calculation of the very time-consuming exponential function is taken by an optimized CORDIC-processor. We determine the number of the necessary FPGAs and do a processing speed comparison between FPGA and DSP referring to an application in speech recognition.
High-speed VMEbus based analog neurocomputing architecture for image classification
Mua D. Tran, Tuan A. Duong, Raoul Tawel, et al.
To fully exploit the real-time computational capabilities of neural networks (NN) -- as applied to image processing applications -- a high performance VMEbus based analog neurocomputing architecture (VMENA) is developed. The inherent parallelism of an analog VLSI NN embodiment enables a fully parallel and hence high speed and high-throughput hardware implementation of NN architectures. The VMEbus interface is specifically chosen to overcome the limited bandwidth of the PC host computer industrial standard architecture (ISA) bus. The NN board is built around cascadable VLSI NN chips (32 X 32 synapse chips and 32 X 32 neuron/synapse composite chips) for a total of 64 neurons and over 8 K synapses. Under software control, the system architecture could be flexibly reconfigured from feedback to feedforward and vice versa, and once selected, the NN topology (i.e. the number of neurons per input, hidden, and output layer and the number of layers) could be carved out from the set of neuron and synapse resources. An efficient hardware-in-the-loop cascade backpropagation (CBP) learning algorithm is implemented on the hardware. This supervised learning algorithm allows the network architecture to dynamically evolve by adding hidden neurons while modulating their synaptic weights using standard gradient-descent backpropagation. As a demonstration, the NN hardware system is applied to a computationally intensive map-data classification problem. Training sets ranging in size from 50 to 2500 pixels are utilized to train the network, and the best result for the hardware-in-the-loop learning is found to be comparable to the best result of the software NN simulation. Once trained, the VMENA subsystem is capable of processing at approximately 75,000 feedforward passes/second, resulting in over twofold computational throughput improvement relative to the ISAbus based neural network architecture.
Analog VLSI signal fuzzifier
Teresa Wong, Mohammed Ismail
In this paper, an analog fuzzifier is designed and verified by SPICE simulation. Building blocks of winner-take-all circuits were developed to build the fuzzifier. The fuzzifier generates the membership functions of a fuzzy system by converting the input analog signals to their corresponding predetermined grade of memberships. It takes voltage-mode inputs and produces current-mode outputs for easy manipulations of the output fuzzy signals by the arithematic operations that follow. Furthermore, the fuzzifier is so flexible such that it can be readily adapted into most fuzzy systems, including neural networks, by changing the reference voltages and the number of building blocks used. The design of each building block is relatively simple as it only consists of four PMOS, four resistors and three current sources. The high speed of analog implementation also makes this fuzzifier more attractive than software implementations of the fuzzification process.
Neural network based data analysis for chemical sensor arrays
Sherif Hashem, Paul E. Keller, Richard T. Kouzes, et al.
Compact, portable systems capable of quickly identifying contaminants in the field are of great importance when monitoring the environment. In this paper, we examine the effectiveness of using artificial neural networks for real-time data analysis of a sensor array. Analyzing the sensor data in parallel may allow for rapid identification of contaminants in the field without requiring highly selective individual sensors. We use a prototype sensor array which consists of nine tin-oxide Taguchi-type sensors, a temperature sensor, and a humidity sensor. We illustrate that by using neural network based analysis of the sensor data, the selectivity of the sensor array may be significantly improved, especially when some (or all) of the sensors are not highly selective.
Applications in Communications
icon_mobile_dropdown
Signal Processing and Neural Network Simulator
Dennis L. Tebbe, Thomas J. Billhartz, John R. Doner, et al.
The signal processing and neural network simulator (SPANNS) is a digital signal processing simulator with the capability to invoke neural networks into signal processing chains. This is a generic tool which will greatly facilitate the design and simulation of systems with embedded neural networks. The SPANNS is based on the Signal Processing WorkSystemTM (SPWTM), a commercial-off-the-shelf signal processing simulator. SPW provides a block diagram approach to constructing signal processing simulations. Neural network paradigms implemented in the SPANNS include Backpropagation, Kohonen Feature Map, Outstar, Fully Recurrent, Adaptive Resonance Theory 1, 2, & 3, and Brain State in a Box. The SPANNS was developed by integrating SAIC's Industrial Strength Neural Networks (ISNN) Software into SPW.
Neural network based buffer allocation scheme for ATM networks
Ercan Sen, Abhijit S. Pandya
In this paper, we propose an artificial neural network (ANN) based buffer allocation scheme for ATM networks. The proposed scheme assumes limited buffer capacity and nonuniform incoming traffic patterns. The effective utilization of limited buffer capacity under nonuniform traffic conditions is a major concern in maintaining acceptable level of quality of service in these high speed communication networks. The proposed ANN based approach takes advantage of the favorable control characteristics of neural networks such as high adaptability and high speed collective computing power for effective buffer utilization. The proposed model uses complete sharing buffer allocation strategy and enhances its performance for high traffic loads and highly asymmetric traffic patterns by regulating the buffer allocation process dynamically via a neural network based controller. In this study, we considered the buffer allocation problem in the context of routing optimization in ATM networks.
Application of neural networks to the dynamic spatial distribution of nodes within an urban wireless network
William S. Hortos
The optimal location of wireless transceivers or communicating sensor devices in an urban area and within large human-made structures is considered. The purpose of the positioning of the devices is formation of a distributed network, either in a mesh or hub-spoke topology, that achieves robust connectivity of the nodes. Real-world examples include wireless local area networks (LANs) within buildings and radio beacons in an outdoor mobile radio environment. Operating environments contain both fixed and moving interferers that correspond to both stationary and time-varying spatial distributions of path distortion of stationary and transient fading and multipath delays that impede connectivity. The positioning of the autonomous wireless devices in an area with an unknown spatial pattern of interferers would normally be a slow incremental process. The proposed objective is determination of the spatial distribution of the devices to achieve the maximum radio connectivity in a minimal number of iterative steps. Impeding the optimal distribution of wireless nodes is the corresponding distribution of environmental interferers in the area or volume of network operation. The problem of network formation is posed as an adaptive learning problem, in particular, a self-organizing map of locally competitive wireless units that recursively update their positions and individual operating configurations at each iterative step of the neural algorithm. The scheme allows the wireless units to adaptively learn the pattern distribution of interferers in their operating environment based on the level of radio interference measured at each node by an equivalent received signal strength from wireless units within the node's hearing distance. Two cases are considered. The first is an indoor human-made environment where the interference pattern is largely deterministic and stationary and the units are positioned to form a wireless LAN. The second situation applies to an outdoor urban environment, where a fixed number of units on mobile platforms operating in a random spatial distribution of interferers.
Hardware Applications II
icon_mobile_dropdown
Mixed analog/digital VLSI design and simulation of an adaptive resonance theory (ART) neural network architecture
Juin J. Liou, Ching S. Ho, Christos G. Christodoulou, et al.
This paper presents a mixed analog/digital circuit design for an adaptive resonance theory (ART) architecture, called the augmented adaptive resonance theory-I neural network (AART1-NN). The circuit is implemented based on the transconductance-mode approach and mixed analog/digital components, in which analog circuits are used to fully incorporate the parallel mechanism of the neural network, whereas digital circuits are used to provide a reduced circuit size as well as more precise multiplication operation. It is shown that the Pspice simulation results of the implemented circuit are in good agreement with the results calculated numerically from the coupled nonlinear differential equations governing the AART1-NN.
Hardware implementation of a neural network controller for a manipulator arm
Rosalyn S. Hobson, Rafael M. Inigo
In a previous paper, a neural network with a reward/punish learning scheme controller for a manipulator arm was described. The inputs to the torque-generating neuron are the position error and the velocity of the joints. The output of the neuron is the torque required to control the arm to its desired position. The reward/punish learning mechanism is implemented to adaptively modify the weights. The neural network controller does not need a dynamic model of the arm. The dynamics are learned through training. In this paper we describe the hardware/software implementation of the neural network to control the shoulder joint of a Mitsubishi RM501 arm. Once the system was checked for correct operation the following tests were performed: (1) training the arm to hold is position at different angles (10, 40, 70, 100 and 120 degrees). The angle was to hold with very small error, even in the presence of significant disturbances, after a training period that varied from 3 to 12 seconds. (2) Training the arm at 50 degrees and then commanding it to follow a cosine trajectory from 50 to 70 degrees. The maximum error in this test was less than 1% of the desired value.
Hybrid optoelectronic neurocomputer: variants of realization
Nickolay N. Evtikhiev, Rostislav S. Starikov, Igor B. Scherbakov, et al.
The optoelectronic devices are the most effective for realization in the form of the vector- matrix multiplier. Proposed optoelectronic neurocomputers (OENC) consist of optical vector- matrix multiplier (OVMM), random access memory (RAM) and electronic control system. There are two variants of realization. The first neurocomputer scheme includes OVMM based on MAOM -- multichannel multifrequency acousto-optic modulator (Bragg cell). MAOM is the fastest up-to-date spatial light modulator. The second neurocomputer is constructed on the basis of planar OVMM (POVMM). Vector-matrix multiplication in POVMM is executed in a very small volume. The POVMM consists of matrix of light emitting diodes and array of linear photodetectors. A special computer program `NEUROEMULATOR' was designed to learn and to test performance of neural networks. Neural networks were trained with gradient and stochastic algorithms. The paper presents results of computer simulation and hardware implementation of neural networks.
Applications in Control and Physics
icon_mobile_dropdown
Adaptive real-time neural network attitude control of chaotic satellite motion
Nenad Koncar, Antonia J. Jones
We present a control strategy that is able to both slow down and direct the orientation of a satellite subjected to chaotic feedback forces. The experiments show that the neural-based control system is able to do this in real-time even on very inexpensive hardware such as a PC. Although no prior knowledge of the moments of inertia of the satellite are assumed, implicit knowledge of the nature of the problem is exploited in the form of a heuristic control strategy employed in the training process. The controller is adaptive to possibly changing satellite dynamics (e.g. modifications and additions to the satellite such as the one done to the Hubble telescope). The control torques produced by the controller are smooth and the system demonstrates the ability to bring the satellite into a desired orientation in the presence of large chaotic external forces in addition to noise in the sensor system. We contrast our approach to that of conventional heuristic, time-optimal and fuel-optimal control methods as well as earlier similar work using a more general neuro-genetic architecture developed by Dracopoulus and Jones.
Real-time stable adaptive control implementation using a neural network processor
Timothy Robinson, Mohammad Bodruzzaman, Kevin L. Priddy, et al.
Helicopters are highly non-linear systems that have dynamics that change significantly with respect to environmental conditions. The system parameters also vary heavily with respect to velocity. These nonlinearities limit the use of traditional fixed controllers, since they can make the aircraft unstable. The purpose of this paper is to make contributions to the development of an `intelligent' control system that can be applied to complex problems such as this in real- time. Using a slowly changing model and a simplified nonlinear model as examples, a neural network based controller is shown to have the ability to learn from these example plants and to generalize this knowledge for previously unseen plants. The adaptability comes from a neural network that adjusts coefficients of the controller in real-time while running on the accurate automation neural network processor.
Hybrid Architectures II
icon_mobile_dropdown
Boltzmann machine generation of initial asset distributions
Wael R. Elwasif, Laurene V. Fausett, Sam Harbaugh
Boltzmann machines have been used to solve a variety of combinatorial optimization problems. The use of simulated annealing allows the network to evolve into a state of maximum consensus (or minimum energy) corresponding to an optimal solution of the given problem. In this paper, Boltzmann machine neural networks (without learning) are used to generate initial configurations of assets for a generic game (e.g. chess). The desired distribution of playing pieces is subject to restrictions on the number of pieces (of several different types) that are present, as well as some preferences for the relative positions of the pieces. The rules implemented in the network allow for flexibility in assigning locations for available resources while the probabilistic nature of the network introduces a degree of variability in the solutions generated. The architecture of the network and a method for assigning the weights are described. Results are given for several different examples.
Applications in Control and Physics
icon_mobile_dropdown
Image and particle track filtering using a "dynamic" cellular automaton coupled to a neural network
Marco Casolino, M. P. Martegani, Piergiorgio Picozza
In this paper the noise removal capabilities of a cellular automaton are applied in two different fields. The first application is performed on 4 Gev pion and electron experimental events taken at Cern PS with a silicon tungsten tracking calorimeter. Particle interaction with the material of the calorimeter can produce different interactions resulting in energy releases and topology patterns dependent on the primary particle nature. The evolution rules devised for the CA have therefore to reckon with these different topologies in order to remove noise and restore interrupted tracks. The distributions of some discriminating parameters are compared with Monte Carlo data before and after filtering by the automaton and the agreement is shown to improve if pions are considered. To successfully take into account electromagnetic showers, more than one different evolutionary rule has to be considered. A neural network accordingly trained selects each step of the evolutions closer to the training classes. Upon convergence of these two different `paths,' obtained with dynamic update rules, the image with the highest output results is filtered and classified. The second use of cellular automata is in DNA sequence autoradiograph films. These images may be filtered by a CA which improves nucleotide readability and speeds up sequencing process.
Applications in Image Processing I
icon_mobile_dropdown
Neural network approach to edge detection and noise reduction in low-contrast images
Christopher M. Johnson, Edward W. Page, Gene A. Tagliarini
Numerous vision applications rely upon efficient techniques for detecting edges in an image. Edge detection is especially difficult in low-contrast images which are characterized by the general lack of sharp variations in the gray-scale intensity values between objects of interest and their backgrounds. In low-contrast images, the application of commonly employed edge detection algorithms may result in excessive noise. This paper presents a neural network model which enhances edges and reduces noise in low-contrast gray-scale images. A neural element is associated with each pixel in an image. Each neuron receives weighted gray-scale inputs from its immediate neighbors. The weights associated with the gray-scale inputs are determined through a fuzzy compatibility function that grades the degree of similarity between the gray-scale intensity values of neighboring pixels. The neural element sums its weighted inputs and subjects the weighted sum to a sigmoid function that produces gray-scale outputs ranging between 0 and 255. The slope of the sigmoid function is chosen to force resulting pixel values away from mid-range values and closer to either 0 or 255. The resulting image is then subjected to the Sobel edge detection algorithm. The technique is illustrated by applying it to several low-contrast infrared images containing military vehicles. The results show significant noise reduction and edge enhancement.
Broccoli/weed/soil discrimination by optical reflectance using neural networks
Federico Hahn
Broccoli is grown extensively in Scotland, and has become one of the main vegetables cropped, due to its high yields and profits. Broccoli, weed and soil samples from 6 different farms were collected and their spectra obtained and analyzed using discriminant analysis. High crop/weed/soil discrimination success rates were encountered in each farm, but the selected wavelengths varied in each farm due to differences in broccoli variety, weed species incidence and soil type. In order to use only three wavelengths, neural networks were introduced and high crop/weed/soil discrimination accuracies for each farm were achieved.
Neural network application in support of software reliability engineering
Taghi M. Khoshgoftaar, David L. Lanning, Abhijit S. Pandya
This paper presents a novel application of neural networks to the problem of classifying software modules into different risk classes based upon source code measures. Neural network models that classify program modules as either high-risk or low-risk are developed. Inputs to these networks include a selection of source code measure data and fault data that were collected from two large commercial systems. The criterion variable for class determination was a quality measure of program faults or changes. Discriminant models using the same data sets provide for a comparative analysis. The neural network technique displayed better classification error rates on both data sets. These successes demonstrate the utility of neural networks in isolating high-risk modules.
Biological Applications
icon_mobile_dropdown
Neurons, psychons, and emotion
Ying Liu
It is proposed by Eccles that (1) The mental world consists of mental units called psychons; (2) The brain world consists of nerve units called dendrons; (3) In mind-brain interaction, psychons and dendrons are linked through quantum physics. In this note, we formalize the above ideas. Emotions are represented by emotion state vectors. The motion of emotion (emotion state vector) depends on the previous psychon state vector, the previous emotion and brain state. The brain state is in turn stimulated by outside stimuli and influenced by psychon state vectors. In this paper, we specify the mind-brain interaction by several operators. A complete theory is presented in this paper.
Neural network model of cortical EEG response to olfactory stimuli
George L. Dunbar, Steve Van Toller
We describe three experiments attempting to model differences in cortical EEG following stimulation with different odors. The data used in these experiments was obtained in previous studies, described briefly here. Subjects sit in an environmentally stabilized low odor cubicle. Twenty-eight electrodes are placed on the scalp and connect the subject to a neurosciences brain imager, which digitizes cortical EEG response. In a given trial, a specific odor is introduced, and the response recorded. In the first experiment, alpha wave data from a subset of ten electrodes and a single subject was used. In the original experiment, the subject was presented with a number of odors and the resulting brain electrical activity was resolved into 16 time slices (5 preceding presentation, 4 during presentation and 7 following presentation). Only data from frames 6, 7 and 8 (during presentation) was used here. A model was constructed to discriminate morning from afternoon responses. The network used measurements from 10 electrodes as input, and backpropagation was used for training. During training, the network was presented with responses to just one odor. Generalization was demonstrated for five other odors. The weights in the network have been analyzed and indicate a role for a specific group of electrode sites in this discrimination. The second experiment involved constructing a network to discriminate cortical EEG responses to two odors. In the original experiment from which we drew our data, fourteen subjects were presented with each odor once. Data from only the frame at first presentation of the odor were used here. Data from three subjects (chosen pseudo-randomly) was selected for use in the generalization phase and dropped from the training set. Output targets were constructed that took account of subjective ratings of `pleasantness.' A feed-forward network with twenty-eight input units was trained using data from the eleven remaining subjects, using conjugate gradient descent.
Practical application of artificial neural networks in the neurosciences
Antonio Pinti
This article presents a practical application of artificial multi-layer perceptron (MLP) neural networks in neurosciences. The data that are processed are labeled data from the visual analysis of electrical signals of human sleep. The objective of this work is to automatically classify into sleep stages the electrophysiological signals recorded from electrodes placed on a sleeping patient. Two large data bases were designed by experts in order to realize this study. One data base was used to train the network and the other to test its generalization capacity. The classification results obtained with the MLP network were compared to a type K nearest neighbor Knn non-parametric classification method. The MLP network gave a better result in terms of classification than the Knn method. Both classification techniques were implemented on a transputer system. With both networks in their final configuration, the MLP network was 160 times faster than the Knn model in classifying a sleep period.
Using neural networks to study concept formation in a sonar discrimination task
Itiel E. Dror, Faith L. Florer, Cynthia F. Moss
Behavioral data show that echolocating bats can discriminate between different speeds of a moving object. A neural network system presented with echoes of simulated bat sounds recorded from targets moving at different speeds exhibits similar performance. In both cases, discriminations were successfully made between targets moving at 30 Hz and a slower variable speed (5, 10, and 20 Hz). However, in both cases, it is not clear what underlying concept was learned and used to perform the task. In other words, did the bat and the neural network perform the tasks based on a relative concept of faster (and thus learned to recognize `this is the faster speed') or did they perform the task based on an absolute concept of 30 Hz speed (and thus learned to recognize `this is the 30 Hz speed')? Both concepts would produce the observed performance of the bat and the neural networks. In this paper, we developed an approach for using neural networks to explore which concept was used to perform the task, and what factors may influence which type of concept is developed. First, we observed the behavior of the neural network when presented with a novel speed of 40 Hz. Its behavior on the novel 40 Hz speed suggested what underlying concept was implemented in the system (if it classified the novel 40 Hz as `30 Hz,' then it used a relative concept of faster; if it classified the novel 40 Hz speed as `not-30 Hz,' then it used an absolute concept, specifically quantifying the 30 Hz speed). Second, we examined what factors influenced the formation of the concept. Our computer simulations of neural networks that discriminated between 30 Hz and the slower variable speeds of 5, 10, and 20 Hz showed that the formation of a concept depends on the magnitude of the difference between the speeds used in training, i.e., the difficulty of the task. Specifically, when the difference between the speeds in the training set was large, the system formulated a relative concept of fast and slow. However, when the difference was small, the system formulated an absolute concept. Our neural network simulation of a system that learns to discriminate between a variety of speeds (i.e., 5, 10, and 20 Hz vs. 30 Hz speeds) demonstrated that the network developed the simplest concept that allowed it to perform the task correctly. Taken together, our results show that the development of concepts is influenced by the underlying computational demands of the task, and that neural networks can be used to explore the processes involved in concept formation.
Pulsed recursive neural networks and resource allocation, part 1: static allocation
Laurent Herault
This paper presents a new recursive neural network to solve optimization problems. It is made of binary neutrons with feedbacks. From a random initial state, the dynamics alternate successively several pulsation and constraint satisfaction phases. Applying the previous neural network has the following advantages: (1) There is no need to precisely adjust some parameters in the motion equations to obtain good feasible solutions. (2) During each constraint satisfaction phase, the network converges to a feasible solution. (3) The convergence time in the constraint satisfaction phases is very fast: only a few updates of each neuron are necessary. (4) The end user can limit the global response time of the network which regularly provides feasible solutions. This paper describes such a neural network to solve a complex real time resource allocation problem and compare the performances to a simulated annealing algorithm.
Hybrid Architectures I
icon_mobile_dropdown
Fuzzy algorithms for learning vector quantization: generalizations and extensions
Nicolaos B. Karayiannis, Pin-I Pai
This paper presents a general methodology for the development of fuzzy algorithms for learning vector quantization (FALVQ). These algorithms can be used to train feature maps to perform pattern clustering through an unsupervised learning process. The development of FALVQ algorithms is based on the minimization of a fuzzy objective function, formed as the weighted sum of the squared Euclidean distances between an input vector, which represents a feature vector, and the weight vectors of the map, which represent the prototypes. This formulation leads to the development of genuinely competitive algorithms, which allow all prototypes to compete for matching each input. The FALVQ 1, FALVQ 2, and FALVQ 3 families of algorithms are developed by selecting admissible generalized membership functions with different properties. The efficiency of the proposed algorithms is illustrated by their use in codebook design required for image compression based on vector quantization.
Fusing human knowledge with neural networks in machine condition monitoring systems
David G. Melvin, J. Penman
There is currently much interest in the application of artificial neural network (ANN) technology to the field of on-line machine condition monitoring (CM) for complex electro- mechanical systems. In this paper the authors discuss, with the help of an industrial case study, a few of the difficulties inherent in the application of neural network based condition monitoring. A method of overcoming these difficulties by utilizing a combination of human knowledge, encoded using techniques borrowed from fuzzy logic, Kohonen neural networks, and statistical K-means clustering has been constructed. The methodology is discussed in the paper by means of a direct comparison between this new approach and a purely neural approach. An analysis of other situations where this approach would be applicable is also presented and the paper discusses other current research work in the area of hybrid AI technologies which should assist further with the alleviation of the problems under consideration.
Adaptive system for generating neural networks using genetic algorithms
Armin Schneider
An adaptive system is described which generates and trains neural networks using genetic algorithms. A genetic algorithm optimizes the network architecture trying to use as few connections as possible. The neurons of the networks generated by this algorithm are not necessarily organized in layers (except input and output). Because of this, classical algorithms for training neural networks can not be used. Therefore a second genetic algorithm is used to optimize the weights for each generated architecture. During simulation it is possible to change the parameters for the genetic algorithms like the mutation probability or the population size, the size of the networks generated as well as the desired size of the input and output layer and even the data used for training the networks. Therefore the system is able to adapt to a changing environment. The system generates C/C++ code for a `recall only' version of the best network found.
Hybrid Architectures II
icon_mobile_dropdown
Application of knowledge-based network processing to automated gas chromatography data interpretation
Alan P. Levis, Robert G. Timpany, Wayne E. Austad, et al.
A method of translating a two-way table of qualified symptom/cause relationships into a four layer expert network for diagnosis of machine or sample preparation failure for gas chromatography is presented. This method has proven to successfully capture an expert's ability to predict causes of failure in a gas chromatograph based on a small set of symptoms, derived from a chromatogram, in spite of poorly defined category delineations and definitions. In addition, the resulting network possesses the advantages inherent in most neural networks: the ability to function correctly in the presence of missing or uncertain inputs and the ability to improve performance through data-based training procedures. Acquisition of knowledge from the domain experts produced a group of imprecise cause-to-symptom relationships. These are reproduced as parallel pathways composed of symptom-filter-combination-cause node chains in the network representation. Each symptom signal is passed through a filter node to determine if the signal should be interpreted as positive or negative evidence and then modified according to the relationship established by the domain experts. The signals from several processed symptoms are then combined in the combination node(s) for a given cause. The resulting value is passed to the cause node and the highest valued cause node is then selected as the most probable cause of failure.
Hopfield networks and scheduling problems
In this paper we present a neural generator method that uses a neural network to generate initial search points for a discrete heuristic. We demonstrate the method for the subset-sum problem (SSP), and consider the SSP to be typical of the sub-problems that a scheduling algorithm must solve while on route to solving an entire scheduling problem. The neural generator method hinges on using the continuous valued activations of the neural system to select a corner of the n-cube that can be used to initialize a discrete search. This can be done at each neural iteration, resulting in many discrete searches over the source of a single neural run. Without the discrete heuristic, the selected corners can be interpreted as instantaneous neural solutions and the best-so-far tabulated as the neural system runs. This allows the neural system to be terminated without losing the full effort of the run, and should the network be run until convergence, the best-so-far result is at least as good as the convergent corner, and usually better. With the discrete heuristic, a search is launched from the instantaneous neural solutions, greatly improving the overall results (again keeping the best-so-far). The results are presented for an n equals 25 SSP.
Use of genetic algorithms for encoding efficient neural network architectures: neurocomputer implementation
Jason James, Cihan H. Dagli
In this study an attempt is being made to encode the architecture of a neural network in a chromosome string for evolving robust, fast-learning, minimal neural network architectures through genetic algorithms. Various attributes affecting the learning of the network are represented as genes. The performance of the networks is used as the fitness value. Neural network architecture design concepts are initially demonstrated using a backpropagation architecture with the standard data set of Rosenberg and Sejnowski for text to speech conversion on Adaptive Solutions Inc.'s CNAPS Neuro-Computer. The architectures obtained are compared with the one reported in the literature for the standard data set used. The study concludes by providing some insights regarding the architecture encoding for other artificial neural network paradigms.
Applications in Pattern Recognition I
icon_mobile_dropdown
Automatic target recognition using neural networks
Steven K. Rogers, John M. Colombi, Curtis E. Martin, et al.
Many applications reported in artificial neural networks are associated with military problems. This paper reviews concepts associated with the processing of military data to find and recognize targets -- automatic target recognition (ATR). A general-purpose automatic target recognition system does not exist. The work presented here is demonstrated on military data, but it can only be considered proof of principle until systems are fielded and proven `under- fire.' ATR data can be in the form of nonimaging one-dimensional sensor returns, such as ultra-high range-resolution radar returns (UHRR) for air-to-air automatic target recognition and vibration signatures from a laser radar for recognition of ground targets. The ATR data can be two-dimensional images. The most common ATR images are infrared, but current systems must also deal with synthetic aperture radar (SAR) images. Finally, the data can be three-dimensional, such as sequences of multiple exposures taken over time from a nonstationary world. Targets move, as do sensors, and that movement can be exploited by the ATR. Hyperspectral data, which are views of the same piece of the world looking at different spectral bands, is another example of multiple image data; the third dimension is now wavelength and not time. ATR system design usually consists of four stages. The first stage is to select the sensor or sensors to produce the target measurements. The next stage is the preprocessing of the data and the location of regions of interest within the data (segmentation). The human retina is a ruthless preprocessor. Physiologically motivated preprocessing and segmentation is demonstrated along with supervised and unsupervised artificial neural segmentation techniques. The third design step is feature extraction and selection: the extraction of a set of numbers which characterize regions of the data. The last step is the processing of the features for decision making (classification). The area of classification is where most ATR related neural network research has been accomplished. The relation of neural classifiers to Bayesian techniques is emphasized along with the more recent use of feature sequences to enhance classification.
Feature space trajectory neural net classifier
Leonard Neiberg, David P. Casasent
A new classifier neural network is described for distortion-invariant multi-class pattern recognition. The input analog neurons are a feature space. All distorted aspect views of one object are described by a trajectory in feature space. Classification of test data involves calculation of the closest feature space trajectory. Pose estimation is achieved by determining the closest line segment on the closest trajectory. Rejection of false class clutter is demonstrated. Comparisons are made to other neural network classifiers, including a radial basis function and a new standard backpropagation neural net. The shapes of the different decision surfaces produced by our feature space trajectory classifier are analyzed.
Automatic ship classification system for inverse synthetic aperture radar (ISAR) imagery
Murali M. Menon
The U.S. Navy has been interested in applying neural network processing architectures to automatically determine the naval class of ships from an inverse synthetic aperture radar (ISAR) on-board an airborne surveillance platform. Currently an operator identifies the target based on an ISAR display. The emergence of the littoral warfare scenario, coupled with the addition of multiple sensors on the platform, threatens to impair the ability of the operator to identify and track targets in a timely manner. Thus, on-board automation is quickly becoming a necessity. Over the past four years the Opto-Radar System Group at MIT Lincoln Laboratory has developed and fielded a neural network based automatic ship classification (ASC) system for ISAR imagery. This system utilizes imagery from the APS-137 ISAR. Previous related work with ASC systems processed either simulated or real ISAR imagery under highly controlled conditions. The focus of this work was to develop a ship classification system capability of providing real-time identification from imagery acquired during an actual mission. The ship classification system described in this report uses both neural network and conventional processing techniques to determine the naval class of a ship from a range- Doppler (ISAR) image. The `learning' capability of the neural network classifier allows a single naval class to be distributed across many categories such that a degree of invariance to ship motion is developed. The ASC system was evaluated on 30 ship class database that had also been used for an operational readiness evaluation of ISAR crews. The results of the evaluation indicate that the ASC system has a performance level comparable to ISAR operators and typically provides a significant improvement in throughput.
Learning IV
icon_mobile_dropdown
Feature extraction and classification of radar targets using neural networks
Doraisamy Nandagopal, M. Palaniswami, N. M. Martin, et al.
In signal processing, artificial neural networks (ANN) have been found to be very useful in solving pattern recognition and classification problems. In this application, the performance of ANNs depends, to a large extent, on the quality of features extracted from the given signal. The features, most often, are extracted using conventional signal processing techniques. In this paper, the feature extraction of radar returns is carried out through the use of neural networks and the final recognition of radar targets is carried out by a second stage neural network. Thus feature extraction and classification of experimental radar targets using feed forward and self organizing neural networks are demonstrated.
Learning I
icon_mobile_dropdown
Temporal generalization capability of simple recurrent networks
Xiaomei Liu, DeLiang Wang, Stanley C. Ahalt
Simple recurrent networks have been widely used in temporal processing applications. In this study we investigate temporal generalization of simple recurrent networks, drawing comparisons between network capabilities and human characteristics. Elman networks were trained to regenerate temporal trajectories sampled at different rates, and then tested with trajectories at both the trained sampling rates and at other sampling rates. The networks were also tested with trajectories representing mixtures of different sampling rates. It was found that for simple trajectories, the networks show interval invariance, but not rate invariance. However, for complex trajectories which contain greater contextual information, these networks do not seem to show any temporal generalization. Similar results were also obtained employing measured speech data. Thus, these results suggest that this class of networks exhibits severe limitations in temporal generalization.
Habituation-based mechanism for encoding temporal information in artificial neural networks
Bryan W. Stiles, Joydeep Ghosh
A novel neural network is proposed for the dynamic classification of spatio-temporal signals. The network is designed to classify signals of different durations, taking into account correlations among different signal segments. Such a network is applicable to SONAR and speech signal classification problems, among others. Network parameters are adapted based on the biologically observed habituation mechanism. This allows the storage of contextual information, without a substantial increase in network complexity. Experiments on classification of high dimensional feature vectors obtained from Banzhaf sonograms, demonstrate that the proposed network performs better than time delay neural networks while using a less complex structure. A mathematical justification of the capabilities of the habituation based mechanism is also provided.
Survey of learning results in adaptive resonance theory (ART) architectures
Michael Georgiopoulos, J. Huang, Gregory L. Heileman
In this paper we investigate the learning properties of ART1, Fuzzy Art, and ARTMAP architectures. These architectures were introduced by Carpenter and Grossberg over the last eight years. some of the learning properties discussed in this paper involve characteristics of the clusters formed in these architectures while other learning properties concentrate on how fast it will take these architectures to converge to a solution for the type of problems that are capable of solving. This latter issue is very important in the neural network literature, and there are very few instances where it has been answered satisfactorily.
Learning II
icon_mobile_dropdown
Interior-point competitive learning of control agents in colony-style systems
Michael D. Lemmon, Peter T. Szymanski, Chris Bett
This paper presents an alternating minimization (AM) algorithm used in training radial basis function (RBF) networks. The AM algorithm can also be viewed as a competitive learning paradigm. Its use is illustrated by optimizing a colony-style control system. The application arises in the context of hybrid control systems. The algorithm is a modification of a small-step interior point method used in solving primal linear programs. The algorithm has a convergence rate of 0((root)nL) iterations where n is a measure of the network size and L is a measure of the resulting solution's accuracy.
Can neural nets learn dynamic systems with attractors?
Daw-Tung Lin
The idea of storing memory states or patterns to be recognized or recalled as static attractors of the neural network dynamics implies that initial configurations of neurons in some neighborhood of a memory stated will be attracted to it. In applications, these attractors could represent memories, pattern classes, or stable control actions. A novel property of autonomous spatio-temporal pattern recognition and production with the Adaptive Time-delay Neural Network (ATNN) has been explored. The ATNN, a paradigm for training a nonlinear neural network with adaptive time-delays and weights, has a rich repertoire of capabilities that are used to perform signal production, and to learn repetitive spatial motions. ATNN has the property that initial segments on signals that contain large amounts of noise can be "cleaned up" to result in trained trajectory motions. The results on noise removal suggest that the trajectories trained into the ATNN networks are in fact attractors. We conducted experiments on the basins of attraction for the circle and figure eight attractors in these networks. The circular and figure eight trajectories can be considered limit cycle attractors or multistate oscillations because of the repetition of points along the closed figures. In spite of the fact that different initial arcs were used, including very noisy arcs, when starting the network trajectory, the nets always arrived at the trajectory for which they were trained. Thus the initial arcs were within the basin of attraction for the trained figures (attractors) both for the circle and figure eight. The impact of this result is that an autonomous or controlled system, such as a robotic arm, or other moving object, could be trained to generate repetitive desired motions, and could attain this repetitive motion from arbitrary starting trajectories.
Learning ranks with neural networks
Khaled A. Al-Ghoneim, Bhagavatula Vijaya Kumar
Complex pattern recognition problems are not usually solved by a single classifier. Multiple classifiers are used instead. Typically, the classifiers are arranged hierarchically, such that the low level classifier (LLC) produces not only a single decision of its best guess of the class of the input pattern but a list of choices ranked according to their likelihood. The high level classifier (HLC) then chooses from this set of classes using additional information that is not usually available to or well represented in a single LLC, such as knowledge of the context or the model. Training neural networks (NNs) as low level classifiers has been traditionally performed independent of what the HLC may do. The traditional performance measure for evaluating classifiers is the classification counting function which counts the number of correct classifications performed by that classifier. It is of course desired that the LLC produces the correct classification (by ranking the correct class as the top choice). Moreover, it is preferred that the LLC ranks the correct class as the second choice if it is not able to correctly classify it (as the first choice). A new cost function (which accounts for the correctness of class rankings) is presented. When this cost function is optimized, it will achieve this desired ranking performance. The parameters of this new cost function will be linked to statistical parameters of our proposed hierarchical model. Unfortunately, this cost function cannot be used to train neural networks because it is not differentiable. Thus we investigate differentiable approximations that are well suited to training NNs using the backpropagation algorithm. Initial simulation results show that superiority of this new error measure over the traditional mean square error measure both in terms of classification and ranking performance.
Application of artificial neural networks to gaming
Norio Baba, Tomio Kita, Kazuhiro Oda
Recently, neural network technology has been applied to various actual problems. It has succeeded in producing a large number of intelligent systems. In this article, we suggest that it could be applied to the field of gaming. In particular, we suggest that the neural network model could be used to mimic players' characters. Several computer simulation results using a computer gaming system which is a modified version of the COMMONS GAME confirm our idea.
Applications I
icon_mobile_dropdown
Distributive vibration control in flexible multibody dynamics
A distributive neural control system is advocated for flexible multibody structures. The proposed neural controller is designed to achieve trajectory slewing of a structural member as well as vibration suppression for precision pointing capability. The motivation to support such an innovation is to pursue a real-time implementation of a robust and fault tolerant structural controller. The proposed control architecture which takes advantage of the geometric distribution of piezoceramic sensors and actuators has provided a tremendous freedom from computational complexity. In the spirit of model reference adaptive control, we utilize adaptive time-delay radial basis function networks as a building block to allow the neural network to function as an indirect closed-loop controller. The horizon-of-one predictive controllers cooperatively regulates the dynamics of the nonlinear structure to follow the prespecified reference models asymptotically. The proposed control strategy is validated in the experimental facility, called the Planar Articulating Controls Experiment which consists of a two-link flexible planar structure constrained to move over a granite table. This paper addresses the theoretical foundation of the architecture and demonstrates its applicability via a realistic structural test bed.
Rejection of unfamiliar patterns with multilayer neural networks
Behrooz Kamgar-Parsi, Behzad Kamgar-Parsi
Most of the pattern recognition applications of multilayer neural networks have been concerned with classification and not rejection of a given pattern. For example, in character recognition all alphabetical characters must be recognized as one of the 26 characters, as there is nothing to reject. However, in many situations, there is no guarantee that all the patterns that will be presented to the network would actually belong to one of the classes on which the network has been trained. In such cases, a useful network must be capable of rejection as well as classification. In this paper we propose a scheme to develop multilayer networks with rejection capabilities. The discriminating power of the proposed technique appears to be comparable to that of the human eye.
Application of neural networks for self-supervised learning
Minoru Sekiguchi, Tamami Sugasaka, Shigemi Nagata
The learning method of layered neural networks can be supervised or unsupervised. Back propagation learning algorithm is a common method of supervised learning that can learn automatically from teaching patterns. However, accurate teaching patterns are not always available for robotic applications and it is necessary to devise a method of producing them. In this paper, two applications of neural network for self-supervised learning are described. One is a system for which a mobile robot learns its behavior by automatically generating and self- evaluating teaching data through a random walk. The other is a control method of an inverted pendulum using a knowledge-based neural network. The system collects the state data of the inverted pendulum such as angles and angular velocities by trial and error. After that, the system generates teaching data by comparing the collected data with stored knowledge. This knowledge expresses the ideal status of the inverted pendulum when it inverts. The system learns from the generated teaching data and the pendulum inverts stably after some trial and error. In both systems, the neural network learns the teaching data that is generated by the system itself.
Learning III
icon_mobile_dropdown
Dynamical Boolean systems: stability analysis and applications
Paul B. Watta, Kaining Wang, Rahul Shringarpure, et al.
In this paper, recurrent neural networks are analyzed from the point of view of sequential machines. Their dynamical behavior is described by a system of coupled Boolean equations, and stability results are presented. The typical stability analysis for recurrent Hopfield-type neural nets is to define an energy function and demonstrate that it is a Liaponov function for the system. This analysis works well for single layer networks, but has not been successfully applied to multilayer networks; although, in theory, an energy function for multi-layered nets may be possible to derive. Alteratively, the stability results presented in this paper are applicable to single layer as well as multilayer recurrent networks. Furthermore, our approach is potentially more systematic and easier to apply than the ad-hoc energy function synthesis methods. As an application of this approach, we show how to design recurrent neural nets to design high performance associative neural memories.
Real-world speech recognition with neural networks
Etienne Barnard, Ronald Cole, Mark Fanty, et al.
We describe a system based on neural networks that is designed to recognize speech transmitted through the telephone network. Context-dependent phonetic modeling is studied as a method of improving recognition accuracy, and a special training algorithm is introduced to make the training of these nets more manageable. Our system is designed for real-world applications, and we have therefore specialized our implementation for this goal; a pipelined DSP structure and a compact search algorithm are described as examples of this specialization. Preliminary results from a realistic test of the system (a field trial for the U.S. Census Bureau) are reported.
Some new competitive learning schemes
James C. Bezdek, Nikhil R. Pal, Richard J. Hathaway, et al.
First, we identify an algorithmic defect of the generalized learning vector quantization (GLVQ) scheme that causes it to behave erratically for a certain scaling of the input data. We demonstrate the problem using the IRIS data. Then, we show that GLVQ can behave incorrectly because its learning rates are reciprocally dependent on the sum of squares of distances from an input vector to the node weight vectors. Finally, we propose a new family of models -- the GLVQ-F family -- that remedies the problem. We derive algorithms for competitive learning using the GLVQ-F model, and prove that they are invariant to all positive scalings of the data. The learning rule for GLVQ-F updates all nodes using a learning rate function which is inversely proportional to their distance from the input data point. We illustrate the failure of GLVQ and success of GLVQ-F with the ubiquitous IRIS data.
ReverbaProp: simultaneous learning of credit and weight
Robert Chris Lacher, Douglas Alan Klotter
We introduce a supervised learning method for feed-forward networks that solves the credit assignment problem for error in concert with solving the error reduction problem normally associated with methods such as backpropagation. The method reverberates between forward and reverse activations of the network. Forward activation using an exemplar computes output for each node in the network using the connection weights as usual. Reverse activation using output error as input computes local error at each node using reverse weights, or responsibilities, on the reverse connections. Reverse-reverse activation (the same as forward activation with linear output functions) using reverse output error as input computes local reverse error at each node. Once local error and local reverse error have been assigned to each node, weights and responsibilities are modified using the standard delta rule and local error and local reverse error, respectively. The method relies on convergence toward an optimal set of responsibilities for reverse error distribution in concert with convergence toward an optimal set of weights, and thus avoids calculation of nonlinear terms in the usual error backpropagation method. Thus the method is free of derivative evaluations, and by allowing credit assignment to optimize simultaneously with error reduction, it promotes clustering of responsibility among the nodes.
Applications in Pattern Recognition II
icon_mobile_dropdown
Clustering, simulation, and neural networks in real-world applications
Mary Lou Padgett, Eleanor M. Josephson, C. R. White, et al.
Real-world applications of neural networks often involve simulation and clustering. Reduction of subjective decisions and increasing the potential for real-time automation of cluster evaluation is a target of the cluster check (CC) method suggested here. CC quantitatively assess the variation within a cluster, produces a `central' pattern for a cluster which is robust in the presence of wide variation and skewed data, and suggests a measure for the distance between clusters. Such a measure of the effectiveness of a segmentation scheme is helpful in many applications, including traditional analysis, neural systems, fuzzy systems and evolutionary systems. This work reports successful use of the CC and companion analytic procedures to measure the consistency of movement of neuroanatomical tracer down neural pathways associated with injection sites (tract tracing). Opposite injection sites produce distinctive L shaped accumulations of tracer in different locations. Consistency of pathways for particular injection sites varies from 0.10 to 0.20 out of a possible 0.80. The pathway rejected by the nonparametric statistics and subdivided by Kohonen's self organizing map (SOM) measures 0.20. These quantitative results are consistent with the expert qualitative inspection traditionally accepted in the study of neuronanatomy of the rat olfactory bulb and tubercle. This work suggests further application of the CC and companion techniques to fault detection, identification and recovery of systems for control of diabetes and systems for control of missiles. Use of managerial decisions in the supervisory portions of these systems may also be facilitated by the consistency measure and distance metric allowing reinforcement of consistent decisions and focus on areas needing reconsideration. Automation of such procedures may facilitate real-time, robust and fault-tolerant control by adding a capability for evaluation needed for automated reinforcement and/or selection in neural, fuzzy and evolutionary systems.
Boundary variance reduction for improved classification through hybrid networks
Kagan Tumer, Joydeep Ghosh
Several researchers have experimentally shown that substantial improvements can be obtained in difficult pattern recognition problems by combining or integrating the outputs of multiple classifiers. This paper provides an analytical framework that quantifies the improvements in classification results due to linear combining. We show that combining networks in output space reduces the variance of the actual decision region boundaries around the optimum boundary. In the absence of network bias, the added classification error is directly proportional to the boundary variance. Moreover, if the network errors are independent, then the reduction in variance boundary location is by a factor of N, the number of classifiers that are combined. In the presence of network bias, the reductions are less than or equal to N, depending on the interaction between network biases. We discuss how the individual networks can be selected to achieve significant gains through combining, and support them with experimental results on 25-dimensional sonar data. The analysis presented here facilitates the understanding of the relationships among error rates, classifier boundary distributions, and combining in output space.
Applications II
icon_mobile_dropdown
Review of efforts combining neural networks and evolutionary computation
David B. Fogel, Peter J. Angeline
Since the widespread recognition of the capacity for neural networks to perform general function approximation, a variety of such mapping functions have been used to address difficult problems in pattern recognition, time series forecasting, automatic control, image compression, and other engineering applications. Although these efforts have met with considerable success, the design and training of neural networks have remained much of an art, relying on human expertise, trial, and error. More recently, methods in evolutionary computation, including genetic algorithms, evolution strategies, and evolutionary programming, have been used to assist in and automate the design and training of neural networks. This presentation offers a review of these efforts and discusses the potential benefits and limitations of such combinations.
Dynamical systems regulation by neurocontrollers
Sigeru Omatu
It is well-known that neural network systems have many typical properties compared with other information processing methods. Some of them are learning ability, generalization, and nonlinear characteristics. We propose a neuro-control method to construct adaptive control systems. Based on three kinds of neuro-control methods, we construct a corresponding neuro- controller to each method. Then we apply the method to a temperature control problem for a water bath and a heating furnace. Experimental results show that the present method provides excellent control results.
Learning to distinguish similar objects
Michael Seibert, Allen M. Waxman, Alan N. Gove
This paper describes how the similarities and differences among similar objects can be discovered during learning to facilitate recognition. The application domain is single views of flying model aircraft captured in silhouette by a CCD camera. The approach was motivated by human psychovisual and monkey neurophysiological data. The implementation uses neural net processing mechanisms to build a hierarchy that relates similar objects to superordinate classes, while simultaneously discovering the salient differences between objects within a class. Learning and recognition experiments both with and without the class similarity and difference learning show the effectiveness of the approach on this visual data. To test the approach, the hierarchical approach was compared to a non-hierarchical approach, and was found to improve the average percentage of correctly classified views from 77% to 84%.
New architecture for automatic fingerprint matching using neural networks as a feature finder and matcher
Qiang Lin, Roy S. Nutter
This research explores the use of Neural Networks (NNs) to implement Automatic Fingerprint Matching (AFM) system. A new three stage architecture, consisting of a preprocessor, feature finder, and matcher stage, is shown to be successful for fingerprint matching. The NN-based AFM system used 20 fingerprints as its training set and 80 fingerprints as its test set. By dividing a fingerprint into 256 16 x 16 pixel blocks, the achieved success matching rates are 95% on the training set and 93.75% on the test set. The Feature Finder based on the Counter Propagation NN realized a high dimension reduction ratio of 256: 1. It finds a feature vector of 256 bytes from a digitized fingerprint of 512 x 512 pixels with 8-bit grayscale. This system also achieved 91.67% matching success m the cross-iferenced fingerprints. Keywords: neural networks, fingerprints, automatic fingerprint matching systems
Newly found 3D illusions and a study on the visual mechanism
From the visual stimuli of disparity given for occluded objects only partially along the contour of an occluding object which is not physically depicted, the human visual system can perceive an entire 3-D illusory occluding object. There are several new findings concerning this illusion: an interaction between several illusory objects; an edge line perceived as an intersection of two illusory surfaces; an occlusion between illusory objects; and a transparency which is unique for binocular fusion. Two types of occlusion cues for 3-D perception are postulated and investigated. In further investigations of these illusions with motion, the author found new types of visual effects, which are named as `dynamic illusion' and could not be expected from the stational case. These newly found illusory phenomena have close relations with the visual functions of the 3-D space perception and can propose new subjects not only in the field of psychophysics and neuro-physiology but also artificial neural networks and computer vision.
Analog VLSI neural systems: trends and challenges
Mohammed Ismail
The demand for intensive computational power for real time information processing applications is rapidly increasing. These information systems are anticipated to support tera operations-per-second in 1996 and beyond. Artificial neural networks represent one of the approaches to enhance computational capabilities in real-time information processing. Analog or mixed digital/analog VLSI neural network processors have been widely developed and are preferred over digital solutions particularly for dedicated applications. We, therefore, see a strong connection between advances in analog VLSI design and the rate of progress of application driven neural network hardware. This paper is devoted primarily to advances and trends in analog VLSI, and discusses the challenges as well as the tremendous potential of analog VLSI neural nets to address real world problems. We also present an interdisciplinary view on VLSI in general and on analog VLSI in particular.
Biological Applications
icon_mobile_dropdown
Pulsed recursive neural networks and resource allocation, part 2: dynamic allocation
Laurent Herault
This paper presents a pulsed recursive neural network to solve a complex resource allocation in which some resources can be allocated to some requests over a time period.
Applications III
icon_mobile_dropdown
Neural estimation and embedology for time-series prediction
Robert E. Garza, James A. Stewart, Steven K. Rogers, et al.
Time series prediction has widespread application, ranging from predicting the stock market to trying to predict future locations of scud missiles. The embedology theorem sets forth the procedures for state space manipulation and reconstruction for time series prediction. This includes embedding the time series into a higher dimensional space in order to form an attractor, a structure defined by the embedded vectors. In this paper, embedology is combined with neural technologies in an effort to create a more accurate prediction algorithm. The algorithms tested are embedology, neural networks, and Euclidean space nearest neighbors. Local linear training methods are compared to the use of the nearest neighbors as the training set for a neural network. The results of these experiments determine that the neural algorithms have the best prediction accuracies. The performance of the nearest neighbor trained neural network validates the applicability of the local linear training set.
Hardware Applications II
icon_mobile_dropdown
Implementation of the adaptive resonance theory (ART1) architecture on a CNAPS neurocomputer
Kanchitpol Ratanapan, Cihan H. Dagli
This study reports the attempt being made to implement adaptive resonance theory (ART) neural network architecture on Adaptive Solution Inc. CNAPSR Neuro-Computer. The architecture is being implemented on a single instruction, multiple data (SIMD) parallel computer, having 256 processors. CNAPS-C language, modified ANSI C having parallel construction, is used in generating the ART code. Five important issues in the implementation are addressed. The performance of the implementation is used to bench-mark against the sequential version using standard data sets.
Learning IV
icon_mobile_dropdown
Using a viewing window and the Hausdorff-Voronoi Network (HAVNET) neural network for the recognition of words within a document
David L. Enke, Cihan H. Dagli
A substantial portion of research and applications for visual image recognition have been limited to the recognition of large, isolated, non-variant images. Performing a visual search for focusing on, locating, and recognizing smaller details within the context of a larger image has proven more difficult. This paper presents a system that is capable of learning words of various lengths, and then locating and recognizing a previously trained word within a noisy document. This system utilizes a fitness function, search routine and viewing window to identify possible word candidates, and then employs the Hausdorff-Voronoi network (HAVNET) for word recognition. After 330 searches of 30 different words, with document noise ranging from 0 - 20%, the system recognition and location accuracy were 97.3%.
Biological Applications
icon_mobile_dropdown
Study of two different biological insect eye maps with artificial neural network processing
Roy E. Williams, Ronald G. Driggers, William R. Clayton, et al.
Insect eyes have a large number of facets or lenses, also known as ommatidia or eyelets, with different arrangements of biological photoreceptors coupled to each eyelet. The output of each photoreceptor is coupled to sets of neurons where the optical information is processed. It is interesting to note that different insects are comprised of entirely different visual systems. These varying eyelet arrangements appear to be particular to the insect's habits and habitats. To test this premise, two very different insect ommatidia maps coupled to artificial neural network (NN) processors were modeled and simulated on a silicon graphics workstation. The performance of each ommatidia/NN system was tested in point source target location tasks and finite target location tasks in order to compare the two to each other and to man-made multi- aperture vision systems. The results of these simulations are presented.
Hybrid Architectures II
icon_mobile_dropdown
Self-training neural network model for tomography data processing
Yuri N. Kulchin, Oleg B. Vitrik, Oleg T. Kamenev, et al.
In this paper we present self-training two-layer neural network model for tomography data processing. This model allows us to reconstruct physical field parameters distribution by use of tomography integral data.
Cooperative system based on soft computing methods to realize higher precision of computer color recipe prediction
Eiji Mizutani, Hideyuki Takagi, David M. Auslander
This paper proposes a combinational model of neural networks (NNs) and a genetic algorithm (GA) to obtain highly precise outputs. Performance of the model is evaluated by application to a computerized color recipe prediction task, which requires relating surface spectral reflectance of a target color to several pigment concentrations. For GA search, first, predictive concentrations of color pigments are initialized by a random initializer, a multi-elite generator based on rules, and an NN which predicts pigment concentrations from the surface spectral reflectance. Then the GA starts searching for more precise pigment concentration vectors depending on a fitness function which is constructed based on three functions: (1) an NN to predict which pigments to use, (2) a rule base to deal with knowledge of color, and (3) an NN to calculate color difference to take into account human visual sensitivity. This hybrid model predicts color pigment concentrations with higher precision by fine-tuning the results of NN approaches. It may possibly show great potential in another precision problem.
Learning IV
icon_mobile_dropdown
Feedforward neural network function represented by morphological transform
Peng Tao, Jie-Gu Li
This paper is based on a binary discrete signal system. We analyze the feedforward neural network (NN) in light of morphological hit-or-miss transform (HMT), indicate that every first- hidden-layer node (the direct successors of input nodes) corresponds to a certain family of structuring elements pairs (called structuring elements pairs family determined by weights and threshold vector (w1, w2,......wM, (mu) ), and the I-O function of the node can be represented as the union of HMT by the members of the family. The structuring elements pairs family can be found by searching an algorithm of artificial intelligence. On the basis of the above analysis, we further investigate the whole feedforward NN, indicate that any recognition function (on binary discrete signal system) can be fulfilled by 3-layer (including 1 hidden-layer) feedforward NN, and can also be fulfilled by the union of HMT by a family of structuring elements pairs.
Optimal robustness in noniterative learning
If M given training patterns are not extremely similar, the analog N-vectors representing them are generally separable in the N-space. Then a one-layered binary perceptron containing P neurons (P equals >log2M) is generally sufficient to do the pattern recognition job. The connection matrix between the input (linear) layer and the neuron layer can be calculated in a noniterative manner. Real-time pattern recognition experiments implementing this theoretical result were reported in this and other national conferences last year. It is demonstrated in these experiments that the noniterative training is very fast, (can be done in real time), and the recognition of the untrained patterns is very robust and very accurate. The present paper concentrates at the theoretical foundation of this noniteratively trained perceptron. The theory starts from an N-dimension Euclidean-geometry approach. An optimally robust learning scheme is then derived. The robustness and the speed of this optimal learning scheme are to be compared with those of the conventional iterative learning schemes.
Applications III
icon_mobile_dropdown
Overview of biomedical applications of neural networks
Russell C. Eberhart, Rebecca Brandmaier
Biomedical applications of neural networks are reviewed. Early applications are briefly mentioned. Applications published in the period 1991 - 1994 provide the main focus for a selected bibliography of over 100 references appearing in the paper. References are organized according to medical discipline. Emergent patterns and trends, and predictions for future developments, are briefly discussed.
Using neural networks to predict the risk of cardiac bypass operations
Richard P. Lippmann, Linda Kukolich
Experiments demonstrate that sigmoid multilayer perceptron (MLP) networks provide slightly better risk prediction than conventional logistic regression and Bayesian models when used to predict the risk of death using a data base of 41,385 patients who underwent coronary artery bypass operations in 1993. MLP networks with no hidden layers (single-layer MLPs), networks with one hidden layer (two-layer MLPs), and networks with two hidden layers (three-layer MLPs) were trained using stochastic gradient descent with early stopping. All prediction techniques used the same input features and were evaluated by training on 20,698 patients and testing on a separate 20,687 patients. Receiver operating characteristic (ROC) curve areas for predicting mortality were roughly 75% for all classifiers. Risk stratification or accuracy of posterior probability prediction was slightly better with three-layer MLP networks which did not inflate risk for high-risk patients. Simple approaches were developed to calculate effective odds ratios for MLP networks and to generate confidence intervals for MLP risk predictions using an auxiliary `confidence MLP.' The confidence MLP is trained to reproduce confidence intervals that were generated during training using the outputs of 50 MLP networks trained with different bootstrap samples.
Analysis and implementation of the Lagrange programming neural network for image restoration
Bin Wang, Thomas F. Krile
In this paper, we propose a modification to the Lagrange programming neural network (LPNN) and its implementation procedure for maximum entropy image restoration with signal independent noise. Our approach has better transient behavior and convergence speed without imposing strict restrictions on the initial information and starting point. Gray level real images with practical size (256 X 256) can be restored in less than 10 iterations, which greatly reduces computational complexity and brings the maximum entropy image restoration technique to a practical stage. Computer simulation results and detailed discussions are provided, with comparisons to restoration using a Hopfield neural network.
Foreground/background segmentation of optical character recognition (OCR) labels by a single-layer recurrent neural network
This paper describes the development of a recurrent neural network to segment gray scale label images into binary label images. To determine a pixel label, the neural network takes into account three sources of information: pixel intensities, correlations between neighboring labels, and edge gradients. These three sources of information are succinctly combined via the network's energy function. By changing its label state to minimize the energy function, the network satisfies constraints imposed by the input image and the current label values. To be mappable to analog hardware, it is desirable that the neural equations be deterministic. Two deterministic networks are developed and compared. The first operates at the zero temperature limit, the original Hopfield network. The second employs the mean field annealing algorithm. It is shown that with only a moderate increase in computational requirements, the mean field approach produces far superior results.
Applications in Speech
icon_mobile_dropdown
Word spotting with the gamma neural model
Craig Fancourt, Neil Euliano, Jose C. Principe
This paper discusses the application of the gamma neural model to word spotting. The gamma model is a dynamic neural model where the conventional tap delay line of the TDNN is replaced by a local recursive memory structure. This model is able to find the best memory depth for a given processing task when the number of taps in the memory is specified. It can also compensate for time warping. In our approach, word spotting is the detection of a signature (the keyword under analysis) in a noisy background (other words of continuous speech). Unlike other approaches, we do not segment the input, and the neural net learns over time how to recognize the patterns associated with a given word. We test two gamma model topologies for their sensitivity to time warping and amplitude variations.
Isolated digit recognition without time alignment
Jeffrey M. Gay, Martin P. DeSimio
A method for isolated digit recognition without time alignment is examined in this paper. Rather than providing a classifier with feature vectors generated from frames of data (typically at rates near 100 per second) over the word's duration, this method uses only one feature vector per word. A baseline speaker-independent recognition accuracy of 98.1% is established with intraword time alignment from the male speaker/digit subset of a Texas Instruments database using dynamic time warping (DTW) and 12 LPC cepstral coefficients as features. Without intraword time alignment and 12 time-averages LPC cepstral coefficients as feature vectors with a multilayer perceptron (MLP) classifier, the recognition accuracy is 78.4%. By augmenting the feature vectors with 9 time-averaged critical band energy elements and 10 time-averaged LPC coefficients, the accuracy increases to 97.1%. This difference between methods is not statistically significant at the 95% confidence level. Thus, time alignment is demonstrated not to be a critical factor for the digit recognition task. Advantages of the proposed method are that (1) intraword time alignment is not required, and (2) only a single feature vector is computed per utterance. The advantages come at the expense of requiring additional information in the feature vectors relative to a DTW-based classifier.
Multisensor fusion techniques for tactical speaker recognition
Multi-sensor fusion techniques have been widely used for target and object recognition, but are relatively unheard of in the speech processing community. Multi-sensor fusion deals with the combination of complementary, and sometimes contradictory, sensor data into a reliable estimate of the environment to achieve a sum which is better than the parts. Rome Laboratory developed a tactical speaker recognition algorithm which incorporates both feature and classifier fusion. The strategy is to exploit the fact that different classifies err in different ways and multiple features, like multiple sensors, can improve recognition performance over the performance of any one feature set (or even a composite feature set). The feature sets used are LPC cepstra, Hamming liftered cepstra, RASTA liftered cepstra, and delta cepstra. Each feature set is used to train separate classifiers: K Nearest Neighbor, Hypersphere, Multilayer Perceptron and Vector Quantization classifiers. This paper details experiments whereby the results of each feature and classifier pair are fused using two different methods: a simple voting scheme (or majority strategy) and a weighted voting scheme. Results are quoted on a simulated data set and the Rome Laboratory Greenflag tactical database.
Identification of pain from infant cry vocalizations using artificial neural networks (ANNs)
Marco Petroni, Alfred S. Malowany, C. Celeste Johnston, et al.
The analysis of infant cry vocalizations has been the focus of a number of efforts over the past thirty years. Since the infant cry is one of the only means that an infant has for communicating with its care-giving environment, it is thought that information regarding the state of an infant, such as hunger or pain, can be determined from an infant's cry. To date, research groups have determined that adult listeners can differentiate between different types of cries auditorialy, and at least one group has attempted to automate this classification process. This paper presents the results of another attempt at automating the discrimination process, this time using artificial neural networks (ANNs). The input data consists of successive frames of one or two parametric representations generated from the first second of a cry following the application of either an anger, fear, or pain stimulus. From tests conducted to date, it is determined that ANNs are a useful tool for cry classification and merit further study in this domain.
Applications in Pattern Recognition III
icon_mobile_dropdown
Use of sensors to deal with uncertainty in realistic robotic environments
Enrique Cervera, Angel P. del Pobil, Edward Marta, et al.
This work is an application of Kohonen's self-organizing feature maps, to deal with uncertainty in realistic robotic environments. The neural network is fed with the signals of a force sensor attached to the wrist of the robot arm. The learning process consists of two phases: a training phase and a labeling phase. After training, the clusters of the map are associated to contact states. There are potential uncertainties in the positions of the elements, which cause error situations, like incorrect grasping of a piece or bad insertions. The error situations can be associated with clusters in the network. As a result of the application of this learning scheme, we can state that Kohonen's self-organizing feature maps are well suited for dealing with uncertainty in realistic robotic environments, particularly robot pick-and-place operations. They are easy to apply and powerful, and are a step towards the solution of more complex intelligent tasks.
Studies on using the higher order neural network for pattern recognition and optimization
Jinyan Li, Ying Lin Yu, Wangchao Li
The higher order neural network used for pattern recognition and optimization is studied in this paper and the results in two different aspects have been obtained. (1) Theoretically, the capacity formula of the Hopfield neural network with the second order weights has been obtained. Compared with the first order network, the capacity of the second order network is about three times greater than that of the first order one and the cost to reach such an efficiency is to add higher order weights. The simulated experiments according to the theory by digital computer are satisfied. (2) Theoretically, the method of how to solve the optimization problems, whose energy functions are more general than Lyapunov function, has been put forward at the end of this paper.
Artificial neural system with Lie germs for affine invariant pattern analysis
Thomas R. Tsao
A computational theory and neural architecture for affine invariant pattern analysis is presented. The orbit of an image pattern under the two dimensional affine transformation group is defined as an invariant pattern class. An analog neural dynamical system with Lie germs is able to compute the distance between the orbit of a given image pattern and a template. Any pattern that is affine reachable to the template (i.e., on the same affine orbit as the template) will have zero distance while others will have larger than zero distances. The key component of this neural system is a type of artificial neuron named Lie germs. Via their receptive fields, these neurons perform the function of the infinitesimal transforms of the affine Lie group on the Gabor representation domain. The responses of the Lie germs generate the vector field of the neural dynamical system tangent to the affine orbits and make the affine orbital motion of the neural dynamical system. A computer simulation of the artificial neural system is also presented.
Ortho-ordent initialization of feedforward artificial neural networks (FFANNs) to improve their generalization ability
There are several models for neurons and their interconnections. Among them, feedforward artificial neural networks (FFANNs) are very popular for being quite simple. However, to make them truly reliable and smart information processing systems, such characteristics as learning speed, local minima, and generalization ability need more understanding. Difficulties such as long learning-time and local minima, may not affect them as much as the question of generalization ability, because in many applications a network needs only one training, and then it may be used for a long time. However, the question of generalization ability of ANNs is of great interest for both theoretical understanding and practical use, because generalization ability is a measure of a learning system that indicates how closely its actual output approximates to the desired output for an input that it has never seen. We investigate novel techniques for systematic initializations (as opposed to purely random initializations) of FFANN architectures for possible improvement of their generalization ability. Our preliminary work has successfully employed row-vectors of Hadamard matrices to generate initializations; this initialization method has produced networks with better generalization ability.
Eliminating order dependency of classification in artificial resonance theory (ART1) networks
Astrid Leuba, Billy V. Koen
Incorrect classification of patterns can occur with ART1 networks when data are presented in certain sequences. The reason for this problem is the coding of the category templates, which are memory-less and give more importance to 1s than to 0s. This paper modifies the ART1 network architecture to alter these two features by adding a second set of top-down LTMs, in effect defining a second template. Computer simulations show that this modification ensures that patterns are always classified in the same category and that information is never lost. As a result, no pre-processing of the data is necessary, and ART1 networks can be used to classify patterns on-line without errors.
Learning IV
icon_mobile_dropdown
Solution of a linear system using a neural network based on iterative techniques
Heriberto Jose Delgado, Laurene V. Fausett
A neural network is presented for solving a general linear system using an approach that is similar to successive over relaxation (SOR). The network is trained to find the appropriate relaxation parameter. A derivation of the algorithm and its relation to the SOR algorithm is given. The performance of the standard SOR and Jacobi methods are compared with the neural network for two sample problems.
Jeffreys' prior for layered neural networks
Yoichi Motomura
In this paper, Jeffreys' prior for neural networks is discussed in the framework of the Bayesian statistics. For a good performance of generalization, the regularization methods which reduce both cost function and regularization term are commonly used. In the Bayesian statistics, the regularization term can be naturally derived from prior distribution of parameters. Jeffreys' prior is known as a typical non-informative objective prior. In the case of neural networks, however, it is not easy to express Jeffreys' prior as a simple function of parameters. In this paper, some numerical analysis of Jeffreys' prior for neural networks is given. The approximation of Jeffreys' prior is given from a parameter transformation getting to make Jeffreys' prior as a simple function. Some learning techniques are also discussed as applications of these results.
Fractal transform network in digital image content analysis
Rick Darby
The fractal transform is a recently developed tool for image analysis and pattern recognition. Problems relating to pixel based image resolution are highlighted. The fractal transform is outlined and shown to be a solution to pixel based image resolution related problems. MatchMaker, a technique based on fractal transform analysis is analyzed and its ability to find objects in complex scenes is confirmed. A new neural network paradigm, called the fractal transform network, is described. Built from neural network elements commonly described in the classic neural network literature, it is shown that this new paradigm can implement the MatchMaker technique and other fractal transform processes as highly parallel networks.
Applications in Pattern Recognition III
icon_mobile_dropdown
Multilayer Kohonen network and its separability analysis
Chao-yuan Liu, Jie-Gu Li
This paper presents a model of a multilayer Kohonen network. Because of obeying the winner- take-all learning rule and projecting high dimensional patterns into one or two dimensional space, the conventional Kohonen network has many limitations in its applications, such as pattern separability limitation and open ended limitation. Taking advantage of the innovation for learning method and its multilayer structure, the multilayer Kohonen network has the performance of nonlinear pattern partition. Owing to labeling pattern clusters with appropriate category names or numbers only, the network is an open ended system, so it is far more powerful than the conventional Kohonen network. The mechanism of the multilayer Kohonen network is explained in detail, and its nonlinear pattern separability is analyzed theoretically. As a result of an experiment made by two layer Kohonen network, a set of human head contour figures assigned into diverse by categories is shown.
Applications in Image Processing II
icon_mobile_dropdown
Effects of data representation and network architecture variation on multiaperture vision system performance
William R. Clayton, Ronald G. Driggers, Roy E. Williams, et al.
This research focuses on the effects of data representation and variations in neural network architecture on the tracking accuracy of a multi-aperture vision system (MAVS). A back- propagation neural network (BPNN) is used as a target location processor. Six different MAVS optical configurations are simulated in software. The system's responses to a point source target, in the form of detector voltages, and the known target location form a training record for the BPNN. Neural networks were trained for each of the optical configurations using different coordinate systems to represent the location of the point source target relative to the optical axis of the central eyelet. The number of processing elements in the network's hidden layer was also varied to determine the impact of these variations on the task of target location determination. A figure-of-merit (FOM) for the target location systems is developed to facilitate a direct comparison between the different optical and BPNN models. The results are useful in designing a MAVS tracker.
Face recognition using the Hausdorff-Voronoi Network (HAVNET) neural network
Usamah M.S. Altaf, Cihan H. Dagli
Some cognitive tasks that are easy for humans are not so for computer systems. Face recognition is one of these tasks. A face recognition prototype model using the HAVNET neural network is implemented and tested. The applications of such a model are tremendous and demanding. The prototype model uses a neural network that behaves as a binary pattern classifier. The neural network used, HAVNET, utilizes the Hausdorff distance as a metric of similarity between patterns and it employs a learned version of the Voronoi surface to perform the comparison. Different human faces' images are used for training and testing the model. The recognition results as well as the different sensitive factors that affect the recognition process are discussed.
Neural network applied to direction map extraction in fingerprint images
Flavio Augusto Pe Soares, Rui Seara, Orlando Jose Tobias
We present a new technique for extracting the direction map from fingerprints. The fingerprint image is first partitioned into small image blocks. Then, a set of parameters is extracted from each block and fed into a neural network that outputs the preferential direction for each block. The technique performed very well in operational conditions. It was developed to be employed in an Automatic Fingerprint Classification System.
Neural network based feature extraction scheme for heart rate variability
Ben Raymond, Doraisamy Nandagopal, Jagan Mazumdar, et al.
Neural networks are extensively used in solving a wide range of pattern recognition problems in signal processing. The accuracy of pattern recognition depends to a large extent on the quality of the features extracted from the signal. We present a neural network capable of extracting the autoregressive parameters of a cardiac signal known as hear rate variability (HRV). Frequency specific oscillations in the HRV signal represent heart rate regulatory activity and hence cardiovascular function. Continual monitoring and tracking of the HRV data over a period of time will provide valuable diagnostic information. We give an example of the network applied to a short HRV signal and demonstrate the tracking performance of the network with a single sinusoid embedded in white noise.
Neural network feature selection for breast cancer diagnosis
Catherine M. Kocur, Steven K. Rogers, Kenneth W. Bauer Jr., et al.
More than 50 million women over the age of 40 are currently at risk for breast cancer in the United States. Computer-aided diagnosis, as a second opinion to radiologists, will aid in decreasing the number of false readings of mammograms. Neural network benefits are exploited at both the classification and feature selection stages in the development of a computer-aided breast cancer diagnostic system. The multilayer perceptron is used to classify and contrast three features (angular second moment, eigenmasses, and wavelets) developed to distinguish benign from malignant lesion in a database of 94 difficult-to-diagnose digitized microcalcification cases. System performance of 74 percent correct classifications is achieved. Feature selection techniques are presented which further improve performance. Neural and decision boundary-based methods are implemented, compared, and validated to isolate and remove useless features. The contribution from this analysis is an increase to 88 percent correct classification in system performance. These feature selection techniques can also process risk factor data.
Optimal fusion of TV and infrared images using artificial neural networks
Thomas Fechner, Grzegorz Godlewski
This paper describes the application of a neural network for pixel-level fusion of visible (TV) and infrared (FLIR) images taken from the same scene. The goal of image fusion is to produce a single composite image in which information of interest from both input images is retained. Therefore, the applied fusion method should preserve those image details that are most relevant for human perception while suppressing noise. The proposed fusion method exploits the pattern recognition capabilities of artificial neural networks. Moreover, the learning capability of neural networks makes it feasible to customize the image fusion process. Some experimental results are presented and compared with existing image fusion methods.
Applications in Image Processing and ATR
icon_mobile_dropdown
Recognition of digits using spatiotemporal neural networks
Chong Sik Lee, Jae Ho Chung
In this paper, a new approach for Korean digit recognition using the Spatio-Temporal Neural Network (STNN) is reported. Two approaches are proposed, and the digit recognition rate of 95% is achieved. In the first approach, the LPC-cepstrums are used as STNN's input patterns. The LPC-cepstrums are derived from the linear predictive coding (LPC) coefficients that computed through the vocal tract analysis. The recognition rate of 90% is achieved, which is higher than the performance rate of 83.5% that is achieved by STNN with LPC coefficients as the input patterns. Using the LPC-cepstrums as the input patterns, in the second approach, when the difference between the highest two scores of ten STNNs' outputs is less than the predefined threshold value, the distortions of the two digit candidates from the input signal are computed using the Euclidean cepstral distance measure. Comparing the two distortions we then determine which STNN between the two produces smaller distortion, and the corresponding digit is declared as the recognized final digit. This simple added feature improves the performance of the STNN significantly from 90% to 95%.
Artificial neural networks for acoustic target recognition
James A. Robertson, John C. Mossing, Bruce A. Weber
Acoustic sensors can be used to detect, track and identify non-line-of-sight targets passively. Attempts to alter acoustic emissions often result in an undesirable performance degradation. This research project investigates the use of neural networks for differentiating between features extracted from the acoustic signatures of sources. Acoustic data were filtered and digitized using a commercially available analog-digital convertor. The digital data was transformed to the frequency domain for additional processing using the FFT. Narrowband peak detection algorithms were incorporated to select peaks above a user defined SNR. These peaks were then used to generate a set of robust features which relate specifically to target components in varying background conditions. The features were then used as input into a backpropagation neural network. A K-means unsupervised clustering algorithm was used to determine the natural clustering of the observations. Comparisons between a feature set consisting of the normalized amplitudes of the first 250 frequency bins of the power spectrum and a set of 11 harmonically related features were made. Initial results indicate that even though some different target types had a tendency to group in the same clusters, the neural network was able to differentiate the targets. Successful identification of acoustic sources under varying operational conditions with high confidence levels was achieved.
Neural network based moving target indicator for radar applications
Although techniques of radar signal processing, including the moving target indicator (MTI), have been vastly improved by the availability of digital computers in recent years, these methods are generally based on complex mathematical procedures which make the engineering and design of radar receivers rather costly and vulnerable to electronic faults. On the other hand, biological systems (e.g., insects, birds) have capabilities far beyond those of the conventional MTI processors. This paper provides some evidence that Doppler shifts can easily be extracted with neural networks even in situations where only a limited number of noisy pulses are available for processing. Furthermore, it is easier to shape the frequency response of the neural network-based MTI (NN-MTI) as desired without needing the complex process of pole placement, which is traditionally required in both digital and analog filter design procedures. The nonlinear processing capability of a neural network is utilized to efficiently combine the Doppler processing and integration performed by conventional MTI and its coprocessor (i.e., integrator). The MTI implementation with neural networks reduces the number of required independent pulses for Doppler shift extraction in the presence of clutter. There are several other conflicting requirements for the optimum MTI design where the algorithmic procedures may not be as efficient. Therefore, shaping the magnitude frequency response of MTI filters demands the flexibilities offered by neural networks.
Applications in Space Technology
icon_mobile_dropdown
Tracking, code recognition, and memory management with high-order neural networks
Clark D. Jeffries
A variety of tracking, recognition, and routing problems can be expressed as choices of rectangular 0,1 matrices. For some such problems of practical significance, neural network systems can give higher quality solutions than any other method. For example, in a tracker, how well predicted position i fits observed position j can generate through a metric function (possibly a backpropagation neural net), the initial ij entry in a large rectangular matrix. Using that matrix as an initial state, an associator dynamical system neural network can then choose a 0,1 matrix of the same dimensions with at most one 1 in each row and at most one 1 in each column. The association optimizes some overall goodness of fit criteria, not in general the same as pairwise goodness of fit. Related high order neural networks can be used to solve other finite choice problems including error-correction of binary codes and computer memory management. This paper is an overview of mathematical foundations of such dynamical systems.
Robust training of multilayer perceptrons: some experimental results
Andrew J. Myles, Alan F. Murray, Andrew M. Wallace
Robust regression methods are a useful alternative to least squares for modelling real data sets, which generally contain outliers. Various authors have previously examined the use of these methods for modelling such data using multilayer perceptrons (MLPs), with impressive results. This paper describes some experimental experiences with the use of some simple robust methods for MLP training. The use of robust error measures for testing is demonstrated, and the use of more than one robust error for testing MLPs during the training process is recommended. The failure of one simple robust training method due to leverage points is demonstrated, and preliminary results from a method which may assist in the identification of these points are provided. Finally, the problem of overfitting when using robust methods is discussed briefly.
Two spacecrafts' attitude determination using neural networks and image processing
Juan Seijas, Jose L. Sanz-Gonzalez
This paper presents a procedure for optimizing a neural network architecture, used in a system for spacecraft attitude and position determination. The procedure establishes the neural network structure and the training algorithm. A new version of basic-evolutive algorithm is presented, basic-evolutive 1 and basic-evolutive 2 algorithms are capable of setting the appropriate dimension of the neural network and the adequate weights interconnecting the neurons. The results produced by both versions are tested with a very wide set of different spacecrafts' maneuvers simulations. The algorithms performances are contrasted with backpropagation training algorithm performances. The capability of the resulting neural network architecture for generalizing is also verified.
Classification of rf transients in space using digital signal processing and neural network techniques
Kurt R. Moore, Phil C. Blain, Scott D. Briles, et al.
The FORTE' (fast on-orbit recording of transient events) small satellite experiment scheduled for launch in October 1995 will attempt to measure and classify electromagnetic transients as sensed from space. The FORTE' payload will employ an event classifier to perform onboard classification of radio frequency transients from terrestrial sources such as lightning. These transients are often dominated by a constantly changing assortment of man-made `clutter' such as TV, FM, and radar signals. The FORTE' event classifier, or EC, uses specialized hardware to implement various signal processing and neural network algorithms. The resulting system can process and classify digitized records of several thousand samples onboard the spacecraft at rates of about a second per record. In addition to reducing downlink rates, the EC minimizes command uplink data by normally using uploaded algorithm sequences rather than full code modules (although it is possible for full code modules to be uploaded from the ground). The FORTE' event classifier experiment combines science and engineering in an evolutionary step toward useful and robust adaptive processing systems in space.
Neural network approach to star-field recognition
Maximillian J. Domeika, Edward W. Page, Gene A. Tagliarini
Automatic recognition of stellar fields viewed by an imaging camera has numerous applications ranging from spacecraft navigation to pointing of spaceborne instruments. The usual approach to recognition is to develop an efficient algorithm for matching stars identified in the imager's field of view with star data recorded in an onboard catalog. Matching stars within a field of view with corresponding stars stored in a catalog requires finding a subset of the stars in the catalog that have positions and magnitudes that match those of the stars in the field of view. This paper presents a neural network approach to the problem of star field recognition. A Hopfield neural network is used to find a subset of the stars in the catalog that provides a good match to stars in the imager's field of view. The matching process employs a compatibility function, similar to a fuzzy membership function, to grade the similarity between stars in the field of view and those in the catalog.
General Aspects of Neural Networks
icon_mobile_dropdown
Computational problems in the linear perceptron: a possible solution--dynamic definition of net topology
With respect to Rosenblatt linear perceptron, two classical limitation theorems demonstrated by M. Minsky and S. Papert are discussed. These two theorems, `(Psi) One-in-a-box' and `(Psi) Parity,' ultimately concern the intrinsic limitations of parallel calculations in pattern recognition problems. We demonstrate a possible solution of these limitation problems by substituting the static definition of characteristic functions and of their domains in the `geometrical' perceptron, with their dynamic definition. This dynamic consists in the mutual redefinition of the characteristic function and of its domain depending on the matching with the input.
Prototype extraction in material attractor neural networks with stochastic dynamic learning
Stefano Fusi
Dynamic learning of random stimuli can be described as a random walk among the stable synaptic values. It is shown that prototype extraction can take place in material attractor neural networks when the stimuli are correlated and hierarchically organized. The network learns a set of attractors representing the prototypes in a completely unsupervised fashion and is able to modify its attractors when the input statistics change. Learning and forgetting rates are computed.
Particle searches with neural nets
Georg Stimpfl-Abele
The efficiency and robustness of neural feed-forward nets in particle searches is studied using the search for the standard Higgs Boson at LEP-200 as an example. Methods to select the most efficient variables, to define standard cuts, and to recognize significant differences between the training-data sample and a test-data sample are presented. The efficiencies of the neural nets are significantly better than those of standard methods.
Comparison of statistical methods and fuzzy systems for atmospheric pressure wave prediction
Francesco Masulli, Franco Casalino, Renato Caviglia, et al.
The prediction of complex phenomena, like the atmospheric system, represents a challenging application field for soft-computing methods. Neural networks and fuzzy systems allow one to obtain nonlinear, model-free regression methods that involve a large number of parameters and that are able to exploit much a-priori numerical knowledge (or training sets). The experiments described in this article point out that time series made up of daily average measures of atmospheric pressure and its waves are characterized by a positive first Lyapunov exponent, hence they are chaotic signals. Moreover, we compare the forecasting performances of two statistical methods, namely, the autoregressive moving average (ARMA) method and the linear predictor code (LPC) and of the adaptive fuzzy system (AFS). The AFS shows the higher prediction accuracy in each experiment, as compared with ARMA and LPC. In addition, for the AFS single waves are easier to predict than the global phenomenon, and the more accurate predictions are obtained for longer waves.
Neural Networks for Offline Analysis in High-Energy Physics I
icon_mobile_dropdown
Neural networks for offline analysis in high-energy physics
Alessandro D. de Angelis
Feed-forward neural networks are nowadays a standard tool in the toolbox of high energy physicists. This talk summarizes the fields of application in offline analysis, and discusses some open problems.
Tagging the s quark in the hadronic decays of the Z
G. Cosmo, Alessandro D. de Angelis
The selection of samples of events enriched in hadronic decays of the Z into ss pairs is made difficult by the poor knowledge of jet fragmentation. A feed-forward neural network can help in maximizing the performance of the classification. This offers the possibility to test the universality of the standard model of the electroweak interaction, and to perform measurements in the sector of nonperturbative QCD.
Standard Model Higgs boson search with neural networks
Klas Hultqvist, Richard Jacobsson, Erik Johansson, et al.
The mass window open for the standard model Higgs boson at LEP1 is at present restricted to a region where the production rate is very small. Moreover, Higgs particle events in this region are very difficult to separate from the background, which is why new analysis techniques are needed. We have employed a classifier based on a feed-forward neural network for the discrimination against the very large background. With a simple preselection followed by a neural network we have obtained a combined background rejection factor of about 29,000 and a detection efficiency of about 54% for a Higgs particle with a mass of 55 GeV/c2. With a different transformation of the input variables to the network, the detection efficiency was improved by a factor of 1.10, with the same background rejection.
Neural nets with varying topology for high-energy particle recognition: theory and applications
Antonio Luigi Perrone, Gianfranco Basti, Roberto Messi, et al.
In this paper we start from a critical analysis of the fundamental problems of the parallel calculus in linear structures and of their extension to the partial solutions obtained with non- linear architectures. Then, we briefly present a new dynamic architecture able to solve the limitations of the previous architectures through an automatic redefinition of the topology. This architecture is applied to real time recognition of particle tracks in high energy accelerators and in astrophysics experiments.
Neural Networks for Offline Analysis in High-Energy Physics II
icon_mobile_dropdown
Tau decay analysis with neural networks
F. Daniel Bertrand, Veronique Lefebure
Feed forward neural networks are used in the DELPHI experiment in order to identify exclusive decay channels of the (tau) lepton. A first step consists in determining the event topology in charged particles. Another network is then used in order to distinguish the corresponding charged leptons or hadrons as well as to establish the presence of neutral pions in the decay products. Migration matrices are computed to evaluate the mutual contaminations of various disintegration modes. The neural network method allows the user to reach higher efficiencies than the classic analysis using linear selection criteria.
Analysis of tau 1-prong hadronic inclusive branching ratio using neural networks
Francisco Matorras, Alberto A. Ruiz-Jimeno
A FFNN has been used to classify the 1-prong (tau) decays. The net is able to separate hadronic decays from leptonic with 90% efficiency and 93% purity. Applied to the data taken by the DELPHI detector at LEP collider during 1992, the (tau) inclusive 1-prong hadronic branching ratio has been measured to be B1h equals 0.5050 +/- 0.0032-0.0031+0.0046.
Identification of gluon jets with a neural network technique
Olof Barring, Torsten Akesson, Vincent Hedberg
The separation of jets originating from quarks and gluons in three-jet events at LEP by using a neural network program has been studied. The analysis has been done in two steps, one in which only the jet-energy is used for the identification and a second step in which jet fragmentation variables are used as well. A new training procedure of the neural network combined with a new normalization of the fragmentation variables gives a method of identifying gluon jets from their fragmentation properties without distorting the energy spectrum of the identified jets. Since the fragmentation process can only be studied with phenomenological models it is important that the identification procedure is to a large extent model independent and the study has been made with two different jet fragmentation models.
Use of neural networks to analyze Z approaches bb events in ALEPH
Pierre Henrard
Neural networks have been used in the ALEPH Collaboration at LEP to tag bb final states in Z decays, and to search for rare decays of B-hadrons. It is shown that the neural network approach, due to its nonlinearity, significantly improves the b tagging compared to linear multivariate analyses and to single-variable methods. For both analyses, the uncertainties due to the use of neural network are negligible and hence this technique is well adapted not only to select pure samples of bb events but also to achieve high precision measurements.
Pattern Recognition and Online Analysis
icon_mobile_dropdown
Neural nets in high-energy physics: applications in high-speed data acquisition
Bruce Denby
The paper provides an introduction to experimental methods in high energy physics (HEP) followed by a motivation for triggering applications of hardware neural networks in high speed data acquisition systems. A few examples of such applications are then treated in detail. The paper concludes with a survey of planned future applications of neural network triggers.
Signal/background classification in a cosmic ray space experiment by a modular neural system
Roberto Bellotti, Marcello Castellano, Carlo Nicola De Marzo, et al.
In the cosmic ray space experiments, the separation of the signal from background is a hard task. Due to the well-known critical conditions that characterize this class of experiments, some changes of the detector performances can be observed during the data taking. As a consequence, differences between the test and real data are found as systematic errors in the classification phase. In this paper, a modular classification system based on neural networks is proposed for the signal/background discrimination task in cosmic ray space experiments, without a priori knowledge of the discriminating feature distributions. The system is composed by two neural modules. The first one is a self organizing map (SOM) that both clusters the real data space in suitable classes of similarity and builds a prototype for each of them; a skilled inspection of the prototypes defines the signal and background. The second one, a multi layer perceptron (MLP) with a single hidden layer, adapts the classification model based on training/test data to the real experimental conditions. The MLP synaptic weights adaptive formation takes into account the labelled real data set as defined in the first system-phase. The modular neural system has been applied in the context of TRAMP-Si experiment, performed on the NASA Balloon-Borne Magnet Facility, for the positron/proton discrimination.
Feature extraction using self-organizing networks
Jochen Dahm, C. Voigt, K.-H. Becks, et al.
Self-organizing networks are used to extract kinematical features and to study correlations of high dimensional variable spaces in new physics areas. This method is applied in the search for heavy neutrinos at LEP200. We show that the extracted knowledge improves the distinction between heavy neutrino candidates and background.
Neural networks with statistical preprocessing for particle discrimination in high-energy physics
Enrico Pasqualucci
A typical task of an analysis program in high energy physics is the discrimination between different kinds of particles interacting with a detector. Neural networks can be easily used to perform this task. In this paper, the performances of a feed-forward neural network as a particle identifier are studied and compared with results from discriminant analysis. A typical task, the (pi) -(mu) separation at 250 MeV/c is presented as an example application. Experimental data collected during a test run of the KLOE electromagnetic calorimeter are used. The effects of the introduction of a statistical pre-processing on physical variables is studied. It allows us to obtain better results both in terms of learning time and in terms of efficiency and background rejection.
Robust tracking by cellular automata and neural networks with nonlocal weights
Gennadii A. Ososkov
A modified rotor model of the Hopfield neural networks (HNN) is proposed for finding tracks in multiwire proportional chambers. That requires us to apply both raw data prefiltering by cellular automaton and HNN weights furnishing by a special robust multiplier. Then this model is developed to be applicable for a more general type of data and detectors. As an example, data processing of ionospheric measurements are considered. For handling tracks detected by high pressure drift chambers with their up-down ambiguity a modification of deformable templates method is proposed. A new concept of controlled HNN is proposed for solving the so-called track-match problem.
Hardware Implementations of Neural Networks
icon_mobile_dropdown
Survey of neural network hardware
Clark S. Lindsey, Thomas Lindblad
We survey the currently available neural network hardware, including VLSI chips (digital, analog, and hybrid), PC accelerator cards, and multi-board neurocomputers. We concentrate on commercial hardware, but also include a few prototypes of special interest. As examples of applications, some systems developed for high energy physics experiments that use this hardware are presented.
LANN27: an electronic implementation of an analog attractor neural network with stochastic learning
Davide Badoni, Stefano Bertazzoni, Stefano Buglioni, et al.
We describe and discuss an electronic implementation of an attractor neural network with plastic synapses. The synaptic dynamics are unsupervised and autonomous, in that they are driven exclusively and perpetually by neural activities. The latter follow the network activity via the developing synapses and the influence of external stimuli. Such a network self- organizes and is a device which converts the gross statistical characteristics of the stimulus input stream into a set of attractors (reverberations). To maintain for a long time the acquired memory the analog synaptic efficacies are discretized by a stochastic refresh mechanism. The discretized synaptic memory has indefinitely long life time in the absence of activity in the network. It is modified only by the arrival of new stimuli. The stochastic refresh mechanism produces transitions at low probability which ensures that transient stimuli do not create significant modifications and that the system has large palimpsestic memory. The electronic implementation is completely analog, stochastic and asynchronous. The circuitry of the first prototype is discussed in some detail as well as the tests performed on it. In carrying out the implementation we have been guided by biological considerations and by electronic constraints.
Applications in Control and Physics
icon_mobile_dropdown
Manipulator trajectory control using neural networks: from application to theory and back again
Yichuang Jin, A. G. Pipe, A. Winfield
In this paper we first briefly review neural networks and some previous results of their applications on manipulator trajectory control. Then we go into the main part of the paper, i.e., theoretical analysis of neuro-manipulator control systems. It consists of control structure designs, off-line/on-line learning algorithms, and system stability proof. Two control structures are presented, both having a stability guarantee. Simulation on a Puma 560 robot and an experiment on a Mentor robot are also presented to demonstrate how to use the theoretical results and to evaluate performance of the developed control structures.
Learning IV
icon_mobile_dropdown
United Kingdom's industrial online neurocomputing system: Department of Trade and Industry (DTI) Neurocomputing Web
Gary Whittington, Robert Wiggins
The Department of Trade and Industry (DTI) established in February 1993 a 3 year, pounds sterling 5.7 M neural computing technology transfer (NCTT) program to promote awareness and exploitation of neurocomputing in the United Kingdom (UK). This paper describes an on-line neurocomputing resource for UK industry that has been established as part of this program: the `DTI Neurocomputing Web.' This resource has initially been implemented as an Internet- based world wide web service.
Applications in Pattern Recognition I
icon_mobile_dropdown
Prestructuring ANNs via a priori knowledge
George G. Lendaris, Karl Mathia
The work to be described has as its objective development of a constructive method that uses certain a priori information about a problem domain to design the starting structure of an artificial neural network (ANN). For the prestructuring process, there is motivation to move away from homogeneous structures to ones that comprise modules of smaller ANNs. Issues of concern include physical realizability, scalability of training time with large numbers of connections, and successful generalization. The method explored is based on a general systems theory methodology (here called GSM) that calculates a kinds of structural information of the problem domain via analyzing I/O pairs from that domain. This GSM-based information is used for developing a modularized ANN starting structure. Extensive experiments on 3-input, 1-output Boolean mappings verify our predictions. In addition, the experiments indicate that the GSM-based modularized-ANN design is `conservative' in the sense that the PS of the modularized ANN contains at least all the mappings included in the GSM category used to design the ANN. Experiments with 5-input, 1-output Boolean functions provide further support of the conclusions.