A What-and-Where Fusion Neural Network for Recognition and Tracking of Multiple Radar Emitters

OpenBU

Show simple item record

dc.contributor.author Granger, Eric en_US
dc.contributor.author Rubin, Mark en_US
dc.contributor.author Grossberg, Stephen en_US
dc.contributor.author Lavoie, Pierre en_US
dc.date.accessioned 2011-11-14T19:02:49Z
dc.date.available 2011-11-14T19:02:49Z
dc.date.issued 2000-09 en_US
dc.identifier.uri http://hdl.handle.net/2144/2276
dc.description.abstract A neural network recognition and tracking system is proposed for classification of radar pulses in autonomous Electronic Support Measure systems. Radar type information is combined with position-specific information from active emitters in a scene. Type-specific parameters of the input pulse stream are fed to a neural network classifier trained on samples of data collected in the field. Meanwhile, a clustering algorithm is used to separate pulses from different emitters according to position-specific parameters of the input pulse stream. Classifier responses corresponding to different emitters are separated into tracks, or trajectories, one per active emitter, allowing for more accurate identification of radar types based on multiple views of emitter data along each emitter trajectory. Such a What-and-Where fusion strategy is motivated by a similar subdivision of labor in the brain. The fuzzy ARTMAP neural network is used to classify streams of pulses according to radar type using their functional parameters. Simulation results obtained with a radar pulse data set indicate that fuzzy AIUMAP compares favorably to several other approaches when performance is measured in terms of accuracy and computational complexity. Incorporation into fuzzy ARTMAP of negative match tracking (from ARTMAP-IC) facilitated convergence during training with this data set. Other modifications improved classification of data that include missing input pattern components and missing training classes. Fuzzy ARTMAP was combined with a bank of Kalman filters to group pulses transmitted from different emitters based on their position-specific parameters, and with a module to accumulate evidence from fuzzy ARTMAP responses corresponding to the track defined for each emitter. Simulation results demonstrate that the system provides a high level of performance on complex, incomplete and overlapping radar data. en_US
dc.description.sponsorship Defense Advanced Research Projects Agency and the Office of Naval Research (N00014-95-1-0409) (S. G. and M. A. R.); National Science Foundation (IRI-97-20333) (S. G.); Natural Sciences and Engineering Research Council of Canada (E. G.); Office of Naval Research (N00014-95-l-0657) (S.G.). en_US
dc.language.iso en_US en_US
dc.publisher Boston University Center for Adaptive Systems and Department of Cognitive and Neural Systems en_US
dc.relation.ispartofseries BU CAS/CNS Technical Reports;CAS/CNS-TR-2000-029 en_US
dc.rights Copyright 2000 Boston University. Permission to copy without fee all or part of this material is granted provided that: 1. The copies are not made or distributed for direct commercial advantage; 2. the report title, author, document number, and release date appear, and notice is given that copying is by permission of BOSTON UNIVERSITY TRUSTEES. To copy otherwise, or to republish, requires a fee and / or special permission. en_US
dc.subject Radar en_US
dc.subject Electronic support measures en_US
dc.subject Pattern recognition en_US
dc.subject Data fusion en_US
dc.subject Neural network en_US
dc.subject ARTMAP en_US
dc.subject Kalman filter en_US
dc.title A What-and-Where Fusion Neural Network for Recognition and Tracking of Multiple Radar Emitters en_US
dc.type Technical Report en_US
dc.rights.holder Boston University Trustees en_US

Files in this item

This item appears in the following Collection(s)

Show simple item record

Search OpenBU


Advanced Search

Browse

Deposit Materials