Takane, Y., Oshima-Takane, Y., & Shultz, T. R. (1994). Methods for analyzing internal representations of neural networks. In T. Kubo (Ed.), Proceedings of the 22nd Annual Meeting of the Behaviormetric Society (pp. 246-247). Tokyo: The Behaviormetric Society.
Neural network models have recently been very popular in artificial intelligence, cognitive psychology, pattern recognition, etc. They appear to work remarkably well even for problems for which conventional statistical methods typically fail. However, how they achieve what they achieve is not understood sufficiently well. Feed-forward networks may be viewed as approximating (nonlinear) functions that connects inputs to outputs. We explore methods to analyze how the approximations are done. These methods range from a simple graphing technique to two-way and three-way constrained and unconstrained principal component analyses.
Abstracts, papers, chapters, and other documents are posted on this site as an efficient way to distribute reprints. The respective authors and publishers of these works retain all of the copyrights to this material. Anyone copying, downloading, bookmarking, or printing any of these materials agrees to comply with all of the copyright terms. Other than having an electronic or printed copy for fair personal use, none of these works may be reposted, reprinted, or redistributed without the explicit permission of the relevant copyright holders.
To obtain a PDF reprint of this particular article, signal your agreement with these copyright terms by clicking on the statement below.