http://scholars.ntou.edu.tw/handle/123456789/17026
Title: | Scale Shrinking Transformation and Applications | Authors: | Yu-Chen Chen Keng-Hsuan Wu Jyun-Ting Lai Jung-Hua Wang |
Keywords: | neural networks;Shrinking Transformation;Function Approximation | Issue Date: | 20-Sep-2006 | Conference: | 3rd International Conference on Soft Computing and Intelligent Systems and 7th International Symposium on advanced Intelligent Systems SCIS&ISIS2006 Tokyo, Japan |
Abstract: | This paper presents a novel information processing technique called scale shrinking transformation (SST). SST comprises three steps: initialization, matrix transformation, and using the column vectors of the transformed matrix as the new input vectors. The essence of SST is that the structural correlation between original inputs can be obtained. More significantly, the transformed matrix contains elements with much smaller scale variation. When applied to existing feedforward neural networks, it can alleviate problems commonly encountered in tasks of function approximation, separating nonlinearly classes, and noise filtering. When the column vectors are used as the new input to a feedforward network that comprises hidden layers, training speed can be reduced. The input scale divergence problem that plagues higher-order neural networks can also be alleviated with SST. |
URI: | https://www.jstage.jst.go.jp/article/softscis/2006/0/2006_0_698/_article/-char/ja/ http://scholars.ntou.edu.tw/handle/123456789/17026 |
DOI: | 10.14864/softscis.2006.0.698.0 |
Appears in Collections: | 電機工程學系 |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.