콤팩트한 구조를 위한 I-ELM의 개선 = Enhancing I-ELM for More Compact Architecture

서성효 2022년
논문상세정보
' 콤팩트한 구조를 위한 I-ELM의 개선 = Enhancing I-ELM for More Compact Architecture' 의 주제별 논문영향력
논문영향력 선정 방법
논문영향력 요약
주제
  • Compact Architecture
  • I-ELM
  • elm
  • 신경망
동일주제 총논문수 논문피인용 총횟수 주제별 논문영향력의 평균
148 0

0.0%

' 콤팩트한 구조를 위한 I-ELM의 개선 = Enhancing I-ELM for More Compact Architecture' 의 참고문헌

  • Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions
    Guang-Bin Huang and Haroon A Babri . 9 ( 1 ) :224 ? 229 [1998]
  • Universal approximation with convex optimization : Gimmick or reality ? [ discussion forum ]
    Jose C Principe and Badong Chen 10 ( 2 ) :68 ? 77 [2015]
  • Universal approximation using radial-basisfunction networks
    Jooyoung Park and Irwin W Sandberg 3 ( 2 ) :246 ? 257 [1991]
  • Universal approximation bounds for superpositions of a sigmoidal function .
    Andrew R Barron 39 ( 3 ) :930 ? 945 [1993]
  • The perceptron : a probabilistic model for information storage and organization in the brain
    Frank Rosenblatt 65 ( 6 ) :386 , [1958]
  • Stochastic choice of basis functions in adaptive function approximation and the functional-link net
    Boris Igelnik and Yoh-Han Pao 6 ( 6 ) :1320 ? 1329 [1995]
  • Smooth function approximation using neural networks
    Silvia Ferrari and Robert F Stengel 16 ( 1 ) :24 ? 38 [2005]
  • Self-adaptive evolutionary extreme learning machine . Neural processing letters
  • Recent advances in finding best operating conditions
    Richard Loree Anderson 48 ( 264 ) :789 ? 798 [1953]
  • Orthogonal incremental extreme learning machine for regression and multiclass classification . Neural computing and applications
    Li Ying 27 ( 1 ) :111 ? 120 [2016]
  • On the capabilities of multilayer perceptrons .
    Eric B Baum 4 ( 3 ) :193 ? 215 [1988]
  • Generating random weights and biases in feedforward neural networks with random hidden nodes .
    Grzegorz Dudek 56 [2019]
  • Further contributions to the theory of generalized inverse of matrices and its applications
  • Fault diagnosis of tennessee-eastman process using orthogonal incremental extreme learning machine based on driving amount
    Weidong Zou , Yuanqing Xia , and Huifang Li . 48 ( 12 ) :3403 ? 3410 [2018]
  • Differential evolution ? a simple and efficient heuristic for global optimization over continuous spaces
    Rainer Storn and Kenneth Price . 11 ( 4 ) :341 ? 359 [1997]
  • Convolutional neural network based on an extreme learning machine for image classification
  • Convex incremental extreme learning machine . Neurocomputing
    Guang-Bin Huang and Lei Chen 70 ( 16-18 ) :3056 ? 3062 [2007]
  • Constructive hidden nodes selection of extreme learning machine for regression
    Yuan Lan , Yeng Chai Soh , and Guang-Bin Huang 73 ( 16-18 ) :3191 ? 3199 [2010]
  • Combination of activation functions in extreme learning machines for multivariate calibration
  • Capabilities of a four-layered feedforward neural network : four layers versus three
  • Approximation capabilities of multilayer feedforward networks . Neural networks
    Kurt Hornik . 4 ( 2 ) :251 ? 257 [1991]
  • An improved evolutionary extreme learning machine based on particle swarm optimization
  • Adaptive radial basis function nonlinearities , and the problem of generalisationpages 171 ? 175
    David Lowe No . 313 [1989]
  • A weight initialization method for improving training speed in feedforward neural network
    Jim YF Yam and Tommy WS Chow . 30 ( 1-4 ) :219 ? 232 [2000]
  • A novel orthogonal extreme learning machine for regression and classification problems
  • A new method in determining initial weights of feedforward neural networks for training enhancement
  • A logical calculus of the ideas immanent in neurons activity
    Warren S McClulloch and Walter Pitts . 5 ( 115-133 ) :10 , [1943]
  • 8. Moshe Leshno, Vladimir Ya Lin, Allan Pinkus, and Shimon Schocken. Multilayer feedforward networks with a nonpolynomial activation function can approximate any function. Neural networks, 6(6):861–867, 1993.
    [1993]
  • 64. Jiexiong Tang, Chenwei Deng, and Guang-Bin Huang. Extreme learning machine for multilayer perceptron. IEEE transactions on neural networks and learning systems, 27(4):809–821, 2015.
    [2015]
  • 63. LA Rastrigin. The convergence of the random search method in the extremal control of a many parameter system. Automaton & Remote Control, 24:1337–1342, 1963.
    [1963]
  • 61. Yong-Ping Zhao, Zhi-Qiang Li, Peng-Peng Xi, Dong Liang, Liguo Sun, and Ting- Hao Chen. Gram–schmidt process based incremental extreme learning machine. Neurocomputing, 241:1–17, 2017.
    [2017]
  • 6. Geoffrey E Hinton, Simon Osindero, and Yee-Whye Teh. A fast learning algorithm for deep belief nets. Neural computation, 18(7):1527–1554, 2006.
    [2006]
  • 59. Yong Peng, Wanzeng Kong, and Bing Yang. Orthogonal extreme learning machine for image classification. Neurocomputing, 266:458–464, 2017.
    [2017]
  • 56. Yat-Fung Yam and Tommy WS Chow. Determining initial weights of feedforward neural networks based on least squares method. Neural Processing Letters, 2(2):13– 17, 1995.
    [1995]
  • 55. Denis Serre. Matrices: Theory and applications. 2002. Graduate texts in mathematics, 2000.
  • 50. Morteza Mohammadzaheri, Lei Chen, Ali Ghaffari, and John Willison. A combination of linear and nonlinear activation functions in neural networks for modeling a de-superheater. Simulation Modelling Practice and Theory, 17(2):398–407, 2009.
    [2009]
  • 5. Yann LeCun, L´eon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278– 2324, 1998.
    [1998]
  • 47. Syukron Abu Ishaq Alfarozi, Noor Akhmad Setiawan, Teguh Bharata Adji, Kuntpong Woraratpanya, Kitsuchart Pasupa, and Masanori Sugimoto. Analytical incremental learning: Fast constructive learning method for neural network. In International Conference on Neural Information Processing, pages 259–268. Springer, 2016.
    [2016]
  • 46. Fei Han, Min-Ru Zhao, Jian-Ming Zhang, and Qing-Hua Ling. An improved incremental constructive single-hidden-layer feedforward networks for extreme learning machine based on particle swarm optimization. Neurocomputing, 228:133–142, 2017.
    [2017]
  • 44. You Xu and Yang Shu. Evolutionary extreme learning machine–based on particle swarm optimization. In International Symposium on Neural Networks, pages 644– 652. Springer, 2006.
    [2006]
  • 43. Tiago Matias, Francisco Souza, Rui Ara´ujo, and Carlos Henggeler Antunes. Learning of a single-hidden layer feedforward neural network using an optimized extreme learning machine. Neurocomputing, 129:428–436, 2014.
    [2014]
  • 42. A Kai Qin, Vicky Ling Huang, and Ponnuthurai N Suganthan. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE transactions on Evolutionary Computation, 13(2):398–417, 2008.
    [2008]
  • 4. Yann LeCun, Bernhard Boser, John Denker, Donnie Henderson, Richard Howard, Wayne Hubbard, and Lawrence Jackel. Handwritten digit recognition with a backpropagation network. Advances in neural information processing systems, 2, 1989.
    [1989]
  • 38. Guorui Feng, Guang-Bin Huang, Qingping Lin, and Robert Gay. Error minimized extreme learning machine with growth of hidden nodes and incremental learning. IEEE Transactions on Neural Networks, 20(8):1352–1357, 2009.
    [2009]
  • 37. Yoan Miche, Mark Van Heeswijk, Patrick Bas, Olli Simula, and Amaury Lendasse. Trop-elm: a double-regularized elm using lars and tikhonov regularization. Neurocomputing, 74(16):2413–2421, 2011.
    [2011]
  • 36. Yoan Miche, Antti Sorjamaa, Patrick Bas, Olli Simula, Christian Jutten, and Amaury Lendasse. Op-elm: optimally pruned extreme learning machine. IEEE transactions on neural networks, 21(1):158–162, 2009.
    [2009]
  • 35. Yoan Miche, Patrick Bas, Christian Jutten, Olli Simula, and Amaury Lendasse. A methodology for building regression models using extreme learning machine: Op-elm. In ESANN, pages 247–252, 2008.
    [2008]
  • 34. Guang-Bin Huang and Lei Chen. Enhanced random search based incremental extreme learning machine. Neurocomputing, 71(16-18):3460–3468, 2008.
    [2008]
  • 32. Hai-Jun Rong, Yew-Soon Ong, Ah-Hwee Tan, and Zexuan Zhu. A fast prunedextreme learning machine for classification problem. Neurocomputing, 72(1-3):359– 366, 2008.
    [2008]
  • 31. Qin-Yu Zhu, A Kai Qin, Ponnuthurai N Suganthan, and Guang-Bin Huang. Evolutionary extreme learning machine. Pattern recognition, 38(10):1759–1763, 2005.
    [2005]
  • 30. Guang-Bin Huang, Lei Chen, and Chee Kheong Siew. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Networks, 17(4):879–892, 2006.
    [2006]
  • 3. David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 323(6088):533–536, 1986.
    [1986]
  • 28. Chen Pan, Dong Sun Park, Yong Yang, and Hyouck Min Yoo. Leukocyte image segmentation by visual attention and extreme learning machine. Neural Computing and Applications, 21(6):1217–1227, 2012.
    [2012]
  • 27. Abdul Adeel Mohammed, Rashid Minhas, QM Jonathan Wu, and Maher A Sid- Ahmed. Human face recognition based on multidimensional pca and extreme learning machine. Pattern Recognition, 44(10-11):2588–2597, 2011.
    [2011]
  • 26. Rashid Minhas, Aryaz Baradarani, Sepideh Seifzadeh, and QM Jonathan Wu. Human action recognition using extreme learning machine based on visual vocabularies. Neurocomputing, 73(10-12):1906–1917, 2010.
    [2010]
  • 25. Hongming Zhou, Guang-Bin Huang, Zhiping Lin, Han Wang, and Yeng Chai Soh. Stacked extreme learning machines. IEEE transactions on cybernetics, 45(9):2013– 2025, 2014.
    [2014]
  • 24. C-WT Yeu, Meng-Hiot Lim, Guang-Bin Huang, Amit Agarwal, and Yew-Soon Ong. A new machine learning paradigm for terrain reconstruction. IEEE Geoscience and Remote Sensing Letters, 3(3):382–386, 2006.
    [2006]
  • 23. Jian-Xin Xu,WeiWang, JCH Goh, and Grace Lee. Internal model approach for gait modeling and classification. In 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, pages 7688–7691. IEEE, 2006.
    [2006]
  • 22. Stephanus Daniel Handoko, Kwoh Chee Keong, Ong Yew Soon, Guang Lan Zhang, and Vladimir Brusic. Extreme learning machine for predicting hla-peptide binding. In International Symposium on Neural Networks, pages 716–721. Springer, 2006.
    [2006]
  • 21. Nan-Ying Liang, Paramasivan Saratchandran, Guang-Bin Huang, and Narasimhan Sundararajan. Classification of mental tasks from eeg signals using extreme learning machine. International journal of neural systems, 16(01):29–38, 2006.
    [2006]
  • 14. Simon Haykin. Neural networks and learning machines, 2009.
    [2009]
  • 13. Risto Miikkulainen. Topology of a neural network., 2010.
    [2010]
  • 12. David Hunter, Hao Yu, Michael S Pukish III, Janusz Kolbusz, and Bogdan M Wilamowski. Selection of proper neural network sizes and architectures—a comparative study. IEEE Transactions on Industrial Informatics, 8(2):228–240, 2012.
    [2012]
  • 11. Guang-Bin Huang, Qin-Yu Zhu, and Chee-Kheong Siew. Extreme learning machine: theory and applications. Neurocomputing, 70(1-3):489–501, 2006.
    [2006]