Abstract
Extreme learning machine (ELM) is a machine learning technique based on competitive single-hidden layer feedforward neural network (SLFN). However, traditional ELM and its variants are only based on random assignment of hidden weights using a uniform distribution, and then the calculation of the weights output using the least-squares method. This paper proposes a new architecture based on a non-linear layer in parallel by another non-linear layer and with entries of independent weights. We explore the use of a deterministic assignment of the hidden weight values using low-discrepancy sequences (LDSs). The simulations are performed with Halton and Sobol sequences. The results for regression and classification problems confirm the advantages of using the proposed method called PL-ELM algorithm with the deterministic assignment of hidden weights. Moreover, the PL-ELM algorithm with the deterministic generation using LDSs can be extended to other modified ELM algorithms.
Original language | English |
---|---|
Pages (from-to) | 109-116 |
Number of pages | 8 |
Journal | Neurocomputing |
Volume | 226 |
DOIs | |
State | Published - 22 Feb 2017 |
Keywords
- Classification
- Extreme learning machine
- Low-discrepancy points
- Parallel layers
- Regression