Regularized One-Layer Neural Networks for Distributed and Incremental Environments
| UDC.coleccion | Investigación | |
| UDC.conferenceTitle | IWANN 2021, 16th International Work-Conference on Artificial Neural Networks | |
| UDC.departamento | Ciencias da Computación e Tecnoloxías da Información | |
| UDC.endPage | 355 | |
| UDC.grupoInv | Laboratorio de Investigación e Desenvolvemento en Intelixencia Artificial (LIDIA) | |
| UDC.institutoCentro | CITIC - Centro de Investigación de Tecnoloxías da Información e da Comunicación | |
| UDC.startPage | 343 | |
| UDC.volume | LNTCS, volume 12862 | |
| dc.contributor.author | Fontenla-Romero, Óscar | |
| dc.contributor.author | Guijarro-Berdiñas, Bertha | |
| dc.contributor.author | Pérez-Sánchez, Beatriz | |
| dc.date.accessioned | 2026-01-14T19:49:32Z | |
| dc.date.available | 2026-01-14T19:49:32Z | |
| dc.date.issued | 2021-08 | |
| dc.description | This version of the conference paper has been accepted for publication, after peer review, but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: https://doi.org/10.1007/978-3-030-85099-9_28 . Presented at: 16th International Work-Conference on Artificial Neural Networks, IWANN 2021, Virtual Event, June 16–18, 2021. | |
| dc.description.abstract | [Abstract]: Deploying machine learning models at scale is still a major challenge; one reason is that performance degrades when they are put into production. It is therefore very important to ensure the maximum possible generalization capacity of the models and regularization plays a key role in avoiding overfitting. We describe Regularized One-Layer Artificial Neural Network (ROLANN), a novel regularized training method for one-layer neural networks. Despite its simplicity, this network model has several advantages: it is noniterative, has low complexity, and is capable of incremental and privacy-preserving distributed learning, while maintaining or improving accuracy over other state-of-the-art methods as demonstrated by the experimental study in which it has been compared with ridge regression, lasso and elastic net over several data sets. | |
| dc.description.sponsorship | This work has been supported by grant Machine Learning on the Edge (Ayudas Fundaci ´on BBVA a Equipos de Investigaci´on Cient´ıfica 2019), also by the National Plan for Scientific and Technical R&I of the Spanish Government (Grant PID2019-109238GBC2), and by the Xunta de Galicia (Grant ED431C 2018/34) with the European Union ERDF funds. CITIC is partially funded by “Conseller´ıa de Cultura, Educaci´on e Universidades from Xunta de Galicia” (Grant ED431G 2019/01). | |
| dc.description.sponsorship | Xunta de Galicia; ED431C 2018/34 | |
| dc.description.sponsorship | Xunta de Galicia; ED431G 2019/01 | |
| dc.identifier.citation | Fontenla-Romero, O., Guijarro-Berdiñas, B., Pérez-Sánchez, B. (2021). Regularized One-Layer Neural Networks for Distributed and Incremental Environments. In: Rojas, I., Joya, G., Català, A. (eds) Advances in Computational Intelligence. IWANN 2021. Lecture Notes in Computer Science(), vol 12862. Springer, Cham. https://doi.org/10.1007/978-3-030-85099-9_28 | |
| dc.identifier.doi | 10.1007/978-3-030-85099-9_28 | |
| dc.identifier.isbn | 978-3-030-85099-9 | |
| dc.identifier.isbn | 978-3-030-85098-2 | |
| dc.identifier.issn | 1611-3349 | |
| dc.identifier.uri | https://hdl.handle.net/2183/46869 | |
| dc.language.iso | eng | |
| dc.publisher | Springer Nature Switzerland AG | |
| dc.relation.projectID | info:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/PID2019-109238GB-C22/ES/APRENDIZAJE AUTOMATICO ESCALABLE Y EXPLICABLE/ | |
| dc.relation.uri | https://doi.org/10.1007/978-3-030-85099-9_28 | |
| dc.rights | ©2021 Springer Nature Switzerland AG. This version of the conference paper is subject to Springer Nature’s AM terms of use. | |
| dc.rights.accessRights | open access | |
| dc.subject | Regularization | |
| dc.subject | Big data | |
| dc.subject | Incremental learning | |
| dc.subject | Distributed learning | |
| dc.subject | Privacy-preserving | |
| dc.subject | Singular value decomposition | |
| dc.title | Regularized One-Layer Neural Networks for Distributed and Incremental Environments | |
| dc.type | conference output | |
| dspace.entity.type | Publication | |
| relation.isAuthorOfPublication | 3eef0200-4ae7-4fc8-9ffe-2e7928ffd1cd | |
| relation.isAuthorOfPublication | d839396d-454e-4ccd-9322-d3e89a876865 | |
| relation.isAuthorOfPublication | 1729347a-a5bc-4ab0-a914-6c7a1dce7eb9 | |
| relation.isAuthorOfPublication.latestForDiscovery | 3eef0200-4ae7-4fc8-9ffe-2e7928ffd1cd |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Fontenla_Romero_Oscar_2021_Regularized_One_Layer_Neural_Networks.pdf
- Size:
- 445.85 KB
- Format:
- Adobe Portable Document Format

