| Nome: | Descrição: | Tamanho: | Formato: | |
|---|---|---|---|---|
| 352.95 KB | Adobe PDF |
Orientador(es)
Resumo(s)
Kolmogorov-Arnold Networks (KANs) represent a breakthrough in deep learning, diverging from Multi-Layer Perceptrons (MLPs) by generalizing the Kolmogorov-Arnold representation theorem (KAT) to networks of arbitrary depth and width. This theorem facilitates the decomposition of multivariate functions into constituent one-dimensional elements, with learnable activation functions on weights and the sum operator on nodes. KANs have been shown to exhibit robust performance in function approximation, validated across mathematical, physical, and practical domains such as traffic prediction and medical diagnostics. Our study investigates KANs’ efficacy through comprehensive evaluations on OpenML, Kaggle and UCI datasets, with a focus on enhancing Human Activity Recognition systems. They demonstrate high classification performance compared to conventional machine learning approaches and MLPs. These findings underscore KANs’ potential as scalable, interpretable tools in modern machine learning applications given their favorable neural scaling laws.
Descrição
Publisher Copyright:
© 2024 Copyright held by the owner/author(s).
Palavras-chave
Classification Human Activity Recognition Kernel function Kolmogorov-Arnold Networks Multivariate functions Artificial Intelligence
Contexto Educativo
Citação
Editora
ACM - Association for Computing Machinery
