From the GitHub repository, Kolmogorov-Arnold Networks (KANs) are emerging as compelling alternatives to Multi-Layer Perceptrons (MLPs). While MLPs are grounded in the universal approximation theorem, KANs are supported by the Kolmogorov-Arnold representation theorem. The key difference between the two lies in their structure: KANs apply activation functions to edges, whereas MLPs apply them to nodes. This structural variation enables KANs to outperform MLPs in terms of both accuracy and interpretability.
This is the implementation of KAN on Iris dataset. The Iris dataset consists of 150 examples. Train and test sets were split as the following:
- Training set: 90%
- Test set: 10%
- Train set accuracy: 96.3%
- Test set accuracy: 93.3%
Note: Adjusting hyperparameters didn't really change anything. Hyperparameter adjustment might affect algorithm's performance when dealing with bigger datasets