Are the "efficient-kan" and "official-kan" equivalent in terms of algorithms? #35
Replies: 14 comments 1 reply
-
As I know almost the same, only official version looks to have additional bias after each layer. Also, I am not sure if initialization is the same. + regularization loss is changed because of optimizations. |
Beta Was this translation helpful? Give feedback.
-
@Indoxer Thanks, you are so kindly. |
Beta Was this translation helpful? Give feedback.
-
No, I'm not quite sure *Including the use of the official LBFGS training strategy |
Beta Was this translation helpful? Give feedback.
-
I think this is acceptable, after all, the model is very efficient, and some losses are normal. It's strange if there are no losses at all. While it effectively retains the characteristics of the official model, it also combines training optimization |
Beta Was this translation helpful? Give feedback.
-
@WhatMelonGua, are you sure, that you didn't train spline_scaler and base_weights? Also did you have the same parameters in LBFGS optimizer (number of steps, etc.)? |
Beta Was this translation helpful? Give feedback.
-
(spline_scaler not trained, base_weights not trained) (I am using my modified version (but the same algorithm as efficient kan), so I am not sure) |
Beta Was this translation helpful? Give feedback.
-
Oh, yes, forgive me for forgetting |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
reg_ is regularization loss. |
Beta Was this translation helpful? Give feedback.
-
Here are my results and code, so you can compare |
Beta Was this translation helpful? Give feedback.
-
AFAIK the only difference is that the "efficient" regularization loss is different from the official one. But I'm not sure if the parallel associativity will introduce numerical error that's large enough to break some important features. |
Beta Was this translation helpful? Give feedback.
-
Just found that I missed the bias term after each layer. Will update that soon. I scanned over this long thread few days ago and totally missed the comment by @Indoxer lol |
Beta Was this translation helpful? Give feedback.
-
Hi is there any plan to support update_grid_from_samples and initialize_from_another_model from the original KANs. A lot of use cases work much better when we use these APIs so I think its critical for KANs |
Beta Was this translation helpful? Give feedback.
-
Its seems like the algos are somewhat equivalent. As someone who is much more interested in the symbolic representation aspect, is there enough consistency between the original KAN implementation and what's here to do formula approximation? |
Beta Was this translation helpful? Give feedback.
-
as title
Beta Was this translation helpful? Give feedback.
All reactions