Difficult-Race-1188
Difficult-Race-1188 OP t1_iylz2aw wrote
Reply to comment by druffischnuffi in [D] Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
What people mean when they say AI is not truly learning is that often the most impressive results are coming from extremely big models. For example, almost all the top AI scientist takes dig on Large language models, because we don't know whether they learned something or it just memorized all the possible combinations. Why people believe AI is not truly learning is that there are papers that show that AI was unable to generalize to simple mathematical equations.
x³ + xy² + y (mod 97), AI was unable to generalize to this simple equation.
https://medium.com/aiguys/paper-review-grokking-generalization-and-over-fitting-9dbbec1055ae
Difficult-Race-1188 OP t1_iylv75t wrote
Reply to comment by Difficult-Race-1188 in [D] Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
Because VC dimension of kernel SVM is infinite and thus they can create extremely curved boundaries in lower hyperdimension.
Difficult-Race-1188 OP t1_iylv1fs wrote
Reply to comment by Deep-Station-1746 in [D] Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
What I mean to say by that the biggest cause for Adversarial attacks is that NN creates linear boundaries locally, that's why Kernel SVM are the best defense against adversarial attacks.
Difficult-Race-1188 OP t1_iyluwu2 wrote
Reply to comment by Blakut in [D] Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
Even I don't know how VAE learn that. But recent paper that Neural networks can be written exactly like decision trees proved mathematically that NN are also decision tress but with added hyperspace.
Difficult-Race-1188 OP t1_iyc8451 wrote
Reply to comment by BrotherAmazing in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
https://arxiv.org/pdf/2210.05189.pdf
Read this paper, it's been proven that neural networks are decision trees, not a mere approximation but precisely that only. 3rd line in the abstract.
Difficult-Race-1188 OP t1_iyaxhbe wrote
Reply to comment by BrotherAmazing in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
The argument goes much further, NNs are not exactly learning the data distribution. If they had, the affine transformation problem would have been already taken care of, there would have been no need for data augmentation by rotating or flipping. Also approximating any algorithm doesn't necessarily mean the underlying data is following a distribution made out of any known algorithm. Also, Neural network struggle even to learn simple mathematical functions, all they do in the approximation is make piecewise assumptions of algorithms.
Here's the grokking paper review that told that NN couldn't generalize to this equation:
x³ + xy² + y (mod 97)
Article: https://medium.com/p/9dbbec1055ae
Original paper: https://arxiv.org/abs/2201.02177
Difficult-Race-1188 OP t1_iya8hg7 wrote
Reply to comment by Creepy_Disco_Spider in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
I've tried adding information from a lot of other resources. Not just one paper. And all of them are mentioned in the article.
Difficult-Race-1188 OP t1_iy7pdev wrote
Reply to comment by freaky1310 in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
So the paper which talks about the Spline theory of DL says that even in latent representation NN are incapable of interpolation and that's a very important thing to know about. If we know this then we can design loss functions that works to better understand the global manifold structures.
Difficult-Race-1188 OP t1_iy7p779 wrote
Reply to comment by ivan_kudryavtsev in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
It might behave in a similar fashion to DT, but DT doesn't make abstract feature representation and that is something important.
Difficult-Race-1188 OP t1_iy7p4ap wrote
Reply to comment by xtof54 in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
It does, if we know that NN behaves like DT then we can design new loss functions that take the internal structure into account. One of the research areas in this regard is Lipschitz Regularization. Adding such regularization makes NN behave more smoothly.
Difficult-Race-1188 OP t1_iy7gm4d wrote
Reply to comment by jazzzzzzzzzzzzzzzy in Neural Networks are just a bunch of Decision Trees by Difficult-Race-1188
This paper is a bit short, I've drawn conclusions from multiple papers like Spline's theory of Deep learning, why adversarial attacks exist, and the interpolation/extrapolation regime of Neural Nets.
Difficult-Race-1188 OP t1_j1888fo wrote
Reply to comment by ktpr in [D] Different types of pooling in Neural Nets by Difficult-Race-1188
https://arxiv.org/ftp/arxiv/papers/2009/2009.07485.pdf, I 've updated in the article also.