Difficult-Race-1188

Difficult-Race-1188 OP t1_iylz2aw wrote

What people mean when they say AI is not truly learning is that often the most impressive results are coming from extremely big models. For example, almost all the top AI scientist takes dig on Large language models, because we don't know whether they learned something or it just memorized all the possible combinations. Why people believe AI is not truly learning is that there are papers that show that AI was unable to generalize to simple mathematical equations.

x³ + xy² + y (mod 97), AI was unable to generalize to this simple equation.

https://medium.com/aiguys/paper-review-grokking-generalization-and-over-fitting-9dbbec1055ae

https://arxiv.org/abs/2201.02177

−1

Difficult-Race-1188 OP t1_iyaxhbe wrote

The argument goes much further, NNs are not exactly learning the data distribution. If they had, the affine transformation problem would have been already taken care of, there would have been no need for data augmentation by rotating or flipping. Also approximating any algorithm doesn't necessarily mean the underlying data is following a distribution made out of any known algorithm. Also, Neural network struggle even to learn simple mathematical functions, all they do in the approximation is make piecewise assumptions of algorithms.

Here's the grokking paper review that told that NN couldn't generalize to this equation:

x³ + xy² + y (mod 97)

Article: https://medium.com/p/9dbbec1055ae

Original paper: https://arxiv.org/abs/2201.02177

1