Visual-Arm-7375
Visual-Arm-7375 OP t1_iz5iyf3 wrote
Reply to comment by rahuldave in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
And why the overfitting depends on the number of comparisons, isn't the overfitting something relation to each model separately?
Visual-Arm-7375 OP t1_iz5inba wrote
Reply to comment by rahuldave in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Thank's for the answer! I don't understand the separation you are doing btw training and validation. Didn't we have train/test and we applied cv to the train? The validation sets would be 1 fold at each cv iteration. What I am not understanding here?
Visual-Arm-7375 OP t1_iz2fh6e wrote
Reply to comment by rahuldave in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Thank you very much for the answer!
One question, what do you mean in this context by estimates? Hyperparameters?
>But if you are comparing 3-4 estimates of the error on the test set to choose the best model class this is not a large comparison, and so the test set is "not so contaminated" by this comparison, and can be used for other purposes.
Could you explain this in another way pls, I'm not sure I am understanding it :(
Visual-Arm-7375 OP t1_iz2e6z6 wrote
Reply to comment by MUSEy69 in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Thank you for the reply! Step 5 is because I have to submit the predictions for a separated from which I don't know the labels. So my idea was to use all the data.
Visual-Arm-7375 OP t1_iz065iq wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
I don't have a clear opinion, I'm trying to learn and I'm proposing a situation and you're not listening. You are evaluating the performance of the model with the same accuracy you are selecting hyperparameters, this does not make sense.
Anyway, thank you for your help, really appreciate it.
Visual-Arm-7375 OP t1_iz04hi2 wrote
Reply to comment by Visual-Arm-7375 in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
What do you think about it?
Visual-Arm-7375 OP t1_iz04glk wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Have a look at this: https://towardsdatascience.com/train-test-split-c3eed34f763b
Visual-Arm-7375 OP t1_iz03pqj wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Is not validation data, it is test data. You haven't checked the accuracy on the test data as another fold you average for getting the mean accuracy in the cross-validation. There you can see how the model is generalizing with the hyperparameters selected in the cross-validation.
Visual-Arm-7375 OP t1_iz03bcf wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Mmmm okay. But imagine you have 1000 datapoints and you want to compare a random forest and a DNN and select which one is the best to put it into production, how would you do it?
Visual-Arm-7375 OP t1_iz02j8x wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
I don't know if you know what I mean, how can you test if the hyperparameters are overfitting if you have selected them as the ones that maximize the mean accuracy across all the folds.
Visual-Arm-7375 OP t1_iz02byy wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Not really, because you are taking the average across all the folds, so at some point you are using that validation splits for training some of the folds, not with the test split.
Visual-Arm-7375 OP t1_iz01vfm wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Yeah, I understand that. But what's the point of separating the test set then? You are using cross-validation to select the hyperparameters, you are not seeing how they work in new data...
Visual-Arm-7375 OP t1_iz016u8 wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Okay, got two answers, one that says separate always a test set and the other one that is useless :(
Submitted by Visual-Arm-7375 t3_zd6a6j in MachineLearning
Visual-Arm-7375 OP t1_is4vqus wrote
Reply to comment by graphicteadatasci in [P] Understanding LIME | Explainable AI by Visual-Arm-7375
But is this LIME's problem? I mean, it is the model that is not taking into account the correlated feature, not LIME. LIME just looks at the original model.
Visual-Arm-7375 OP t1_is4vmw7 wrote
Reply to comment by TenaciousDwight in [P] Understanding LIME | Explainable AI by Visual-Arm-7375
Damn! The paper is really interesting u/TenaciousDwight! Thanks :) Appreciate it.
Visual-Arm-7375 OP t1_is2qev6 wrote
Reply to comment by ktpr in [P] Understanding LIME | Explainable AI by Visual-Arm-7375
Nice! Thank you :)
Visual-Arm-7375 OP t1_is2qdv9 wrote
Reply to comment by danabxy in [P] Understanding LIME | Explainable AI by Visual-Arm-7375
Thank you very much for the comment! I don't take it badly at all. Constructive criticism is always welcome. Regarding my accent, I really try to give the best of myself but it's innate haha, I'm sorry. I will try to improve it although I don't know how. Also, I recorded with the laptop, I have no micro :( However, I provided detailed subtitles in the video!!
Visual-Arm-7375 OP t1_is0jefe wrote
Reply to comment by flinsypop in [P] Understanding LIME | Explainable AI by Visual-Arm-7375
Thank you very much for your opinion u/flinsypop! Appreciate it a lot!
I completely agree with what you say, and I'll take that in mind for the next videos.
Submitted by Visual-Arm-7375 t3_y1zg5r in MachineLearning
Visual-Arm-7375 OP t1_iz7eoyv wrote
Reply to comment by rahuldave in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Really appreciate your time! Thank you very much!