Cross-validation and out-of-bag bootstrap applications
$begingroup$
I have a question regarding steps on which a specific resample method should be used in general.
As far as I know:
out-of-bag bootstrap is the resample method with replacement, which has lower variance and higher bias compared to the cross-validation method.
cross-validation is the resample method without replacement, which has higher variance and lower bias compared to the out-of-bag bootstrap method.
Following descriptions above, isn't it better (in general) to tune models with out-of-bag bootstrap resampling and use cross-validation only for final scoring of the tuned models?
Tuning is the process, which should answer the question: What is the best model configuration according to the specific evaluation metric? So, we're not interested in finding exact performance measure, but we would rather focus only on comparison of different configurations. Lower variance means a lower risk of choosing a model with a high mean performance, although few sets have a very poor performance.
On the other hand, cross-validation performed on the tuned model gives us an answer to the question: What is the approximate performance of the tuned model?
machine-learning cross-validation performance sampling hyperparameter-tuning
New contributor
$endgroup$
add a comment |
$begingroup$
I have a question regarding steps on which a specific resample method should be used in general.
As far as I know:
out-of-bag bootstrap is the resample method with replacement, which has lower variance and higher bias compared to the cross-validation method.
cross-validation is the resample method without replacement, which has higher variance and lower bias compared to the out-of-bag bootstrap method.
Following descriptions above, isn't it better (in general) to tune models with out-of-bag bootstrap resampling and use cross-validation only for final scoring of the tuned models?
Tuning is the process, which should answer the question: What is the best model configuration according to the specific evaluation metric? So, we're not interested in finding exact performance measure, but we would rather focus only on comparison of different configurations. Lower variance means a lower risk of choosing a model with a high mean performance, although few sets have a very poor performance.
On the other hand, cross-validation performed on the tuned model gives us an answer to the question: What is the approximate performance of the tuned model?
machine-learning cross-validation performance sampling hyperparameter-tuning
New contributor
$endgroup$
add a comment |
$begingroup$
I have a question regarding steps on which a specific resample method should be used in general.
As far as I know:
out-of-bag bootstrap is the resample method with replacement, which has lower variance and higher bias compared to the cross-validation method.
cross-validation is the resample method without replacement, which has higher variance and lower bias compared to the out-of-bag bootstrap method.
Following descriptions above, isn't it better (in general) to tune models with out-of-bag bootstrap resampling and use cross-validation only for final scoring of the tuned models?
Tuning is the process, which should answer the question: What is the best model configuration according to the specific evaluation metric? So, we're not interested in finding exact performance measure, but we would rather focus only on comparison of different configurations. Lower variance means a lower risk of choosing a model with a high mean performance, although few sets have a very poor performance.
On the other hand, cross-validation performed on the tuned model gives us an answer to the question: What is the approximate performance of the tuned model?
machine-learning cross-validation performance sampling hyperparameter-tuning
New contributor
$endgroup$
I have a question regarding steps on which a specific resample method should be used in general.
As far as I know:
out-of-bag bootstrap is the resample method with replacement, which has lower variance and higher bias compared to the cross-validation method.
cross-validation is the resample method without replacement, which has higher variance and lower bias compared to the out-of-bag bootstrap method.
Following descriptions above, isn't it better (in general) to tune models with out-of-bag bootstrap resampling and use cross-validation only for final scoring of the tuned models?
Tuning is the process, which should answer the question: What is the best model configuration according to the specific evaluation metric? So, we're not interested in finding exact performance measure, but we would rather focus only on comparison of different configurations. Lower variance means a lower risk of choosing a model with a high mean performance, although few sets have a very poor performance.
On the other hand, cross-validation performed on the tuned model gives us an answer to the question: What is the approximate performance of the tuned model?
machine-learning cross-validation performance sampling hyperparameter-tuning
machine-learning cross-validation performance sampling hyperparameter-tuning
New contributor
New contributor
edited yesterday
Michał Kardach
New contributor
asked yesterday
Michał KardachMichał Kardach
14
14
New contributor
New contributor
add a comment |
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "557"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Michał Kardach is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f46154%2fcross-validation-and-out-of-bag-bootstrap-applications%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Michał Kardach is a new contributor. Be nice, and check out our Code of Conduct.
Michał Kardach is a new contributor. Be nice, and check out our Code of Conduct.
Michał Kardach is a new contributor. Be nice, and check out our Code of Conduct.
Michał Kardach is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Data Science Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f46154%2fcross-validation-and-out-of-bag-bootstrap-applications%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown