Adding recommendations to the output of a classification model
$begingroup$
I have built a binary classification model using:
- logit
- decision trees
- random forest
- bagging classifier
- gradientboost
- xgboost
- adaboost
I have evaluated the above models and chose xgboost based on training/test and validation metrics (accuracy, prediction, recall, f1 and AUC).
I want to now productionalize it and share the output with the business. The output would basically have a list of items with the predicted class and that could be filtered based on business needs.
However, Instead of simply giving the business the predicted classes, I want to add insights/recommendations as to why a specific item was predicted with class X and how you could go about working on the item to change its class from say X to Y.
How do I go about this? I thought of using feature importance, but my input data shape is [800,000 * 1,050] and I am not sure if it would the best way to proceed.
Are there any existing industry standard methodologies that can add interpretability to such models and convert them from a black box models to prescriptive models?
machine-learning python classification data-science-model
$endgroup$
bumped to the homepage by Community♦ 9 mins ago
This question has answers that may be good or bad; the system has marked it active so that they can be reviewed.
add a comment |
$begingroup$
I have built a binary classification model using:
- logit
- decision trees
- random forest
- bagging classifier
- gradientboost
- xgboost
- adaboost
I have evaluated the above models and chose xgboost based on training/test and validation metrics (accuracy, prediction, recall, f1 and AUC).
I want to now productionalize it and share the output with the business. The output would basically have a list of items with the predicted class and that could be filtered based on business needs.
However, Instead of simply giving the business the predicted classes, I want to add insights/recommendations as to why a specific item was predicted with class X and how you could go about working on the item to change its class from say X to Y.
How do I go about this? I thought of using feature importance, but my input data shape is [800,000 * 1,050] and I am not sure if it would the best way to proceed.
Are there any existing industry standard methodologies that can add interpretability to such models and convert them from a black box models to prescriptive models?
machine-learning python classification data-science-model
$endgroup$
bumped to the homepage by Community♦ 9 mins ago
This question has answers that may be good or bad; the system has marked it active so that they can be reviewed.
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05
add a comment |
$begingroup$
I have built a binary classification model using:
- logit
- decision trees
- random forest
- bagging classifier
- gradientboost
- xgboost
- adaboost
I have evaluated the above models and chose xgboost based on training/test and validation metrics (accuracy, prediction, recall, f1 and AUC).
I want to now productionalize it and share the output with the business. The output would basically have a list of items with the predicted class and that could be filtered based on business needs.
However, Instead of simply giving the business the predicted classes, I want to add insights/recommendations as to why a specific item was predicted with class X and how you could go about working on the item to change its class from say X to Y.
How do I go about this? I thought of using feature importance, but my input data shape is [800,000 * 1,050] and I am not sure if it would the best way to proceed.
Are there any existing industry standard methodologies that can add interpretability to such models and convert them from a black box models to prescriptive models?
machine-learning python classification data-science-model
$endgroup$
I have built a binary classification model using:
- logit
- decision trees
- random forest
- bagging classifier
- gradientboost
- xgboost
- adaboost
I have evaluated the above models and chose xgboost based on training/test and validation metrics (accuracy, prediction, recall, f1 and AUC).
I want to now productionalize it and share the output with the business. The output would basically have a list of items with the predicted class and that could be filtered based on business needs.
However, Instead of simply giving the business the predicted classes, I want to add insights/recommendations as to why a specific item was predicted with class X and how you could go about working on the item to change its class from say X to Y.
How do I go about this? I thought of using feature importance, but my input data shape is [800,000 * 1,050] and I am not sure if it would the best way to proceed.
Are there any existing industry standard methodologies that can add interpretability to such models and convert them from a black box models to prescriptive models?
machine-learning python classification data-science-model
machine-learning python classification data-science-model
edited Nov 16 '18 at 10:23
praveen
asked Nov 15 '18 at 12:19
praveenpraveen
1212
1212
bumped to the homepage by Community♦ 9 mins ago
This question has answers that may be good or bad; the system has marked it active so that they can be reviewed.
bumped to the homepage by Community♦ 9 mins ago
This question has answers that may be good or bad; the system has marked it active so that they can be reviewed.
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05
add a comment |
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
a link. This is a link where someone has answered a similar question like that of yours. Have a read to see if it helps.
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "557"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f41259%2fadding-recommendations-to-the-output-of-a-classification-model%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
a link. This is a link where someone has answered a similar question like that of yours. Have a read to see if it helps.
$endgroup$
add a comment |
$begingroup$
a link. This is a link where someone has answered a similar question like that of yours. Have a read to see if it helps.
$endgroup$
add a comment |
$begingroup$
a link. This is a link where someone has answered a similar question like that of yours. Have a read to see if it helps.
$endgroup$
a link. This is a link where someone has answered a similar question like that of yours. Have a read to see if it helps.
answered Nov 17 '18 at 22:25
SudhiSudhi
415
415
add a comment |
add a comment |
Thanks for contributing an answer to Data Science Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f41259%2fadding-recommendations-to-the-output-of-a-classification-model%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
If the features are sparse, you might want to simply use standardized coefficients from logistic regression, or select some individual decision trees. Explaining how can the label change is however a lot more problematic, and an optimization problem of its own. Generally speaking, if you need explanatory ability, you'll have to move away from black-box models.
$endgroup$
– anymous.asker
Nov 17 '18 at 15:05