working with an extremely unbalanced and poorly correlated dataset
Hi I am working with a difficult data set, in that the classes are both highly unbalanced and extremely uncorrelated (The set has 96,000 values, of which less than 200 are 1s.). I tried a few methods, and with each the precision and accuracy were always high, however only a few (less than 5) values are being classified as 1. I wonder if there is a way to force the machine to classify more 1s. If I could classify correctly just 25% of the time, this would be a great result.
I have tried using random forest's 'class weight' parameter, but this doesn't seem to have any effect on the result.
Thanks
import numpy as np
import pandas as pd
import sklearn as sklearn
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import seaborn as sns
df = pd.read_pickle('/Users/shellyganga/Downloads/ola.pickle')
print(df.describe())
#filtering the df to improve results
df = df[(df['trip_duration'] > 5) & (df['Smooth_Driving_Score'] < 99)]
print(df.describe())
maxVal = 1
df.unsafe = df['unsafe'].where(df['unsafe'] <= maxVal, maxVal)
df.drop(df.columns[0], axis=1, inplace=True)
df.drop(df.columns[-2], axis=1, inplace=True)
#setting features and labels
labels = np.array(df['unsafe'])
features= df.drop('unsafe', axis = 1)
# Saving feature names for later use
feature_list = list(features.columns)
# Convert to numpy array
features = np.array(features)
from sklearn.model_selection import train_test_split
# 30% examples in test data
train, test, train_labels, test_labels = train_test_split(features, labels,
stratify = labels,
test_size = 0.4,
random_state = 12)
from sklearn.ensemble import RandomForestClassifier
# Create the model with 100 trees
model = RandomForestClassifier(n_estimators=100,
random_state=12,
max_features = 'sqrt',
n_jobs=-1, verbose = 1, class_weight={0:1, 1:1})
# Fit on training data
model.fit(train, train_labels)
predictions = model.predict(test)
print(np.mean(predictions))
print(predictions.shape)
from sklearn.metrics import classification_report
print(classification_report(test_labels, predictions)
ouput
precision recall f1-score support
0 1.00 1.00 1.00 38300
1 1.00 0.01 0.02 90
avg / total 1.00 1.00 1.00 38390
edit: I tried using {class_weight = 'balanced'} and provided a different result, but I am having trouble understanding it.
micro avg 1.00 1.00 1.00 38390
macro avg 1.00 0.51 0.51 38390
weighted avg 1.00 1.00 1.00 38390
How do I know how many positives it has predicted?
python scikit-learn classification weight
add a comment |
Hi I am working with a difficult data set, in that the classes are both highly unbalanced and extremely uncorrelated (The set has 96,000 values, of which less than 200 are 1s.). I tried a few methods, and with each the precision and accuracy were always high, however only a few (less than 5) values are being classified as 1. I wonder if there is a way to force the machine to classify more 1s. If I could classify correctly just 25% of the time, this would be a great result.
I have tried using random forest's 'class weight' parameter, but this doesn't seem to have any effect on the result.
Thanks
import numpy as np
import pandas as pd
import sklearn as sklearn
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import seaborn as sns
df = pd.read_pickle('/Users/shellyganga/Downloads/ola.pickle')
print(df.describe())
#filtering the df to improve results
df = df[(df['trip_duration'] > 5) & (df['Smooth_Driving_Score'] < 99)]
print(df.describe())
maxVal = 1
df.unsafe = df['unsafe'].where(df['unsafe'] <= maxVal, maxVal)
df.drop(df.columns[0], axis=1, inplace=True)
df.drop(df.columns[-2], axis=1, inplace=True)
#setting features and labels
labels = np.array(df['unsafe'])
features= df.drop('unsafe', axis = 1)
# Saving feature names for later use
feature_list = list(features.columns)
# Convert to numpy array
features = np.array(features)
from sklearn.model_selection import train_test_split
# 30% examples in test data
train, test, train_labels, test_labels = train_test_split(features, labels,
stratify = labels,
test_size = 0.4,
random_state = 12)
from sklearn.ensemble import RandomForestClassifier
# Create the model with 100 trees
model = RandomForestClassifier(n_estimators=100,
random_state=12,
max_features = 'sqrt',
n_jobs=-1, verbose = 1, class_weight={0:1, 1:1})
# Fit on training data
model.fit(train, train_labels)
predictions = model.predict(test)
print(np.mean(predictions))
print(predictions.shape)
from sklearn.metrics import classification_report
print(classification_report(test_labels, predictions)
ouput
precision recall f1-score support
0 1.00 1.00 1.00 38300
1 1.00 0.01 0.02 90
avg / total 1.00 1.00 1.00 38390
edit: I tried using {class_weight = 'balanced'} and provided a different result, but I am having trouble understanding it.
micro avg 1.00 1.00 1.00 38390
macro avg 1.00 0.51 0.51 38390
weighted avg 1.00 1.00 1.00 38390
How do I know how many positives it has predicted?
python scikit-learn classification weight
1
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
1
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17
add a comment |
Hi I am working with a difficult data set, in that the classes are both highly unbalanced and extremely uncorrelated (The set has 96,000 values, of which less than 200 are 1s.). I tried a few methods, and with each the precision and accuracy were always high, however only a few (less than 5) values are being classified as 1. I wonder if there is a way to force the machine to classify more 1s. If I could classify correctly just 25% of the time, this would be a great result.
I have tried using random forest's 'class weight' parameter, but this doesn't seem to have any effect on the result.
Thanks
import numpy as np
import pandas as pd
import sklearn as sklearn
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import seaborn as sns
df = pd.read_pickle('/Users/shellyganga/Downloads/ola.pickle')
print(df.describe())
#filtering the df to improve results
df = df[(df['trip_duration'] > 5) & (df['Smooth_Driving_Score'] < 99)]
print(df.describe())
maxVal = 1
df.unsafe = df['unsafe'].where(df['unsafe'] <= maxVal, maxVal)
df.drop(df.columns[0], axis=1, inplace=True)
df.drop(df.columns[-2], axis=1, inplace=True)
#setting features and labels
labels = np.array(df['unsafe'])
features= df.drop('unsafe', axis = 1)
# Saving feature names for later use
feature_list = list(features.columns)
# Convert to numpy array
features = np.array(features)
from sklearn.model_selection import train_test_split
# 30% examples in test data
train, test, train_labels, test_labels = train_test_split(features, labels,
stratify = labels,
test_size = 0.4,
random_state = 12)
from sklearn.ensemble import RandomForestClassifier
# Create the model with 100 trees
model = RandomForestClassifier(n_estimators=100,
random_state=12,
max_features = 'sqrt',
n_jobs=-1, verbose = 1, class_weight={0:1, 1:1})
# Fit on training data
model.fit(train, train_labels)
predictions = model.predict(test)
print(np.mean(predictions))
print(predictions.shape)
from sklearn.metrics import classification_report
print(classification_report(test_labels, predictions)
ouput
precision recall f1-score support
0 1.00 1.00 1.00 38300
1 1.00 0.01 0.02 90
avg / total 1.00 1.00 1.00 38390
edit: I tried using {class_weight = 'balanced'} and provided a different result, but I am having trouble understanding it.
micro avg 1.00 1.00 1.00 38390
macro avg 1.00 0.51 0.51 38390
weighted avg 1.00 1.00 1.00 38390
How do I know how many positives it has predicted?
python scikit-learn classification weight
Hi I am working with a difficult data set, in that the classes are both highly unbalanced and extremely uncorrelated (The set has 96,000 values, of which less than 200 are 1s.). I tried a few methods, and with each the precision and accuracy were always high, however only a few (less than 5) values are being classified as 1. I wonder if there is a way to force the machine to classify more 1s. If I could classify correctly just 25% of the time, this would be a great result.
I have tried using random forest's 'class weight' parameter, but this doesn't seem to have any effect on the result.
Thanks
import numpy as np
import pandas as pd
import sklearn as sklearn
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import seaborn as sns
df = pd.read_pickle('/Users/shellyganga/Downloads/ola.pickle')
print(df.describe())
#filtering the df to improve results
df = df[(df['trip_duration'] > 5) & (df['Smooth_Driving_Score'] < 99)]
print(df.describe())
maxVal = 1
df.unsafe = df['unsafe'].where(df['unsafe'] <= maxVal, maxVal)
df.drop(df.columns[0], axis=1, inplace=True)
df.drop(df.columns[-2], axis=1, inplace=True)
#setting features and labels
labels = np.array(df['unsafe'])
features= df.drop('unsafe', axis = 1)
# Saving feature names for later use
feature_list = list(features.columns)
# Convert to numpy array
features = np.array(features)
from sklearn.model_selection import train_test_split
# 30% examples in test data
train, test, train_labels, test_labels = train_test_split(features, labels,
stratify = labels,
test_size = 0.4,
random_state = 12)
from sklearn.ensemble import RandomForestClassifier
# Create the model with 100 trees
model = RandomForestClassifier(n_estimators=100,
random_state=12,
max_features = 'sqrt',
n_jobs=-1, verbose = 1, class_weight={0:1, 1:1})
# Fit on training data
model.fit(train, train_labels)
predictions = model.predict(test)
print(np.mean(predictions))
print(predictions.shape)
from sklearn.metrics import classification_report
print(classification_report(test_labels, predictions)
ouput
precision recall f1-score support
0 1.00 1.00 1.00 38300
1 1.00 0.01 0.02 90
avg / total 1.00 1.00 1.00 38390
edit: I tried using {class_weight = 'balanced'} and provided a different result, but I am having trouble understanding it.
micro avg 1.00 1.00 1.00 38390
macro avg 1.00 0.51 0.51 38390
weighted avg 1.00 1.00 1.00 38390
How do I know how many positives it has predicted?
python scikit-learn classification weight
python scikit-learn classification weight
edited Nov 18 '18 at 8:15
user3107977
asked Nov 17 '18 at 18:14
user3107977user3107977
233
233
1
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
1
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17
add a comment |
1
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
1
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17
1
1
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like
{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like
{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
1
1
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53354123%2fworking-with-an-extremely-unbalanced-and-poorly-correlated-dataset%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Stack Overflow!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53354123%2fworking-with-an-extremely-unbalanced-and-poorly-correlated-dataset%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
class_weight is the way to do it. What values did you try? From the docs, it looks like you should try class_weight='balanced' which will automatically create something roughly like
{0:(200.0/96000), 1:1}
– Luke
Nov 17 '18 at 20:46
1
thanks, I added an edit with the output I got but I still don't understand it
– user3107977
Nov 18 '18 at 8:17