Python sklearn.metrics.zero_one_loss() Examples
The following are 14
code examples of sklearn.metrics.zero_one_loss().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
sklearn.metrics
, or try the search function
.
Example #1
Source File: multi_class_classification.py From edge2vec with BSD 3-Clause "New" or "Revised" License | 11 votes |
def multi_class_classification(data_X,data_Y): ''' calculate multi-class classification and return related evaluation metrics ''' svc = svm.SVC(C=1, kernel='linear') # X_train, X_test, y_train, y_test = train_test_split( data_X, data_Y, test_size=0.4, random_state=0) clf = svc.fit(data_X, data_Y) #svm # array = svc.coef_ # print array predicted = cross_val_predict(clf, data_X, data_Y, cv=2) print "accuracy",metrics.accuracy_score(data_Y, predicted) print "f1 score macro",metrics.f1_score(data_Y, predicted, average='macro') print "f1 score micro",metrics.f1_score(data_Y, predicted, average='micro') print "precision score",metrics.precision_score(data_Y, predicted, average='macro') print "recall score",metrics.recall_score(data_Y, predicted, average='macro') print "hamming_loss",metrics.hamming_loss(data_Y, predicted) print "classification_report", metrics.classification_report(data_Y, predicted) print "jaccard_similarity_score", metrics.jaccard_similarity_score(data_Y, predicted) # print "log_loss", metrics.log_loss(data_Y, predicted) print "zero_one_loss", metrics.zero_one_loss(data_Y, predicted) # print "AUC&ROC",metrics.roc_auc_score(data_Y, predicted) # print "matthews_corrcoef", metrics.matthews_corrcoef(data_Y, predicted)
Example #2
Source File: link_prediction.py From edge2vec with BSD 3-Clause "New" or "Revised" License | 7 votes |
def evaluation_analysis(true_label,predicted): ''' return all metrics results ''' print "accuracy",metrics.accuracy_score(true_label, predicted) print "f1 score macro",metrics.f1_score(true_label, predicted, average='macro') print "f1 score micro",metrics.f1_score(true_label, predicted, average='micro') print "precision score",metrics.precision_score(true_label, predicted, average='macro') print "recall score",metrics.recall_score(true_label, predicted, average='macro') print "hamming_loss",metrics.hamming_loss(true_label, predicted) print "classification_report", metrics.classification_report(true_label, predicted) print "jaccard_similarity_score", metrics.jaccard_similarity_score(true_label, predicted) print "log_loss", metrics.log_loss(true_label, predicted) print "zero_one_loss", metrics.zero_one_loss(true_label, predicted) print "AUC&ROC",metrics.roc_auc_score(true_label, predicted) print "matthews_corrcoef", metrics.matthews_corrcoef(true_label, predicted)
Example #3
Source File: tests.py From scikit-mdr with MIT License | 6 votes |
def test_mdr_custom_score(): """Ensure that the MDR 'score' function outputs the right custom score passed in from the user""" features = np.array([[2, 0], [0, 0], [0, 1], [0, 0], [0, 0], [0, 0], [0, 1], [0, 0], [0, 0], [0, 1], [0, 0], [0, 0], [0, 0], [1, 1], [1, 1]]) classes = np.array([1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0]) mdr = MDRClassifier() mdr.fit(features, classes) assert mdr.score(features = features, class_labels = classes, scoring_function = accuracy_score) == 12. / 15 assert mdr.score(features = features, class_labels = classes, scoring_function = zero_one_loss) == 1 - 12. / 15 assert mdr.score(features = features, class_labels = classes, scoring_function = zero_one_loss, normalize=False) == 15 - 12
Example #4
Source File: test_classification.py From Mastering-Elasticsearch-7.0 with MIT License | 5 votes |
def test_multilabel_zero_one_loss_subset(): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) assert_equal(zero_one_loss(y1, y2), 0.5) assert_equal(zero_one_loss(y1, y1), 0) assert_equal(zero_one_loss(y2, y2), 0) assert_equal(zero_one_loss(y2, np.logical_not(y2)), 1) assert_equal(zero_one_loss(y1, np.logical_not(y1)), 1) assert_equal(zero_one_loss(y1, np.zeros(y1.shape)), 1) assert_equal(zero_one_loss(y2, np.zeros(y1.shape)), 1)
Example #5
Source File: lm_plots.py From numpy-ml with GNU General Public License v3.0 | 5 votes |
def plot_logistic(): np.random.seed(12345) fig, axes = plt.subplots(4, 4) for i, ax in enumerate(axes.flatten()): n_in = 1 n_ex = 150 X_train, y_train, X_test, y_test = random_classification_problem( n_ex, n_classes=2, n_in=n_in, seed=i ) LR = LogisticRegression(penalty="l2", gamma=0.2, fit_intercept=True) LR.fit(X_train, y_train, lr=0.1, tol=1e-7, max_iter=1e7) y_pred = (LR.predict(X_test) >= 0.5) * 1.0 loss = zero_one_loss(y_test, y_pred) * 100.0 LR_sk = LogisticRegression_sk( penalty="l2", tol=0.0001, C=0.8, fit_intercept=True, random_state=i ) LR_sk.fit(X_train, y_train) y_pred_sk = (LR_sk.predict(X_test) >= 0.5) * 1.0 loss_sk = zero_one_loss(y_test, y_pred_sk) * 100.0 xmin = min(X_test) - 0.1 * (max(X_test) - min(X_test)) xmax = max(X_test) + 0.1 * (max(X_test) - min(X_test)) X_plot = np.linspace(xmin, xmax, 100) y_plot = LR.predict(X_plot) y_plot_sk = LR_sk.predict_proba(X_plot.reshape(-1, 1))[:, 1] ax.scatter(X_test[y_pred == 0], y_test[y_pred == 0], alpha=0.5) ax.scatter(X_test[y_pred == 1], y_test[y_pred == 1], alpha=0.5) ax.plot(X_plot, y_plot, label="mine", alpha=0.75) ax.plot(X_plot, y_plot_sk, label="sklearn", alpha=0.75) ax.legend() ax.set_title("Loss mine: {:.2f} Loss sklearn: {:.2f}".format(loss, loss_sk)) ax.xaxis.set_ticklabels([]) ax.yaxis.set_ticklabels([]) plt.tight_layout() plt.savefig("plot_logistic.png", dpi=300) plt.close("all")
Example #6
Source File: test_gentleboost.py From skboost with MIT License | 5 votes |
def test_gentleboost_musk_fitting(): c = GentleBoostClassifier( base_estimator=DecisionTreeRegressor(max_depth=1), n_estimators=30, learning_rate=1.0 ) data = MUSK1() c.fit(data.data, np.sign(data.labels)) assert_array_less(c.estimator_errors_, 0.5) assert zero_one_loss(np.sign(data.labels), c.predict(data.data)) < 0.1
Example #7
Source File: test_gentleboost.py From skboost with MIT License | 5 votes |
def test_gentleboost_hastie_fitting(): c = GentleBoostClassifier( base_estimator=DecisionTreeRegressor(max_depth=1), n_estimators=30, learning_rate=1.0 ) data = Hastie_10_2() c.fit(data.data, np.sign(data.labels)) assert_array_less(c.estimator_errors_, 0.5) assert zero_one_loss(np.sign(data.labels), c.predict(data.data)) < 0.2
Example #8
Source File: test_milboost.py From skboost with MIT License | 5 votes |
def test_milboost_musk_fitting_lse(): c = MILBoostClassifier( base_estimator=DecisionTreeClassifier(max_depth=1), softmax=LogSumExponential(5.0), n_estimators=30, learning_rate=1.0 ) data = MUSK1() c.fit(data.data, data.labels) assert_array_less(c.estimator_errors_, 0.5) assert zero_one_loss(np.sign(data.labels), c.predict(data.data)) < 0.30
Example #9
Source File: test_milboost.py From skboost with MIT License | 5 votes |
def test_milboost_hastie_fitting(): c = MILBoostClassifier( base_estimator=DecisionTreeClassifier(max_depth=1), softmax=LogSumExponential(5.0), n_estimators=30, learning_rate=1.0 ) data = Hastie_10_2() c.fit(data.data, data.labels) assert_array_less(c.estimator_errors_, 0.5) assert zero_one_loss(np.sign(data.labels), c.predict(data.data)) < 0.40
Example #10
Source File: test_logitboost.py From skboost with MIT License | 5 votes |
def test_logitboost_musk_fitting(): c = LogitBoostClassifier( base_estimator=DecisionTreeRegressor(max_depth=1), n_estimators=30, learning_rate=1.0 ) data = MUSK1() c.fit(data.data, np.sign(data.labels)) assert_array_less(c.estimator_errors_, 0.6) assert zero_one_loss(np.sign(data.labels), c.predict(data.data)) < 0.05
Example #11
Source File: adaBoostClassifier.py From TextDetector with GNU General Public License v3.0 | 5 votes |
def fit(self, data, target): no_of_stages = self.no_of_stages decision_stump = DecisionTreeClassifier(criterion='gini', splitter='best', max_depth=1, max_features=1) #No. of samples m = data.shape[0] weight = numpy.ones(m) weight = numpy.float32(weight)/m Alpha = numpy.zeros(no_of_stages) classifiers = [] for i in range(no_of_stages): decision_stump = decision_stump.fit(data, target, sample_weight = weight) classifiers.append(decision_stump) pred = decision_stump.predict(data) error = zero_one_loss(target, pred, normalize=True, sample_weight = weight) if error > 0.5: print 'error value is greater than 0.5!' beta = error/(1-error) if beta != 0: weight[pred == target] = weight[pred==target]*beta weight = weight / weight.sum() print weight # beta_mat = (pred==target)*beta # beta_mat[beta_mat==0] = 1 # weight = numpy.multiply(weight, beta_mat) if beta > 0: alpha = math.log(1/beta) else: alpha = 10000 # make alpha extremly large if decision stump is totally correct. Alpha[i] = alpha self.Alpha = Alpha self.classifiers = classifiers
Example #12
Source File: test_metrics.py From pandas-ml with BSD 3-Clause "New" or "Revised" License | 5 votes |
def test_zero_one_loss(self): result = self.df.metrics.zero_one_loss() expected = metrics.zero_one_loss(self.target, self.pred) self.assertEqual(result, expected)
Example #13
Source File: metrics_np.py From cs-ranking with Apache License 2.0 | 5 votes |
def subset_01_loss(y_true, y_pred): return zero_one_loss(y_true, y_pred)
Example #14
Source File: test_classification.py From twitter-stock-recommendation with MIT License | 5 votes |
def test_multilabel_zero_one_loss_subset(): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) assert_equal(zero_one_loss(y1, y2), 0.5) assert_equal(zero_one_loss(y1, y1), 0) assert_equal(zero_one_loss(y2, y2), 0) assert_equal(zero_one_loss(y2, np.logical_not(y2)), 1) assert_equal(zero_one_loss(y1, np.logical_not(y1)), 1) assert_equal(zero_one_loss(y1, np.zeros(y1.shape)), 1) assert_equal(zero_one_loss(y2, np.zeros(y1.shape)), 1)