S. Iqbal S. Iqbal - 1 year ago 232
Python Question

sklearn SGDClassifier's decision_function odd behavior

So I am using a simple SGDClassifier on the MNIST dataset (as per the Hands-on ML book) and I can't seem to figure out the behavior of its decision_function.

I changed the last line of the original_decision function to specifically check if anything was different. The variables with suffix "check" are the ones returned by the decision_function here. The code:

from sklearn.datasets import fetch_mldata
mnist = fetch_mldata("MNIST original")
import numpy as np
from sklearn.linear_model import SGDClassifier
from sklearn.utils.extmath import safe_sparse_dot
from sklearn.utils import check_array

X, y = mnist["data"], mnist["target"]
X_train, X_test, y_train, y_test = X[:60000], X[60000:], y[:60000], y[60000:]

shuffle_index = np.random.permutation(60000)
X_train, y_train = X_train[shuffle_index], y_train[shuffle_index]

# converting the problem into a binary classification problem.
y_train_5 = (y_train == 5)
y_test_5 = (y_test == 5)

sgd_clf = SGDClassifier(random_state=42)
sgd_clf.fit(X_train, y_train_5)

# modified decision_func to ouput vars
X_check, coef_check, intcpt_check = sgd_clf.decision_function([X_train[36000]])

print((X_check == X_train[36000]).all())
print((coef_check == sgd_clf.coef_).all())
print((intcpt_check == sgd_clf.intercept_).all())

# using same funcs as used by decision_function to calc.
# i.e. check_array safe_sparse_dot
X_mod = check_array(X[36000].reshape(1,-1), "csr")
my_score = safe_sparse_dot(X_mod,sgd_clf.coef_.T) + sgd_clf.intercept_
sk_score = safe_sparse_dot(X_check, coef_check.T) + intcpt_check


Here is the output (for a single run):

[[ 49505.1725926]]

Here is the modification I made to the decision_function (2nd last line before the commented out original one):

scores = safe_sparse_dot(X, self.coef_.T,
dense_output=True) + self.intercept_
return X, self.coef_, self.intercept_
#return scores.ravel() if scores.shape[1] == 1 else scores

Even though the 3 entities involved (instance X, coefficients, intercept) all match to be true with my variables, the multiplication still leads to vastly different outcomes.

Why is this happening?

Edit: Curiously I found out that if I comment out the two lines responsible for shuffling the dataset namely:

shuffle_index = np.random.permutation(60000)
X_train, y_train = X_train[shuffle_index], y_train[shuffle_index]

The problem disappears...

Answer Source

Its probably due to X_mod using the original "X" (unshuffled) matrix while X_check uses the X_train(shuffled) index.

i.e. X_train[36000]!=X[36000] so of course, your scores should not be the same.

Recommended from our users: Dynamic Network Monitoring from WhatsUp Gold from IPSwitch. Free Download