The code below does a lot in only a few lines. To help explain things, here are the steps that code is doing:
- Split the raw data into three folds. Select one for testing and two for training.
- Preprocess the data by scaling the training features.
- Train a support vector classifier on the training data.
- Apply the classifier to the test data.
- Record the accuracy score.
- Repeat steps 1-5 two more times, once for each fold.
- Calculate the mean score for all the folds.
from sklearn.datasets import load_iris from sklearn.pipeline import make_pipeline from sklearn import preprocessing from sklearn import cross_validation from sklearn import svm
For this tutorial we will use the famous iris dataset. The iris data contains four measurements of 150 iris flowers and their species. We will use a support vector classifier to predict the species of the iris flowers.
# Load the iris test data iris = load_iris()
# View the iris data features for the first three rows iris.data[0:3]
array([[ 5.1, 3.5, 1.4, 0.2], [ 4.9, 3. , 1.4, 0.2], [ 4.7, 3.2, 1.3, 0.2]])
# View the iris data target for first three rows. '0' means it flower is of the setosa species. iris.target[0:3]
array([0, 0, 0])
Create Classifier Pipeline
Now we create a pipeline for the data. First, the pipeline preprocesses the data by scaling the feature variable's values to mean zero and unit variance. Second, the pipeline trains a support classifier on the data with
C is the cost function for the margins. The higher the C, the less tolerant the model is for observations being on the wrong side of the hyperplane.
# Create a pipeline that scales the data then trains a support vector classifier classifier_pipeline = make_pipeline(preprocessing.StandardScaler(), svm.SVC(C=1))
Scikit provides a great helper function to make it easy to do cross validation. Specifically, the code below splits the data into three folds, then executes the classifier pipeline on the iris data.
Important note from the scikit docs: For integer/None inputs, if y is binary or multiclass, StratifiedKFold used. If the estimator is a classifier or if y is neither binary nor multiclass, KFold is used.
# KFold/StratifiedKFold cross validation with 3 folds (the default) # applying the classifier pipeline to the feature and target data scores = cross_validation.cross_val_score(classifier_pipeline, iris.data, iris.target, cv=3)
Here is the output of our 3 KFold cross validation. Each value is the accuracy score of the support vector classifier when leaving out a different fold. There are three values because there are three folds. A higher accuracy score, the better.
array([ 0.98039216, 0.90196078, 0.97916667])
To get an good measure of the model's accuracy, we calculate the mean of the three scores. This is our measure of model accuracy.