Test datasets are small contrived datasets that let you test a machine learning algorithm or test harness.
The data from test datasets have welldefined properties, such as linearly or nonlinearity, that allow you to explore specific algorithm behavior. The scikitlearn Python library provides a suite of functions for generating samples from configurable test problems for regression and classification.
In this tutorial, you will discover test problems and how to use them in Python with scikitlearn.
After completing this tutorial, you will know:
 How to generate multiclass classification prediction test problems.
 How to generate binary classification prediction test problems.
 How to generate linear regression prediction test problems.
Let’s get started.
Tutorial Overview
This tutorial is divided into 3 parts; they are:
 Test Datasets
 Classification Test Problems
 Regression Test Problems
Test Datasets
A problem when developing and implementing machine learning algorithms is how do you know whether you have implemented them correctly. They seem to work even with bugs.
Test datasets are small contrived problems that allow you to test and debug your algorithms and test harness. They are also useful for better understanding the behavior of algorithms in response to changes in hyperparameters.
Below are some desirable properties of test datasets:
 They can be generated quickly and easily.
 They contain “known” or “understood” outcomes for comparison with predictions.
 They are stochastic, allowing random variations on the same problem each time they are generated.
 They are small and easily visualized in two dimensions.
 They can be scaled up trivially.
I recommend using test datasets when getting started with a new machine learning algorithm or when developing a new test harness.
scikitlearn is a Python library for machine learning that provides functions for generating a suite of test problems.
In this tutorial, we will look at some examples of generating test problems for classification and regression algorithms.
Classification Test Problems
Classification is the problem of assigning labels to observations.
In this section, we will look at three classification problems: blobs, moons and circles.
Blobs Classification Problem
The make_blobs() function can be used to generate blobs of points with a Gaussian distribution.
You can control how many blobs to generate and the number of samples to generate, as well as a host of other properties.
The problem is suitable for linear classification problems given the linearly separable nature of the blobs.
The example below generates a 2D dataset of samples with three blobs as a multiclass classification prediction problem. Each observation has two inputs and 0, 1, or 2 class values.

# generate 2d classification dataset X, y = make_blobs(n_samples=100, centers=3, n_features=2) 
The complete example is listed below.

from sklearn.datasets.samples_generator import make_blobs from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_blobs(n_samples=100, centers=3, n_features=2) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = 0:‘red’, 1:‘blue’, 2:‘green’ fig, ax = pyplot.subplots() grouped = df.groupby(‘label’) for key, group in grouped: group.plot(ax=ax, kind=‘scatter’, x=‘x’, y=‘y’, label=key, color=colors[key]) pyplot.show() 
Running the example generates the inputs and outputs for the problem and then creates a handy 2D plot showing points for the different classes using different colors.
Note, your specific dataset and resulting plot will vary given the stochastic nature of the problem generator. This is a feature, not a bug.
We will use this same example structure for the following examples.
Moons Classification Problem
The make_moons() function is for binary classification and will generate a swirl pattern, or two moons.
You can control how noisy the moon shapes are and the number of samples to generate.
This test problem is suitable for algorithms that are capable of learning nonlinear class boundaries.
The example below generates a moon dataset with moderate noise.

# generate 2d classification dataset X, y = make_moons(n_samples=100, noise=0.1) 
The complete example is listed below.

from sklearn.datasets import make_moons from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_moons(n_samples=100, noise=0.1) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = 0:‘red’, 1:‘blue’ fig, ax = pyplot.subplots() grouped = df.groupby(‘label’) for key, group in grouped: group.plot(ax=ax, kind=‘scatter’, x=‘x’, y=‘y’, label=key, color=colors[key]) pyplot.show() 
Running the example generates and plots the dataset for review, again coloring samples by their assigned class.
Circles Classification Problem
The make_circles() function generates a binary classification problem with datasets that fall into concentric circles.
Again, as with the moons test problem, you can control the amount of noise in the shapes.
This test problem is suitable for algorithms that can learn complex nonlinear manifolds.
The example below generates a circles dataset with some noise.

# generate 2d classification dataset X, y = make_circles(n_samples=100, noise=0.05) 
The complete example is listed below.

from sklearn.datasets import make_circles from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_circles(n_samples=100, noise=0.05) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = 0:‘red’, 1:‘blue’ fig, ax = pyplot.subplots() grouped = df.groupby(‘label’) for key, group in grouped: group.plot(ax=ax, kind=‘scatter’, x=‘x’, y=‘y’, label=key, color=colors[key]) pyplot.show() 
Running the example generates and plots the dataset for review.
Regression Test Problems
Regression is the problem of predicting a quantity given an observation.
The make_regression() function will create a dataset with a linear relationship between inputs and the outputs.
You can configure the number of samples, number of input features, level of noise, and much more.
This dataset is suitable for algorithms that can learn a linear regression function.
The example below will generate 100 examples with one input feature and one output feature with modest noise.

# generate regression dataset X, y = make_regression(n_samples=100, n_features=1, noise=0.1) 
The complete example is listed below.

from sklearn.datasets import make_regression from matplotlib import pyplot # generate regression dataset X, y = make_regression(n_samples=100, n_features=1, noise=0.1) # plot regression dataset pyplot.scatter(X,y) pyplot.show() 
Running the example will generate the data and plot the X and y relationship, which, given that it is linear, is quite boring.
Extensions
This section lists some ideas for extending the tutorial that you may wish to explore.
 Compare Algorithms. Select a test problem and compare a suite of algorithms on the problem and report the performance.
 Scale Up Problem. Select a test problem and explore scaling it up, use progression methods to visualize the results, and perhaps explore model skill vs problem scale for a given algorithm.
 Additional Problems. The library provides a suite of additional test problems; write a code example for each to demonstrate how they work.
If you explore any of these extensions, I’d love to know.
Further Reading
This section provides more resources on the topic if you are looking to go deeper.
Summary
In this tutorial, you discovered test problems and how to use them in Python with scikitlearn.
Specifically, you learned:
 How to generate multiclass classification prediction test problems.
 How to generate binary classification prediction test problems.
 How to generate linear regression prediction test problems.
Do you have any questions?
Ask your questions in the comments below and I will do my best to answer.