12  Splines

Splines, is one way to represent a curve that makes it useful in modeling content, since it allows us to model non-linear relationships between predictors and outcomes. This is a trained method.

Being able to transform a numeric variable that has a non-linear relationship with the outcome into one or more variables that do have linear relationships with the outcome is of great importance, as many models wouldn’t be able to work with these types of variables effectively themselves. Below is a toy example of one such variable

Scatter chart. Predictor along the x-axis and outcome along the y-axis. The data has some wiggliness to it, but it follows a curve. You would not be able to fit a straight line to this data.
Figure 12.1: Non-linear relationship between predictor and outcome.

Here we have a non-linear relationship. It is a fairly simple one, the outcome is high when the predictor takes values between 25 and 50, and outside the ranges, it takes over values. Given that this is a toy example, we do not have any expert knowledge regarding what we expect the relationship to be outside this range. The trend could go back up, it could go down or flatten out. We don’t know.

As we saw in Chapter 11, one way to deal with this non-linearity is to chop up the predictor and emit indicators for which region the values take. While this works, we are losing quite a lot of detail by the rounding that occurs. This is where splines come in. Imagine that instead of indicators for whether a value is within a range, we have a set of functions that gives each predictor value a set of values, related to its location in the distribution.

Facetted line chart. Predictor along the x-axis, value along the y-axis. Each of the curves starts at 0, goes smoothly, and then down to zero. The highpoint for each curve goes further to the right for each curve shown.
Figure 12.2: Each part of the spline detects a part of the data set.

Above we see an example of a Basis function that creates 6 features. The curves represent the area where they are β€œactivated”. So if the predictor has a value of 15 then the first basis function returns 0.40, the second basis function returns 0.20 and so one, with the last basis function returning 0.00 since it is all flat over there.

This is a trained method as the location and shape of these functions are determined by the distribution of the variables we are trying to apply the spline to.

So in this example, we are taking 1 numeric variable and turning it into 6 numeric variables.

Table 12.1: Spline values for different values of the predictor
predictor Spline Feature 1 Spline Feature 2 Spline Feature 3 Spline Feature 4 Spline Feature 5 Spline Feature 6
0 0.01 0.00 0.00 0.00 0.00 0.00
10 0.37 0.12 0.02 0.00 0.00 0.00
35 0.19 0.30 0.26 0.14 0.04 0.01
50 0.06 0.17 0.28 0.27 0.15 0.05
80 0.00 0.01 0.04 0.13 0.29 0.35

This spline is set up in such a way, that each spline function signals if the values are close to a given region. This way we have a smooth transition throughout the distribution.

If we take a look at how this plays out when we bring back the outcome we can look at this visualization

Facetted scatter chart. Predictor along the x-axis, outcome along the y-axis. Each of the facets shows the same non-linear relationship between predictor and outcome. Color is used to show how each spline term highlights a different part of the predictor. The highlight goes further to the right for each facet.
Figure 12.3: Each part of the spline detects a part of the data set.

and we have that since the different spline features highlight different parts of the predictor, we have that at least some of them are useful when we look at the relationship between the predictor and the outcome.

It is important to point out that this transformation only uses the predictor variable to do its calculations. And the fact that it works in a modeling sense is that the outcome predictor relationship in this case, and many real-life cases, can helpfully be explained by β€œthe predictor value has these values”.

Facetted scatter chart. Spline value along the x-axis, outcome along the y-axis. Each facet shows the relationship between one of the spline terms and the outcome. Some of them are non-linear, and a couple of them are fairly linear. A fitted line is overlaid in blue.
Figure 12.4: Some spline terms have a better relationship to the outcome than others.

As we see in the above visualization, some of these new predictors are not much better than the original. But a couple of them do appear to work pretty well, especially the 3rd one. Depending on which model we use, having these 6 variables is gonna give us higher performance than using the original variable alone.

One thing to note is that you will get back correlated features when using splines. Some values of the predictor will influence multiple of the spline features as the spline functions overlap. This is expected but is worth noting. If you are using a model type that doesn’t handle correlated features well, then you should take a look at the methods outlined in Chapter 80 for ways to deal with correlated features.

Correlation chart. The spline basis features are lined up one after another. Neighboring features show high correlation, features 2 apart are slightly correlated, and other features are anti-correlated.
Figure 12.5: Neighboring features are highly correlated and anti-correlated with far away features.

Lastly, the above spline functions you saw were called B-splines, but they are not the only kind of splines you can use.

4 charts in a grid. Each represents a different type of spline. The C-splines here are all increasing at different rates of change. The M-splines appear to have a sigmoidal shape, starting at 0 and ending at 1. The natural splines look very similar to the basic splines we saw earlier. And the last chart shows a periodic b-spline. These splines are the same kind as earlier, but they have been modified to repeat at a specific interval.
Figure 12.6: Neighboring features are highly correlated and anti-correlated with far away features.

Above we see several different kinds of splines. As we can see they are all trying to do different things. You generally can’t go too wrong by picking any of them, but knowing the data can help guide which of them you should use. The M-splines intuitively can be seen as threshold features. The periodic example is also interesting. Many of the types of splines can be formulated to work periodically. This can be handy for data that has a naturally periodic nature to them.

Below is a chart of how well using splines works when using it on our toy example. Since the data isn’t that complicated, a small deg_free is sufficient to fit the data well.

Scatter chart. Predictor along the x-axis and outcome along the y-axis. The data has some wiggliness to it, but it follows a curve. You would not be able to fit a straight line to this data. 4 spline fits are plotted to fit the data. deg_free = 5 appears to fit well without overfitting, the rest are overfitting the data.
Figure 12.7: All the splines follow the data well, the higher degrees appear to overfit quite a bit.

12.2 Pros and Cons

12.2.1 Pros

  • Works fast computationally
  • Good performance compared to binning
  • is good at handling continuous changes in predictors

12.2.2 Cons

  • arguably less interpretable than binning
  • creates correlated features
  • can produce a lot of variables
  • have a hard time modeling sudden changes in distributions

12.3 R Examples

We will be using the ames data set for these examples.

library(recipes)
library(modeldata)

ames |>
  select(Lot_Area, Year_Built)
# A tibble: 2,930 Γ— 2
   Lot_Area Year_Built
      <int>      <int>
 1    31770       1960
 2    11622       1961
 3    14267       1958
 4    11160       1968
 5    13830       1997
 6     9978       1998
 7     4920       2001
 8     5005       1992
 9     5389       1995
10     7500       1999
# β„Ή 2,920 more rows

{recipes} provides a number of steps to perform spline operations, each of them starting with step_spline_. Let us use a B-spline and a M-spline as examples here:

log_rec <- recipe(~ Lot_Area + Year_Built, data = ames) |>
  step_spline_b(Lot_Area) |>
  step_spline_monotone(Year_Built)

log_rec |>
  prep() |>
  bake(new_data = NULL) |>
  glimpse()
Rows: 2,930
Columns: 20
$ Lot_Area_01   <dbl> 0.000000000, 0.000000000, 0.000000000, 0.000000000, 0.00…
$ Lot_Area_02   <dbl> 0.000000e+00, 0.000000e+00, 0.000000e+00, 0.000000e+00, …
$ Lot_Area_03   <dbl> 0.0000000, 0.0000000, 0.0000000, 0.0000000, 0.0000000, 0…
$ Lot_Area_04   <dbl> 0.0000000000, 0.0000000000, 0.0000000000, 0.0000000000, …
$ Lot_Area_05   <dbl> 0.0000000000, 0.0000000000, 0.0000000000, 0.0078526499, …
$ Lot_Area_06   <dbl> 0.000000000, 0.283603274, 0.000000000, 0.507327055, 0.00…
$ Lot_Area_07   <dbl> 0.73399934, 0.71382012, 0.96474057, 0.48414256, 0.971047…
$ Lot_Area_08   <dbl> 2.408258e-01, 2.576602e-03, 3.503161e-02, 6.777374e-04, …
$ Lot_Area_09   <dbl> 2.444735e-02, 3.441535e-09, 2.277849e-04, 0.000000e+00, …
$ Lot_Area_10   <dbl> 7.274651e-04, 0.000000e+00, 3.035128e-08, 0.000000e+00, …
$ Year_Built_01 <dbl> 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,…
$ Year_Built_02 <dbl> 1.0000000, 1.0000000, 1.0000000, 1.0000000, 1.0000000, 1…
$ Year_Built_03 <dbl> 0.9991483, 0.9995281, 0.9977527, 1.0000000, 1.0000000, 1…
$ Year_Built_04 <dbl> 0.9041892, 0.9210803, 0.8649607, 0.9884577, 1.0000000, 1…
$ Year_Built_05 <dbl> 0.20672563, 0.24041792, 0.14882796, 0.54043388, 0.999999…
$ Year_Built_06 <dbl> 4.792231e-04, 1.169978e-03, 2.995144e-05, 3.881537e-02, …
$ Year_Built_07 <dbl> 0.000000e+00, 0.000000e+00, 0.000000e+00, 4.491099e-07, …
$ Year_Built_08 <dbl> 0.00000000, 0.00000000, 0.00000000, 0.00000000, 0.221125…
$ Year_Built_09 <dbl> 0.000000e+00, 0.000000e+00, 0.000000e+00, 0.000000e+00, …
$ Year_Built_10 <dbl> 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0,…

We can set the deg_free argument to specify how many spline features we want for each of the splines.

log_rec <- recipe(~ Lot_Area + Year_Built, data = ames) |>
  step_spline_b(Lot_Area, deg_free = 3) |>
  step_spline_monotone(Year_Built, deg_free = 4)

log_rec |>
  prep() |>
  bake(new_data = NULL) |>
  glimpse()
Rows: 2,930
Columns: 7
$ Lot_Area_1   <dbl> 0.31422525, 0.13110895, 0.16045431, 0.12580964, 0.1557218…
$ Lot_Area_2   <dbl> 0.0521839123, 0.0066461383, 0.0103524317, 0.0060782666, 0…
$ Lot_Area_3   <dbl> 2.888757e-03, 1.123014e-04, 2.226446e-04, 9.788684e-05, 2…
$ Year_Built_1 <dbl> 0.9827669, 0.9841047, 0.9798397, 0.9914201, 0.9999212, 0.…
$ Year_Built_2 <dbl> 0.8614458, 0.8686207, 0.8464715, 0.9129756, 0.9968924, 0.…
$ Year_Built_3 <dbl> 0.5411581, 0.5539857, 0.5156159, 0.6440229, 0.9532064, 0.…
$ Year_Built_4 <dbl> 0.1653539, 0.1729990, 0.1508264, 0.2341901, 0.6731684, 0.…

These steps have more arguments, so we can change other things. The B-splines created by step_spline_b() default to cubic splines, but we can change that by specifying which polynomial degree with want with the degree argument.

log_rec <- recipe(~ Lot_Area + Year_Built, data = ames) |>
  step_spline_b(Lot_Area, deg_free = 3, degree = 1) |>
  step_spline_monotone(Year_Built, deg_free = 4)

log_rec |>
  prep() |>
  bake(new_data = NULL) |>
  glimpse()
Rows: 2,930
Columns: 7
$ Lot_Area_1   <dbl> 0.00000000, 0.00000000, 0.00000000, 0.00000000, 0.0000000…
$ Lot_Area_2   <dbl> 0.89690903, 0.99540159, 0.98247163, 0.99766006, 0.9846078…
$ Lot_Area_3   <dbl> 0.103090969, 0.004598405, 0.017528365, 0.002339940, 0.015…
$ Year_Built_1 <dbl> 0.9827669, 0.9841047, 0.9798397, 0.9914201, 0.9999212, 0.…
$ Year_Built_2 <dbl> 0.8614458, 0.8686207, 0.8464715, 0.9129756, 0.9968924, 0.…
$ Year_Built_3 <dbl> 0.5411581, 0.5539857, 0.5156159, 0.6440229, 0.9532064, 0.…
$ Year_Built_4 <dbl> 0.1653539, 0.1729990, 0.1508264, 0.2341901, 0.6731684, 0.…

12.4 Python Examples

We are using the ames data set for examples. {sklearn} provided the SplineTransformer() method we can use.

from feazdata import ames
from sklearn.compose import ColumnTransformer
from sklearn.preprocessing import SplineTransformer

ct = ColumnTransformer(
    [('spline', SplineTransformer(), ['Lot_Area'])], 
    remainder="passthrough")

ct.fit(ames)
ColumnTransformer(remainder='passthrough',
                  transformers=[('spline', SplineTransformer(), ['Lot_Area'])])
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.
ct.transform(ames)
      spline__Lot_Area_sp_0  ...  remainder__Latitude
0                     0.013  ...               42.054
1                     0.088  ...               42.053
2                     0.072  ...               42.053
3                     0.090  ...               42.051
4                     0.075  ...               42.061
...                     ...  ...                  ...
2925                  0.112  ...               41.989
2926                  0.105  ...               41.988
2927                  0.095  ...               41.987
2928                  0.098  ...               41.991
2929                  0.100  ...               41.989

[2930 rows x 80 columns]