Weight Decay

Great question! Weight decay ideally should be mathematically equivalent to L2 Regularization. But be carefully when you implement it in code, this article indicates that different frameworks might be slightly different.

1 Like

You are maybe right.
Does that mean we can use w and b as bias penalty at the same time or separately?

I think it’s possible. Can we add? @goldpiggy

Hi @StevenJokes and @Steven_Hearnt, great questions! L1 regularization is applicable if you specify how to handle the in-differentiable case (x=0) .

Hi @StevenJokes, I might be late to the party but would like to take a stab at it from a theoretical point of view rather than practical (which has already been covered). If we regularize the bias term (b) by adding the penalty term b^2, the network would end up learning a very small value of bias term in the case where we constraint the model a lot (i.e. lambda is very big). In such a case, model would not have any average value and thus it would be predicting some value near to zero every time. So, in order to avoid such a scenario, maybe bias term isn’t regularized in the last layer.

1 Like

Why
“the network would end up learning a very small value of bias term”?

What does your “lambda” mean?

Lambda is the constraint imposed on the L2 norm in the loss function. It is defined in 4.5.2. If we set lambda to a very large number and include bias term in the L2 norm as well, gradient descent will set bias term to an extremely small number as well. All of this is mentioned in the section 4.5


Is it?
@kusur

Yep, If you include bias term in the l2_penalty, you set this parameter to a huge value, and train the model, you will notice that bias itself becomes very small. This removes the affine aspect of the transformation in the neural network

…still can’t understand “the affine aspect”.
Some related papers? @kusur

In 4.5.1

The technique is motivated by the basic intuition that among all functions f, the function f=0 (assigning the value 0 to all inputs) is in some sense the simplest , and that we can measure the complexity of a function by its distance from zero.

Why is f = 0 considered the simplest? What does it mean for a function to be simple/complex in this context? Why is a function’s distance from zero, a measure of its complexity?

The most common method for ensuring a small weight vector is to add its norm as a penalty term to the problem of minimizing the loss

What does a weight vector being “small” mean? Like, number of elements in the weight vector (length of the weight vector)?

Hi @tinkuge, great questions!

f = 0 means no regularization, that’s why it is the simplest (no computation here :wink: ). Check the Lp distance and it should give you better answer.

“small” refers each weight element’s value is small, such as following within range [-1, 1].

Any idea on how do I implement an L1 regularized weight decay using

torch.optim.SGD

As per PyTorch documentation torch.optim — PyTorch 2.1 documentation it seems like the weight_decay parameter only implements L2 regularization.

It looks like l1 norm has to be implemented using lower-level api

any one has idea about question 6? thx

Is there someone can share the idea about question 6? thx

It seem that a model with no regularization is better than a model with regularization in my case. Although the norm of the weights is significantly lower in the model with regularization, the plots are better in the unregularized one. Another problem is that the loss in my validation data is lower than the training loss. Here is my code:

import torch
from torch.utils.data import TensorDataset, DataLoader
import matplotlib.pyplot as plt

def synthetic_data(true_w, true_b, n):
    X = torch.normal(0, 0.01, (n, len(true_w)))
    y = torch.matmul(X, true_w) + true_b
    y += torch.normal(0, 0.001, y.shape)
    return X, y.reshape((-1, 1))

def load_array(array, batch_size, is_train=False):
    ''' Change array to a torch data iterator '''
    dataset = TensorDataset(*array)
    return DataLoader(dataset, batch_size=batch_size, shuffle=is_train)

n_train, n_test, n_inputs, batch_size = 20, 100, 200, 5

true_w = torch.ones((n_inputs, 1)) * 0.01
true_b = 0.05

train_data, train_labels = synthetic_data(true_w, true_b, n_train)
train_iter = load_array((train_data, train_labels), batch_size, is_train=True)
test_data, test_labels = synthetic_data(true_w, true_b, n_test)
test_iter = load_array((test_data, test_labels), batch_size)

# build the model
def linreg(X):
    return X@W + b

# initialize the weights
W = torch.normal(0, 1, (n_inputs, 1), requires_grad=True)
b = torch.zeros(1, requires_grad=True)

# define the loss function
def MSELoss(y_hat, y):
    return (y_hat - y) ** 2 / 2

# define the L2 regularization term
def L2_penality(W):
    return torch.norm(W) / 2

# Define the optimizer
def SGD(params, batch_size):
    with torch.no_grad():
        for param in params:
            param -= lr * param.grad / batch_size
            param.grad.zero_()

class Accumulator:
    def __init__(self, n):
        self.data = [0.0] * n

    def add(self, *args):
        self.data = [a + float(b) for a, b in zip(self.data, args)]

    def reset(self):
        self.data = [0.0] * len(self.data)

    def __getitem__(self, index):
        return self.data[index]

def evaluate(data_iter):
    metric = Accumulator(2)
    with torch.no_grad():
        for X, y in data_iter:
            l = MSELoss(linreg(X), y)
            metric.add(float(l.sum().item()), len(y))
    return metric[0] / metric[1]
        
# Write the training loop
epochs, lr = 100, 0.003
weight_decay = 0

train_loss = []
val_loss = []

train_metric = Accumulator(2)
for epoch in range(epochs):
    for X, y in train_iter:
        l = MSELoss(linreg(X), y) + weight_decay * L2_penality(W)
        l.sum().backward()
        SGD([W, b], batch_size)
        train_metric.add(float(l.sum().item()), len(y))
        
    train_loss.append(train_metric[0] / train_metric[1])
    train_metric.reset()

    # test the validation loss
    l = evaluate(test_iter)
    val_loss.append(l)

##    print("Epoch {}/{}   loss:  {:.5f}    val_loss:  {:.5f}".format(epoch+1, epochs, train_loss[-1], val_loss[-1]))
print("Weight Norm: ", torch.norm(W).item())


plt.plot(range(len(train_loss)), train_loss, label='Train loss')
plt.plot(range(len(val_loss)), val_loss, label='Validation loss')
plt.xlabel("Epochs")
plt.ylabel("Loss")
plt.title("Training and validation loss")
plt.legend()
plt.show()

First post! :slight_smile: Thanks for this amazing book.

Also, I tried to plot train and test loss vs the choice of lambda. We can see that the training loss increases and test loss first decreases then stays mostly constant (see exact values below)

So, is the first value where test loss decreases before staying stagnant (i.e. lambda=1) is the best regularization parameter for this problem @goldpiggy can you suggest if this is a ok heuristic?

srno lambda train_loss test_loss
0 0 2.360348e-13 1.914223
1 1 3.561247e-04 0.035403
2 2 1.136713e-03 0.018138
3 3 2.143079e-03 0.017133
4 4 3.240126e-03 0.017512
5 5 4.373294e-03 0.017590
6 6 5.604185e-03 0.017622
7 7 6.343381e-03 0.017295
8 8 7.639927e-03 0.017984
9 9 8.753147e-03 0.017798
10 10 9.767077e-03 0.017654

#https://stackoverflow.com/questions/42704283/adding-l1-l2-regularization-in-pytorch

def l1_norm_with_abs(w):
    return torch.abs(w).sum()

Exercises and my answers

  1. Experiment with the value of λ in the estimation problem in this section. Plot training and

test accuracy as a function of λ. What do you observe?

  • as we increase wd it is making better train and test loss graph

  1. Use a validation set to find the optimal value of λ. Is it really the optimal value? Does this

matter?

  • the minimum loss is 0.02 at wd = 21-5 = 16. It is th eoptimal value for the validation set.
  1. What would the update equations look like if instead of ∥w∥

2 we used ∑

i

|wi

| as our penalty

of choice (L1 regularization)?

  • It is strangely coming as linear

# training for l1 norm

model = nn.Sequential(nn.Linear(num_inputs, 1))

for param in model.parameters():

    param.data.normal_()

    print(l1_norm(param.detach()))

    

loss = nn.MSELoss()

optimizer = torch.optim.SGD(model.parameters(), lr = 0.03)

valid_loss_array = []

for epoch in range(20):

    current_loss = 0

    current_number = 0

    for X, y in valid_dataloader:

        l = loss(model(X),y) - l1_norm_with_abs(model[0].weight)

        

        optimizer.zero_grad()

        l.backward()

        optimizer.step()

        

        current_loss += l.detach()

        current_number += len(y)

    

    valid_loss_array.append(current_loss/current_number)

plt.plot(range(20), valid_loss_array)

plt.grid(True)

plt.show()

(cant upload the graph here only onw pic per post)

  1. We know that ∥w∥

2 = w⊤w. Can you find a similar equation for matrices (see the Frobenius

norm in Section 2.3.10)?

  • This is Frobenius norm: f(αx) = |α|f(x). I dont understand.
  1. Review the relationship between training error and generalization error. In addition to

weight decay, increased training, and the use of a model of suitable complexity, what other

ways can you think of to deal with overfitting?

  • larger data size, dropouts
  1. In Bayesian statistics we use the product of prior and likelihood to arrive at a posterior via

P(w | x) ∝ P(x | w)P(w). How can you identify P(w) with regularization?

  • you ask good question but me no understand. how to get to p of w.