I want to train the YOLO v8 in transfer learning on my custom dataset.
I have different classes than the base training on the COCO dataset.
Yet I don't want to learn again the feature extraction. Hence I though following the Ultralytics YOLOv8 Docs - Train.
Yet, When I train on my small dataset I want to freeze the backbone.
How can I do that?
I looked at the documentation and couldn't find how to do so.
You can do the following
def freeze_layer(trainer):
    model = trainer.model
    num_freeze = 10
    print(f"Freezing {num_freeze} layers")
    freeze = [f'model.{x}.' for x in range(num_freeze)]  # layers to freeze 
    for k, v in model.named_parameters(): 
        v.requires_grad = True  # train all layers 
        if any(x in k for x in freeze): 
            print(f'freezing {k}') 
            v.requires_grad = False 
    print(f"{num_freeze} layers are freezed.")
Then add this function as a custom callback function to the model
model = YOLO("yolov8x.pt") 
model.add_callback("on_train_start", freeze_layer)
model.train(data="./dataset.yaml")
Original answer is provided in one of the issues in ultralytics repo Freezing layers yolov8 #793
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With