# ResNet

ResNet is the most popular architecture for classifiers with over 20,000 citations. The authors were able to build a very deep, powerful network without running into the problem of vanishing gradients.

It is also often used as a backbone network for detection and segmentation models.

ResNet is often noted together with a number (e.g., ResNet18, ResNet50, ...). The number depicts the depth of the network, meaning how many layers it contains.

One would assume that stacking additional layers to a deep neural network would improve its performance because it could learn more features.

However, the original ResNet paper authors showed that this is not the case, but that just adding more layers to a network will actually lead to a performance decrease at a certain amount of layers because they simply pass over the same features again and again. Classification error with Cifar-10 data. Graphic taken from the orignal paper.

### Residual Blocks

As a solution, the authors proposed residual blocks, an “identity shortcut connection” that skips one or more layers, as shown in the following figure:

Due to this skip connection, the output of the layer is not the same now. Without using this skip connection, the input ‘x’ gets multiplied by the layer's weights, followed by adding a bias term. Next, this term goes through the activation function f(), and we get our output as H(x).

$H\left(x\right)=f\left(wx + b\right)$

Now with the introduction of the skip connection, the output is changed to:

$H\left(x\right)=f\left(wx + b\right)+x$

The skip connections in ResNet solve the problem of vanishing gradient in deep neural networks by allowing this alternate shortcut path for the gradient to flow through.

Using ResNet has significantly enhanced the performance of neural networks with more layers, as shown above.

### Depth of the resnet model

It is the number of layers in the ResNet model. In model playground, it can be chosen from:

### Weights

It's the weights to use for model initialization, and in Model Playground ResNet18 ImageNetweights are used.

    import torch

# or any of these variants
# model = torch.hub.load('pytorch/vision:v0.9.0', 'resnet34', pretrained=True)
# model = torch.hub.load('pytorch/vision:v0.9.0', 'resnet50', pretrained=True)
# model = torch.hub.load('pytorch/vision:v0.9.0', 'resnet101', pretrained=True)
# model = torch.hub.load('pytorch/vision:v0.9.0', 'resnet152', pretrained=True)
model.eval()
# Sample Execution
import urllib
url, filename = ("https://github.com/pytorch/hub/raw/master/images/dog.jpg", "dog.jpg")
try: urllib.URLopener().retrieve(url, filename)
except: urllib.request.urlretrieve(url, filename)

from PIL import Image
from torchvision import transforms
input_image = Image.open(filename)
preprocess = transforms.Compose([
transforms.Resize(256),
transforms.CenterCrop(224),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),
])
input_tensor = preprocess(input_image)
input_batch = input_tensor.unsqueeze(0) # create a mini-batch as expected by the model

# move the input and model to GPU for speed if available
if torch.cuda.is_available():
input_batch = input_batch.to('cuda')
model.to('cuda')

output = model(input_batch)

# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
print(output)
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output, dim=0)
print(probabilities)

!wget https://raw.githubusercontent.com/pytorch/hub/master/imagenet_classes.txt

with open("imagenet_classes.txt", "r") as f:
categories = [s.strip() for s in f.readlines()]
# Show top categories per image
top5_prob, top5_catid = torch.topk(probabilities, 5)
for i in range(top5_prob.size(0)):
print(categories[top5_catid[i]], top5_prob[i].item())

#Model structure  Top-1 error Top-5 error
#resnet18           30.24      10.92
#resnet34           26.70      8.58
#resnet50           23.85      7.13
#resnet101            22.63        6.44
#resnet152            21.69        5.94


Last updated on Dec 27, 2022

## Get AI confident. Start using Hasty today.

Automate 90% of the work, reduce your time to deployment by 40%, and replace your whole ML software stack with our platform.