Skip to content

Commit ab97ec0

Browse files
authored
refactor: unify the words into one (#338)
1 parent e55b003 commit ab97ec0

27 files changed

+39
-39
lines changed

facebookresearch_WSL-Images_resnext.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ if torch.cuda.is_available():
6565

6666
with torch.no_grad():
6767
output = model(input_batch)
68-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
68+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6969
print(output[0])
7070
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7171
print(torch.nn.functional.softmax(output[0], dim=0))

facebookresearch_semi-supervised-ImageNet1K-models_resnext.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ if torch.cuda.is_available():
7373

7474
with torch.no_grad():
7575
output = model(input_batch)
76-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
76+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
7777
print(output[0])
7878
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7979
print(torch.nn.functional.softmax(output[0], dim=0))

nvidia_deeplearningexamples_efficientnet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
5555
print(f'Using {device} for inference')
5656
```
5757

58-
Load the model pretrained on IMAGENET dataset.
58+
Load the model pretrained on ImageNet dataset.
5959

6060
You can choose among the following models:
6161

nvidia_deeplearningexamples_gpunet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ print(f'Using {device} for inference')
5454
```
5555

5656
### Load Pretrained model
57-
Loads NVIDIA GPUNet-0 model by default pre-trained on IMAGENET dataset. You can switch the default pre-trained model loading from GPUNet-0 to one of the following models listed below.
57+
Loads NVIDIA GPUNet-0 model by default pre-trained on ImageNet dataset. You can switch the default pre-trained model loading from GPUNet-0 to one of the following models listed below.
5858

5959
The model architecture is visible as output of the loaded model. For details architecture and latency info please refer to [architecture section](https://github.com/NVIDIA/DeepLearningExamples/tree/torchhub/PyTorch/Classification/GPUNet#model-architecture) in the original repo and Table#[3](https://arxiv.org/pdf/2205.00841.pdf) in the CVPR-2022 paper, respectively.
6060

nvidia_deeplearningexamples_resnet50.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
5959
print(f'Using {device} for inference')
6060
```
6161

62-
Load the model pretrained on IMAGENET dataset.
62+
Load the model pretrained on ImageNet dataset.
6363
```python
6464
resnet50 = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_resnet50', pretrained=True)
6565
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')

nvidia_deeplearningexamples_resnext.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
6464
print(f'Using {device} for inference')
6565
```
6666

67-
Load the model pretrained on IMAGENET dataset.
67+
Load the model pretrained on ImageNet dataset.
6868
```python
6969
resneXt = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_resneXt')
7070
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')

nvidia_deeplearningexamples_se-resnext.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
6464
print(f'Using {device} for inference')
6565
```
6666

67-
Load the model pretrained on IMAGENET dataset.
67+
Load the model pretrained on ImageNet dataset.
6868
```python
6969
resneXt = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_se_resnext101_32x4d')
7070
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')

pytorch_vision_alexnet.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -85,11 +85,11 @@ for i in range(top5_prob.size(0)):
8585

8686
AlexNet competed in the ImageNet Large Scale Visual Recognition Challenge on September 30, 2012. The network achieved a top-5 error of 15.3%, more than 10.8 percentage points lower than that of the runner up. The original paper's primary result was that the depth of the model was essential for its high performance, which was computationally expensive, but made feasible due to the utilization of graphics processing units (GPUs) during training.
8787

88-
The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
88+
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.
8989

9090
| Model structure | Top-1 error | Top-5 error |
9191
| --------------- | ----------- | ----------- |
92-
| alexnet | 43.45 | 20.91 |
92+
| AlexNet | 43.45 | 20.91 |
9393

9494
### References
9595

pytorch_vision_densenet.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ if torch.cuda.is_available():
6363

6464
with torch.no_grad():
6565
output = model(input_batch)
66-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
66+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6767
print(output[0])
6868
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6969
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -89,7 +89,7 @@ for i in range(top5_prob.size(0)):
8989

9090
Dense Convolutional Network (DenseNet), connects each layer to every other layer in a feed-forward fashion. Whereas traditional convolutional networks with L layers have L connections - one between each layer and its subsequent layer - our network has L(L+1)/2 direct connections. For each layer, the feature-maps of all preceding layers are used as inputs, and its own feature-maps are used as inputs into all subsequent layers. DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
9191

92-
The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
92+
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.
9393

9494
| Model structure | Top-1 error | Top-5 error |
9595
| --------------- | ----------- | ----------- |

pytorch_vision_ghostnet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_googlenet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_hardnet.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ if torch.cuda.is_available():
6363

6464
with torch.no_grad():
6565
output = model(input_batch)
66-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
66+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6767
print(output[0])
6868
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6969
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -95,7 +95,7 @@ were designed for comparing with MobileNet).
9595

9696
Here we have the 4 versions of hardnet models, which contains 39, 68, 85 layers
9797
w/ or w/o Depthwise Separable Conv respectively.
98-
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
98+
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.
9999

100100
| Model structure | Top-1 error | Top-5 error |
101101
| --------------- | ----------- | ----------- |

pytorch_vision_ibnnet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_inception_v3.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ layout: hub_detail
33
background-class: hub-background
44
body-class: hub
55
title: Inception_v3
6-
summary: Also called GoogleNetv3, a famous ConvNet trained on Imagenet from 2015
6+
summary: Also called GoogleNetv3, a famous ConvNet trained on ImageNet from 2015
77
category: researchers
88
image: inception_v3.png
99
author: Pytorch Team
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -85,7 +85,7 @@ for i in range(top5_prob.size(0)):
8585

8686
Inception v3: Based on the exploration of ways to scale up networks in ways that aim at utilizing the added computation as efficiently as possible by suitably factorized convolutions and aggressive regularization. We benchmark our methods on the ILSVRC 2012 classification challenge validation set demonstrate substantial gains over the state of the art: 21.2% top-1 and 5.6% top-5 error for single frame evaluation using a network with a computational cost of 5 billion multiply-adds per inference and with using less than 25 million parameters. With an ensemble of 4 models and multi-crop evaluation, we report 3.5% top-5 error on the validation set (3.6% error on the test set) and 17.3% top-1 error on the validation set.
8787

88-
The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
88+
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.
8989

9090
| Model structure | Top-1 error | Top-5 error |
9191
| --------------- | ----------- | ----------- |

pytorch_vision_meal_v2.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ if torch.cuda.is_available():
6767

6868
with torch.no_grad():
6969
output = model(input_batch)
70-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
70+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
7171
print(output[0])
7272
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7373
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_mobilenet_v2.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_once_for_all.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,7 @@ if torch.cuda.is_available():
105105

106106
with torch.no_grad():
107107
output = model(input_batch)
108-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
108+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
109109
print(output[0])
110110
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
111111
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_proxylessnas.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ if torch.cuda.is_available():
6161

6262
with torch.no_grad():
6363
output = model(input_batch)
64-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
64+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6565
print(output[0])
6666
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6767
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_resnest.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ if torch.cuda.is_available():
6262

6363
with torch.no_grad():
6464
output = model(input_batch)
65-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
65+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6666
print(output[0])
6767
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6868
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_resnet.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ if torch.cuda.is_available():
6464

6565
with torch.no_grad():
6666
output = model(input_batch)
67-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
67+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6868
print(output[0])
6969
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7070
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -91,7 +91,7 @@ for i in range(top5_prob.size(0)):
9191
Resnet models were proposed in "Deep Residual Learning for Image Recognition".
9292
Here we have the 5 versions of resnet models, which contains 18, 34, 50, 101, 152 layers respectively.
9393
Detailed model architectures can be found in Table 1.
94-
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
94+
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.
9595

9696
| Model structure | Top-1 error | Top-5 error |
9797
| --------------- | ----------- | ----------- |

pytorch_vision_resnext.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ if torch.cuda.is_available():
6161

6262
with torch.no_grad():
6363
output = model(input_batch)
64-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
64+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6565
print(output[0])
6666
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6767
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -90,7 +90,7 @@ for i in range(top5_prob.size(0)):
9090
Resnext models were proposed in [Aggregated Residual Transformations for Deep Neural Networks](https://arxiv.org/abs/1611.05431).
9191
Here we have the 2 versions of resnet models, which contains 50, 101 layers repspectively.
9292
A comparison in model archetechure between resnet50 and resnext50 can be found in Table 1.
93-
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
93+
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.
9494

9595
| Model structure | Top-1 error | Top-5 error |
9696
| ----------------- | ----------- | ----------- |

pytorch_vision_shufflenet_v2.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ layout: hub_detail
33
background-class: hub-background
44
body-class: hub
55
title: ShuffleNet v2
6-
summary: An efficient ConvNet optimized for speed and memory, pre-trained on Imagenet
6+
summary: An efficient ConvNet optimized for speed and memory, pre-trained on ImageNet
77
category: researchers
88
image: shufflenet_v2_1.png
99
author: Pytorch Team
@@ -59,7 +59,7 @@ if torch.cuda.is_available():
5959

6060
with torch.no_grad():
6161
output = model(input_batch)
62-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
62+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6363
print(output[0])
6464
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6565
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_snnmlp.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ if torch.cuda.is_available():
6262

6363
with torch.no_grad():
6464
output = model(input_batch)
65-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
65+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6666
print(output[0])
6767
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6868
print(torch.nn.functional.softmax(output[0], dim=0))

pytorch_vision_squeezenet.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ if torch.cuda.is_available():
6161

6262
with torch.no_grad():
6363
output = model(input_batch)
64-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
64+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6565
print(output[0])
6666
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6767
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -90,7 +90,7 @@ Model `squeezenet1_0` is from the [SqueezeNet: AlexNet-level accuracy with 50x f
9090
Model `squeezenet1_1` is from the [official squeezenet repo](https://github.com/DeepScale/SqueezeNet/tree/master/SqueezeNet_v1.1).
9191
It has 2.4x less computation and slightly fewer parameters than `squeezenet1_0`, without sacrificing accuracy.
9292

93-
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
93+
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.
9494

9595
| Model structure | Top-1 error | Top-5 error |
9696
| --------------- | ----------- | ----------- |

pytorch_vision_vgg.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ layout: hub_detail
33
background-class: hub-background
44
body-class: hub
55
title: vgg-nets
6-
summary: Award winning ConvNets from 2014 Imagenet ILSVRC challenge
6+
summary: Award winning ConvNets from 2014 ImageNet ILSVRC challenge
77
category: researchers
88
image: vgg.png
99
author: Pytorch Team
@@ -67,7 +67,7 @@ if torch.cuda.is_available():
6767

6868
with torch.no_grad():
6969
output = model(input_batch)
70-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
70+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
7171
print(output[0])
7272
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7373
probabilities = torch.nn.functional.softmax(output[0], dim=0)

pytorch_vision_wide_resnet.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ if torch.cuda.is_available():
6262

6363
with torch.no_grad():
6464
output = model(input_batch)
65-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
65+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
6666
print(output[0])
6767
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
6868
probabilities = torch.nn.functional.softmax(output[0], dim=0)

simplenet.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ if torch.cuda.is_available():
6767

6868
with torch.no_grad():
6969
output = model(input_batch)
70-
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
70+
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
7171
print(output[0])
7272
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
7373
probabilities = torch.nn.functional.softmax(output[0], dim=0)
@@ -94,7 +94,7 @@ for i in range(top5_prob.size(0)):
9494
SimpleNet models were proposed in "Lets Keep it simple, Using simple architectures to outperform deeper and more complex architectures".
9595
Here we have the 8 versions of simplenet models, which contains 1.5m, 3.2m, 5.7m and 9.5m parameters respectively.
9696
Detailed model architectures can be found in Table 1 and Table 2.
97-
Their 1-crop errors on imagenet dataset with pretrained models are listed below.
97+
Their 1-crop errors on ImageNet dataset with pretrained models are listed below.
9898

9999
The m2 variants
100100

0 commit comments

Comments
 (0)