Model¶
A model represents all the information about a Machine Learning model.
Model is represented by a dict with these common fields:
Name
- the only required field, with the name of the model.Metadata
- a dict of Metadata, or a link to a file (yml/json) containing it.Results
- a list of Result dicts.Paper
- URL to the paper, or a dict with paper title, url andCode
- link to code snippet snippetWeights
- link to download the pretrained weightsConfig
- link to the config file used for trainingREADME
- the content of, or a link to the README.md file for the modelImage
- path or URL to an image for this modelIn Collection
- name of the Collection to which this model belongs.
The fields above will be automatically recognized by model-index, but you can also add any number of custom fields to it.
All field names are case-insensitive.
Metadata¶
Metadata is a dict of common and custom metadata. The common fields are:
FLOPs
- The number of FLOPs of the modelParameters
- The total number of parameters of the modelEpochs
- Number of training epochsBatch Size
- Input batch sizeTraining Data
- Names of dataset on which the models is trained onTraining Techniques
- A list of training techniques (for the full list see Methods on Papers with Code)Training Resources
- The hardware used for trainingTraining Time
- How many hours or days it takes to train.Architecture
- A list of architectural features of the model (for the full list see Methods on Papers with Code)
You can also add any other custom field that is specific to your model.
Result¶
Result is a dict capturing the evaluation results of the model. It has these fields:
Task
- Name of the task (for full see Benchmarks on Papers with Code)Dataset
- Name of the dataset (for full see Datasets on Papers with Code)Metrics
- a list of dictionaries with metrics. For relevant metrics consult the see Benchmarks on Papers with Code.
A full example¶
An example of the full model dict is shown below:
Name: Inception v3
Metadata:
FLOPs: 5731284192
Parameters: 23834568
Epochs: 90
Batch Size: 32
Training Data: ImageNet
Training Techniques:
- RMSProp
- Weight Decay
- Gradient Clipping
- Label Smoothing
Training Resources: 8x V100 GPUs
Training Time: 24 hours
Architecture:
- Auxiliary Classifier
- Inception-v3 Module
Results:
- Task: Image Classification
Dataset: ImageNet
Metrics:
Top 1 Accuracy: 74.67%
Top 5 Accuracy: 92.1%
Paper: https://arxiv.org/abs/1512.00567v3
Code: https://github.com/rwightman/pytorch-image-models/blob/timm/models/inception_v3.py#L442
Weights: https://download.pytorch.org/models/inception_v3_google-1a9a5a14.pth
Config: configs/inception-v3-config.json
README: docs/inception-v3-readme.md