TF-Keras
mobile-vit / README.md
Tonic's picture
Set `library_name` to `tf-keras`. (#1)
f3d1125 verified
metadata
library_name: tf-keras
license: other

Description

  • Year: 2022
  • Organisation: TensorFlow
  • Project Title: Publish fine-tuned MobileViT in TensorFlow Hub TensorFlow Hub is the main TensorFlow model repository with thousands of pre-trained models with documentation, sample code and readily available to use or fine-tune. The idea behind the project is to develop new State-of-the-Art models like MobileViT and publish the pre-trained models on TensorFlow Hub using the ImageNet1k dataset. MobileViT is light-weight and general-purpose vision transformer for mobile devices. MobileViT presents a different perspective for the global processing of information with transformers, i.e., transformers as convolutions. Our results show that MobileViT significantly outperforms CNN- and ViT-based networks across different tasks and datasets. On the ImageNet-1k dataset, MobileViT achieves top-1 accuracy of 78.4% with about 6 million parameters, which is 3.2% and 6.2% more accurate than MobileNetv3 (CNN-based) and DeIT (ViT-based) for a similar number of parameters. On the MS-COCO object detection task, MobileViT is 5.7% more accurate than MobileNetv3 for a similar number of parameters.
  • Mentors: Luis Gustavo Martins & Sayak Paul

Project Report

This repository provides TensorFlow / Keras implementations of different MobileViT [1] variants. It also provides the TensorFlow / Keras models that have been populated with the original MobileViT pre-trained weights available from [2]. These models are not blackbox SavedModels i.e., they can be fully expanded into tf.keras.Model objects and one can call all the utility functions on them (example: .summary()).

As of today, all the TensorFlow / Keras variants of the models listed here are available in this repository. This list includes the ImageNet-1k models.

Refer to the "Using the models" section to get started.

Conversion

TensorFlow / Keras implementations are available in mobilevit/models/mobilevit.py. Conversion utilities are in convert.py.

Models

The converted models will be available on TF-Hub.

There should be a total of 3 different models each having two variants: classifier and feature extractor. You can load any model and get started like so:

import tensorflow as tf

model = tf.keras.models.load_model('model_path')
print(model.summary())

The model names are interpreted as follows:

  • mobilevit_xxs_1k_256: Means that the model was pre-trained on the ImageNet-1k dataset with a resolution of 256x256.

Results

Results are on ImageNet-1k validation set (top-1 accuracy).

name original acc@1 keras acc@1
MobileViT_XXS 69.0 68.59
MobileViT_XS 74.7 74.67
MobileViT_S 78.3 78.36

Differences in the results are primarily because of the differences in the library implementations especially how image resizing is implemented in PyTorch and TensorFlow. Results can be verified with the code in imagenet_1k_eval. Logs are available at this URL.

Using the models

Pre-trained models:

Randomly initialized models:

from mobilevit.models.mobilevit import get_mobilevit_model

model = get_mobilevit_model(
      model_name='mobilevit_xxs', # [mobilevit_xxs, mobilevit_xs, mobilevit_s]
      image_shape=(256, 256, 3),
      num_classes=1000,
    )

print(model.summary())

To view different model configurations, refer here.

Upcoming Contributions

  • Allow the models to accept more input shapes (useful for downstream tasks)
  • Convert the saved_models to TFLite.
  • Fine-tuning notebook
  • Off-the-shelf-classification notebook
  • Publish models on TF-Hub

References

[1] MobileViT Paper: https://arxiv.org/abs/2110.02178

[2] Official MobileViT weights: https://github.com/apple/ml-cvnets

[3] Hugging Face MobileViT: MobileViT-HF

Acknowledgements

πŸ”— Links

portfolio linkedin twitter