Google AI Team Open Sources BiT – Big Transfer: General Visual Representation Learning (Computer Vision)

Google AI team recently open-source BiT (Big Transfer) for general visual representation learning. Current computer vision training generally involves a pre-trained model due to lack of labeled data for computer vision tasks. This has been a common problem for computer vision scientists to collect and train models with a large set of generic data available via tools like OpenImages or Places, but many times this collection of a large set of data (over 1M labeled images) could be prohibitive for an average practitioner.

The available solutions include using pre-trained models that are trained on generic data (example: ImageNet). Despite these pre-trained models working well in practice, this still doesn’t solve the problem for conditions like grasping news concepts and then to understand them in a different context. Just like how BERT and T5 have shown advances in the language domain, it is believed BiT based large-scale pre-training can advance the performance of computer vision models.



BiT Models – Components of all TL blocks:
 | Website

Asif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.

[Announcing Gretel Navigator] Create, edit, and augment tabular data with the first compound AI system trusted by EY, Databricks, Google, and Microsoft