But it also faces a serious barrier: The shortage of labeled training data.

To overcome this hurdle, scientists have explored different solutions to various degrees of success.

Google is one of several organizations that has been exploring its use inmedical imaging.

Google Research: Self-supervised learning is transforming medical imaging

It’s free, every week, in your inbox.

One of the key challenges of deep learning is the need for huge amounts of annotated data.

Large neural networks require millions of labeled examples to reach optimal accuracy.

Article image

In medical parameters, data labeling is a complicated and costly endeavor.

One popular way to address the shortage of medical data is to use supervised pre-training.

In this approach, a CNN is initially trained on a dataset of labeled images such as ImageNet.

self-supervised-learning-contrastive-learning

This phase tunes the parameters of the models layers to the general patterns found in all kinds of images.

The trained deep learning model can then befine-tunedon a limited set of labeled examples for the target task.

However, supervised pre-training also has its limits.

multi-instance-contrastive-learning

The common paradigm for training medical imaging models is transfer learning where models are first pre-trained usingsupervised learningon ImageNet.

Self-supervised pre-training

Self-supervised learning has emerged as a promising area of research in recent years.

Outside of medical configs, Googlehas developed several self-supervised learning techniques to train neural networks for computer vision tasks.

self-supervised-pretraining-for-medical-imaging

Unlabeled data is often available in large quantities in various medical domains.

Putting it all together

The self-supervised learning framework the Google researchers used involved three steps.

First, the target neural internet was trained on examples from the ImageNet dataset using SimCLR.

Finally, the model is fine-tuned on a limited dataset of labeled images for the target software.

The researchers tested the framework on two dermatology and chest x-ray interpretation tasks.

And it requires much less labeled data.

This hopefully translates to significant cost and time savings for developing medical AI models.

you’ve got the option to read the original articlehere.

Also tagged with