News

To solve the problems of small samples, acquisition difficulties, under representation and labeling difficulties in object detection, recognition, and segmentation tasks for underwater all-category ...
CLIP (Contrastive Language-Image Pre-Training) is a neural network trained on a variety of (image, text) pairs. It can map images and text into the same latent space, so that they can be compared ...
In this work, we propose a novel nucleus-aware self-supervised pretraining framework for histopathology images. The framework aims to capture the nuclear morphology and distribution information ...