34 lines
1.5 KiB
Markdown
34 lines
1.5 KiB
Markdown
|
---
|
||
|
language: en
|
||
|
license: mit
|
||
|
tags:
|
||
|
- vision
|
||
|
- image-segmentation
|
||
|
model_name: openmmlab/upernet-convnext-small
|
||
|
---
|
||
|
|
||
|
# UperNet, ConvNeXt small-sized backbone
|
||
|
|
||
|
UperNet framework for semantic segmentation, leveraging a ConvNeXt backbone. UperNet was introduced in the paper [Unified Perceptual Parsing for Scene Understanding](https://arxiv.org/abs/1807.10221) by Xiao et al.
|
||
|
|
||
|
Combining UperNet with a ConvNeXt backbone was introduced in the paper [A ConvNet for the 2020s](https://arxiv.org/abs/2201.03545).
|
||
|
|
||
|
Disclaimer: The team releasing UperNet + ConvNeXt did not write a model card for this model so this model card has been written by the Hugging Face team.
|
||
|
|
||
|
## Model description
|
||
|
|
||
|
UperNet is a framework for semantic segmentation. It consists of several components, including a backbone, a Feature Pyramid Network (FPN) and a Pyramid Pooling Module (PPM).
|
||
|
|
||
|
Any visual backbone can be plugged into the UperNet framework. The framework predicts a semantic label per pixel.
|
||
|
|
||
|

|
||
|
|
||
|
## Intended uses & limitations
|
||
|
|
||
|
You can use the raw model for semantic segmentation. See the [model hub](https://huggingface.co/models?search=openmmlab/upernet) to look for
|
||
|
fine-tuned versions (with various backbones) on a task that interests you.
|
||
|
|
||
|
### How to use
|
||
|
|
||
|
For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/upernet#transformers.UperNetForSemanticSegmentation).
|