Improve model card: Update pipeline tag, add library name, and usage example
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,11 +1,12 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
pipeline_tag: mask-generation
|
| 4 |
base_model:
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
| 7 |
tags:
|
| 8 |
-
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction
|
|
@@ -31,6 +32,41 @@ tags:
|
|
| 31 |
| **SeC (Ours)** | **82.7** | **81.7** | **86.5** | **75.3** | **91.3** | **88.6** | **70.0** |
|
| 32 |
|
| 33 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
## Citation
|
| 35 |
|
| 36 |
If you find this project useful in your research, please consider citing:
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
+
- OpenGVLab/InternVL2.5-4B
|
| 4 |
+
- facebook/sam2.1-hiera-large
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
pipeline_tag: image-segmentation
|
| 7 |
tags:
|
| 8 |
+
- SeC
|
| 9 |
+
library_name: transformers
|
| 10 |
---
|
| 11 |
|
| 12 |
# SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction
|
|
|
|
| 32 |
| **SeC (Ours)** | **82.7** | **81.7** | **86.5** | **75.3** | **91.3** | **88.6** | **70.0** |
|
| 33 |
|
| 34 |
---
|
| 35 |
+
|
| 36 |
+
## Usage
|
| 37 |
+
|
| 38 |
+
You can load the SeC model and processor using the `transformers` library with `trust_remote_code=True`. For comprehensive video object segmentation and detailed usage instructions, please refer to the project's [GitHub repository](https://github.com/OpenIXCLab/SeC), particularly `demo.ipynb` for single video inference and `INFERENCE.md` for full inference and evaluation.
|
| 39 |
+
|
| 40 |
+
```python
|
| 41 |
+
import torch
|
| 42 |
+
from transformers import AutoModel, AutoProcessor
|
| 43 |
+
from PIL import Image
|
| 44 |
+
|
| 45 |
+
# Load model and processor
|
| 46 |
+
model_name = "OpenIXCLab/SeC-4B"
|
| 47 |
+
# Ensure your environment has the necessary PyTorch and transformers versions as specified in the GitHub repo.
|
| 48 |
+
model = AutoModel.from_pretrained(model_name, trust_remote_code=True, torch_dtype=torch.bfloat16).cuda()
|
| 49 |
+
processor = AutoProcessor.from_pretrained(model_name, trust_remote_code=True)
|
| 50 |
+
|
| 51 |
+
# Example: Assuming you have an image (e.g., a frame from a video) and a text query
|
| 52 |
+
# For full video processing, refer to the project's GitHub repository.
|
| 53 |
+
# Placeholder for an actual image path
|
| 54 |
+
# image = Image.open("path/to/your/image.jpg").convert("RGB")
|
| 55 |
+
# text_query = "segment the main object"
|
| 56 |
+
|
| 57 |
+
# # Prepare inputs
|
| 58 |
+
# inputs = processor(images=image, text=text_query, return_tensors="pt").to(model.device)
|
| 59 |
+
|
| 60 |
+
# # Perform inference
|
| 61 |
+
# with torch.no_grad():
|
| 62 |
+
# outputs = model(**inputs)
|
| 63 |
+
|
| 64 |
+
# The output format will vary depending on the model's implementation.
|
| 65 |
+
# Typically, for segmentation tasks, outputs might include logits or predicted masks.
|
| 66 |
+
# You will need to process these outputs further to visualize the segmentation.
|
| 67 |
+
print("Model loaded successfully. For actual inference with video data, please refer to the project's GitHub repository and demo.ipynb.")
|
| 68 |
+
```
|
| 69 |
+
|
| 70 |
## Citation
|
| 71 |
|
| 72 |
If you find this project useful in your research, please consider citing:
|