Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
36 commits
Select commit Hold shift + click to select a range
e1cd786
[`BLIP`] fix doctest (#21217)
younesbelkada Jan 23, 2023
c8d719f
Generate: precision fix in compute_transition_scores doctests (#21251)
gante Jan 23, 2023
91ff7ef
[DETR and friends] Use AutoBackbone as alternative to timm (#20833)
NielsRogge Jan 23, 2023
96b2b2d
Extend Script to enable conversion of Encoder Only T5x Models to Pyt…
ToluClassics Jan 23, 2023
66459ce
Add test_image_processing_common.py (#20785)
amyeroberts Jan 23, 2023
6e4d3f0
[GIT] Convert more checkpoints (#21245)
NielsRogge Jan 23, 2023
eaace0c
Optimize by not computing gradients for parameters set to requires_gr…
raghavanone Jan 23, 2023
cb6b568
Fix reformer CI (#21254)
ydshieh Jan 23, 2023
9291116
Add Japanese translation installation.mdx (#21241)
kambehmw Jan 23, 2023
5603f78
Add scikit-learn dependency to train langage-modeling (#21229)
mostafaelhoushi Jan 23, 2023
cf1a1ee
Add missing checkpoint for doctest (#21258)
amyeroberts Jan 23, 2023
1eda4a4
Generate: save generation config with the models' `.save_pretrained()…
gante Jan 23, 2023
354ea44
Replace reduce_labels with do_reduce_labels (#21218)
amyeroberts Jan 23, 2023
e2bd7f8
Update tests: replace feature extractor tests with image processor (#…
amyeroberts Jan 23, 2023
2218dac
Notebook examples grouping and update (#21265)
MKhalusova Jan 23, 2023
275ad9d
Add: TensorFlow example for semantic segmentation task guide (#21223)
MKhalusova Jan 23, 2023
b80b221
[ci-daily] Fix pipeline tests (#21257)
ArthurZucker Jan 23, 2023
c18b4fb
Add class properties with warnings (#21195)
amyeroberts Jan 23, 2023
d8415ba
[Whisper] fix all issues with unk token (#21250)
ArthurZucker Jan 23, 2023
9e86c4e
Supported pipeline tasks update (#21268)
MKhalusova Jan 23, 2023
fd5cdae
Models docstring (#21225)
sgugger Jan 23, 2023
7119bb0
v4.27.0.dev0
sgugger Jan 23, 2023
bde7378
Skip `test_multi_gpu_data_parallel_forward` for `UperNetModelTest` (#…
ydshieh Jan 24, 2023
f0fc791
[Doc] fix broken link (#21276)
younesbelkada Jan 24, 2023
071529b
Use return_tensors="np" instead of "tf" (#21266)
Rocketknight1 Jan 24, 2023
6731644
Hotifx remove tuple for git config image processor. (#21278)
Narsil Jan 24, 2023
767939a
Use `logger.info` instead of `print` to emit a logging message in `hu…
hkiyomaru Jan 24, 2023
f424b09
Fix MaskFormerImageProcessor.post_process_instance_segmentation (#21256)
alaradirik Jan 24, 2023
e2e393c
[`t5`] Fix T5 inference in `float16` + `bnb` error (#21281)
younesbelkada Jan 24, 2023
9286039
[examples/deepspeed] fix renamed api (#21283)
stas00 Jan 24, 2023
94a7edd
[GenerationConfig] add additional kwargs handling (#21269)
ArthurZucker Jan 24, 2023
14d058b
[W2V2 with LM] Fix decoder test with params (#21277)
sanchit-gandhi Jan 24, 2023
1f98121
Fix `TrainingArguments.label_names` docs to reflect the correct defau…
fredtcaroli Jan 24, 2023
de1ca3a
Update expected values for doctest (#21284)
stevhliu Jan 24, 2023
efdbad5
[GIT] Add test for batched generation (#21282)
NielsRogge Jan 25, 2023
99e7905
Supporting `ImageProcessor` in place of `FeatureExtractor` for pipeli…
Narsil Jan 25, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
34 changes: 17 additions & 17 deletions README.md

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion docs/source/en/main_classes/text_generation.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ of the generation method.

To learn how to inspect a model's generation configuration, what are the defaults, how to change the parameters ad hoc,
and how to create and save a customized generation configuration, refer to the
[text generation strategies guide](./generation_strategies).
[text generation strategies guide](../generation_strategies).

## GenerationConfig

Expand Down
13 changes: 11 additions & 2 deletions docs/source/en/model_doc/mask2former.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -22,8 +22,8 @@ The abstract from the paper is the following:
of semantics defines a task. While only the semantics of each task differ, current research focuses on designing specialized architectures for each task. We present Masked-attention Mask Transformer (Mask2Former), a new architecture capable of addressing any image segmentation task (panoptic, instance or semantic). Its key components include masked attention, which extracts localized features by constraining cross-attention within predicted mask regions. In addition to reducing the research effort by at least three times, it outperforms the best specialized architectures by a significant margin on four popular datasets. Most notably, Mask2Former sets a new state-of-the-art for panoptic segmentation (57.8 PQ on COCO), instance segmentation (50.1 AP on COCO) and semantic segmentation (57.7 mIoU on ADE20K).*

Tips:
- Mask2Former uses the same preprocessing and postprocessing steps as [MaskFormer](maskformer). Use [`MaskFormerImageProcessor`] or [`AutoImageProcessor`] to prepare images and optional targets for the model.
- To get the final segmentation, depending on the task, you can call [`~MaskFormerImageProcessor.post_process_semantic_segmentation`] or [`~MaskFormerImageProcessor.post_process_instance_segmentation`] or [`~MaskFormerImageProcessor.post_process_panoptic_segmentation`]. All three tasks can be solved using [`Mask2FormerForUniversalSegmentation`] output, panoptic segmentation accepts an optional `label_ids_to_fuse` argument to fuse instances of the target object/s (e.g. sky) together.
- Mask2Former uses the same preprocessing and postprocessing steps as [MaskFormer](maskformer). Use [`Mask2FormerImageProcessor`] or [`AutoImageProcessor`] to prepare images and optional targets for the model.
- To get the final segmentation, depending on the task, you can call [`~Mask2FormerImageProcessor.post_process_semantic_segmentation`] or [`~Mask2FormerImageProcessor.post_process_instance_segmentation`] or [`~Mask2FormerImageProcessor.post_process_panoptic_segmentation`]. All three tasks can be solved using [`Mask2FormerForUniversalSegmentation`] output, panoptic segmentation accepts an optional `label_ids_to_fuse` argument to fuse instances of the target object/s (e.g. sky) together.

This model was contributed by [Shivalika Singh](https://huggingface.co/shivi) and [Alara Dirik](https://huggingface.co/adirik). The original code can be found [here](https://github.com/facebookresearch/Mask2Former).

Expand Down Expand Up @@ -55,3 +55,12 @@ The resource should ideally demonstrate something new instead of duplicating an

[[autodoc]] Mask2FormerForUniversalSegmentation
- forward

## Mask2FormerImageProcessor

[[autodoc]] Mask2FormerImageProcessor
- preprocess
- encode_inputs
- post_process_semantic_segmentation
- post_process_instance_segmentation
- post_process_panoptic_segmentation
18 changes: 9 additions & 9 deletions docs/source/en/task_summary.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -85,11 +85,11 @@ Image classification labels an entire image from a predefined set of classes. Li
... )
>>> preds = [{"score": round(pred["score"], 4), "label": pred["label"]} for pred in preds]
>>> print(*preds, sep="\n")
{'score': 0.4403, 'label': 'lynx, catamount'}
{'score': 0.0343, 'label': 'cougar, puma, catamount, mountain lion, painter, panther, Felis concolor'}
{'score': 0.0321, 'label': 'snow leopard, ounce, Panthera uncia'}
{'score': 0.0235, 'label': 'Egyptian cat'}
{'score': 0.023, 'label': 'tiger cat'}
{'score': 0.4335, 'label': 'lynx, catamount'}
{'score': 0.0348, 'label': 'cougar, puma, catamount, mountain lion, painter, panther, Felis concolor'}
{'score': 0.0324, 'label': 'snow leopard, ounce, Panthera uncia'}
{'score': 0.0239, 'label': 'Egyptian cat'}
{'score': 0.0229, 'label': 'tiger cat'}
```

### Object detection
Expand Down Expand Up @@ -131,10 +131,10 @@ Segmentation tasks are helpful in self-driving vehicles to create a pixel-level
... "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/pipeline-cat-chonk.jpeg"
... )
>>> preds = [{"score": round(pred["score"], 4), "label": pred["label"]} for pred in preds]
>>> preds
[{'score': 0.9856, 'label': 'LABEL_184'},
{'score': 0.9976, 'label': 'snow'},
{'score': 0.9962, 'label': 'cat'}]
>>> print(*preds, sep="\n")
{'score': 0.9879, 'label': 'LABEL_184'}
{'score': 0.9973, 'label': 'snow'}
{'score': 0.9972, 'label': 'cat'}
```

### Depth estimation
Expand Down
Loading