Skip to content

Commit f72e75a

Browse files
authored
Update README.md
1 parent c117e30 commit f72e75a

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

LayoutLMv3/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ Note that LayoutLMv3 is identical to LayoutLMv2 in terms of training/inference,
1717
* images need to be resized and normalized, such that they are `pixel_values` of shape `(batch_size, num_channels, heigth, width)`. The channels need to be in RGB format. This was not the case for LayoutLMv2, which expected the channels in BGR format (due to its Detectron2 visual backbone), and normalized the images internally.
1818
* tokenization of text is based on RoBERTa, hence byte-level Byte-Pair-Encoding. This in contrast to LayoutLMv2, which used BERT-like WordPiece tokenization.
1919

20-
Because of this, I've created a new `LayoutLMv3Processor`, which combines a `LayoutLMv3FeatureExtractor` (for the image modality) and a `LayoutLMv3TokenizerFast` (for the text modality) into one. Usage is identical to its predecessor [`LayoutLMv2Processor`](https://huggingface.co/docs/transformers/model_doc/layoutlmv2#usage-layoutlmv2processor).
20+
Because of this, I've created a new `LayoutLMv3Processor`, which combines a `LayoutLMv3ImageProcessor` (for the image modality) and a `LayoutLMv3TokenizerFast` (for the text modality) into one. Usage is identical to its predecessor [`LayoutLMv2Processor`](https://huggingface.co/docs/transformers/model_doc/layoutlmv2#usage-layoutlmv2processor).
2121

2222
The full documentation can be found [here](https://huggingface.co/transformers/model_doc/layoutlmv3.html).
2323

0 commit comments

Comments
 (0)