Update README.md
Browse files
README.md
CHANGED
|
@@ -9,9 +9,9 @@ library_name: transformers
|
|
| 9 |
|
| 10 |
<img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
|
| 11 |
|
| 12 |
-
# olmOCR-7B-1025-FP8
|
| 13 |
|
| 14 |
-
Quantized to FP8 Version of [olmOCR-7B-1025](https://huggingface.co/allenai/olmOCR-7B-1025), using llmcompressor.
|
| 15 |
|
| 16 |
This is a release of the olmOCR model that's fine tuned from Qwen2.5-VL-7B-Instruct using the
|
| 17 |
[olmOCR-mix-1025](https://huggingface.co/datasets/allenai/olmOCR-mix-1025) dataset. It has been additionally
|
|
@@ -50,7 +50,7 @@ This model scores the following scores on [olmOCR-bench](https://huggingface.co/
|
|
| 50 |
</thead>
|
| 51 |
<tbody>
|
| 52 |
<tr>
|
| 53 |
-
<td align="left">olmOCR pipeline v0.4.0 with olmOCR-7B-1025</td>
|
| 54 |
<td align="center">82.9</td>
|
| 55 |
<td align="center">82.1</td>
|
| 56 |
<td align="center">84.3</td>
|
|
@@ -62,7 +62,7 @@ This model scores the following scores on [olmOCR-bench](https://huggingface.co/
|
|
| 62 |
<td align="center">82.3 ± 1.1</td>
|
| 63 |
</tr>
|
| 64 |
<tr>
|
| 65 |
-
<td align="left">olmOCR pipeline v0.4.0 with olmOCR-7B-1025-FP8</td>
|
| 66 |
<td align="center">83.0</td>
|
| 67 |
<td align="center">82.3</td>
|
| 68 |
<td align="center">84.9</td>
|
|
@@ -111,7 +111,7 @@ from olmocr.data.renderpdf import render_pdf_to_base64png
|
|
| 111 |
from olmocr.prompts import build_no_anchoring_v4_yaml_prompt
|
| 112 |
|
| 113 |
# Initialize the model
|
| 114 |
-
model = Qwen2_5_VLForConditionalGeneration.from_pretrained("allenai/olmOCR-7B-1025", torch_dtype=torch.bfloat16).eval()
|
| 115 |
processor = AutoProcessor.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct")
|
| 116 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 117 |
model.to(device)
|
|
|
|
| 9 |
|
| 10 |
<img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
|
| 11 |
|
| 12 |
+
# olmOCR-2-7B-1025-FP8
|
| 13 |
|
| 14 |
+
Quantized to FP8 Version of [olmOCR-2-7B-1025](https://huggingface.co/allenai/olmOCR-2-7B-1025), using llmcompressor.
|
| 15 |
|
| 16 |
This is a release of the olmOCR model that's fine tuned from Qwen2.5-VL-7B-Instruct using the
|
| 17 |
[olmOCR-mix-1025](https://huggingface.co/datasets/allenai/olmOCR-mix-1025) dataset. It has been additionally
|
|
|
|
| 50 |
</thead>
|
| 51 |
<tbody>
|
| 52 |
<tr>
|
| 53 |
+
<td align="left">olmOCR pipeline v0.4.0 with olmOCR-2-7B-1025</td>
|
| 54 |
<td align="center">82.9</td>
|
| 55 |
<td align="center">82.1</td>
|
| 56 |
<td align="center">84.3</td>
|
|
|
|
| 62 |
<td align="center">82.3 ± 1.1</td>
|
| 63 |
</tr>
|
| 64 |
<tr>
|
| 65 |
+
<td align="left">olmOCR pipeline v0.4.0 with olmOCR-2-7B-1025-FP8</td>
|
| 66 |
<td align="center">83.0</td>
|
| 67 |
<td align="center">82.3</td>
|
| 68 |
<td align="center">84.9</td>
|
|
|
|
| 111 |
from olmocr.prompts import build_no_anchoring_v4_yaml_prompt
|
| 112 |
|
| 113 |
# Initialize the model
|
| 114 |
+
model = Qwen2_5_VLForConditionalGeneration.from_pretrained("allenai/olmOCR-2-7B-1025", torch_dtype=torch.bfloat16).eval()
|
| 115 |
processor = AutoProcessor.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct")
|
| 116 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 117 |
model.to(device)
|