jakep-allenai commited on
Commit
f7118f1
·
verified ·
1 Parent(s): c2a058d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -9,9 +9,9 @@ library_name: transformers
9
 
10
  <img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
11
 
12
- # olmOCR-7B-1025-FP8
13
 
14
- Quantized to FP8 Version of [olmOCR-7B-1025](https://huggingface.co/allenai/olmOCR-7B-1025), using llmcompressor.
15
 
16
  This is a release of the olmOCR model that's fine tuned from Qwen2.5-VL-7B-Instruct using the
17
  [olmOCR-mix-1025](https://huggingface.co/datasets/allenai/olmOCR-mix-1025) dataset. It has been additionally
@@ -50,7 +50,7 @@ This model scores the following scores on [olmOCR-bench](https://huggingface.co/
50
  </thead>
51
  <tbody>
52
  <tr>
53
- <td align="left">olmOCR pipeline v0.4.0 with olmOCR-7B-1025</td>
54
  <td align="center">82.9</td>
55
  <td align="center">82.1</td>
56
  <td align="center">84.3</td>
@@ -62,7 +62,7 @@ This model scores the following scores on [olmOCR-bench](https://huggingface.co/
62
  <td align="center">82.3 ± 1.1</td>
63
  </tr>
64
  <tr>
65
- <td align="left">olmOCR pipeline v0.4.0 with olmOCR-7B-1025-FP8</td>
66
  <td align="center">83.0</td>
67
  <td align="center">82.3</td>
68
  <td align="center">84.9</td>
@@ -111,7 +111,7 @@ from olmocr.data.renderpdf import render_pdf_to_base64png
111
  from olmocr.prompts import build_no_anchoring_v4_yaml_prompt
112
 
113
  # Initialize the model
114
- model = Qwen2_5_VLForConditionalGeneration.from_pretrained("allenai/olmOCR-7B-1025", torch_dtype=torch.bfloat16).eval()
115
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct")
116
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
117
  model.to(device)
 
9
 
10
  <img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
11
 
12
+ # olmOCR-2-7B-1025-FP8
13
 
14
+ Quantized to FP8 Version of [olmOCR-2-7B-1025](https://huggingface.co/allenai/olmOCR-2-7B-1025), using llmcompressor.
15
 
16
  This is a release of the olmOCR model that's fine tuned from Qwen2.5-VL-7B-Instruct using the
17
  [olmOCR-mix-1025](https://huggingface.co/datasets/allenai/olmOCR-mix-1025) dataset. It has been additionally
 
50
  </thead>
51
  <tbody>
52
  <tr>
53
+ <td align="left">olmOCR pipeline v0.4.0 with olmOCR-2-7B-1025</td>
54
  <td align="center">82.9</td>
55
  <td align="center">82.1</td>
56
  <td align="center">84.3</td>
 
62
  <td align="center">82.3 ± 1.1</td>
63
  </tr>
64
  <tr>
65
+ <td align="left">olmOCR pipeline v0.4.0 with olmOCR-2-7B-1025-FP8</td>
66
  <td align="center">83.0</td>
67
  <td align="center">82.3</td>
68
  <td align="center">84.9</td>
 
111
  from olmocr.prompts import build_no_anchoring_v4_yaml_prompt
112
 
113
  # Initialize the model
114
+ model = Qwen2_5_VLForConditionalGeneration.from_pretrained("allenai/olmOCR-2-7B-1025", torch_dtype=torch.bfloat16).eval()
115
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct")
116
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
117
  model.to(device)