Skip to content

Commit 4e990e2

Browse files
committed
Merge branch 'main' of https://github.com/allenai/olmocr
2 parents a13a501 + c7ddad0 commit 4e990e2

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

Diff for: olmocr/bench/prompts.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
def build_basic_prompt() -> str:
2-
return "Please provide a natural, plain text representation of the document, formatted in Markdown. For mathematical expressions, use LaTeX notation with \( and \) for inline equations and \[ and \] for display equations. Convert any tables into Markdown format."
2+
return "Please provide a natural, plain text representation of the document, formatted in Markdown. Skip any headers and footers. For ALL mathematical expressions, use LaTeX notation with \( and \) for inline equations and \[ and \] for display equations. Convert any tables into Markdown format."
33

44

55
def claude_response_format_schema() -> dict:

Diff for: olmocr/bench/scripts/convert_all.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -260,7 +260,7 @@ pip install --upgrade vllm==0.8.3
260260

261261

262262
start_server vllm "Qwen/Qwen2.5-VL-7B-Instruct" --max-model-len 8192
263-
python -m olmocr.bench.convert --dir "$BENCH_DIR" server:name=qwen25vl_prompt6:model=Qwen/Qwen2.5-VL-7B-Instruct:temperature=0.1:prompt_template=basic:response_template=plain --parallel 50
263+
python -m olmocr.bench.convert --dir "$BENCH_DIR" server:name=qwen25vl_prompt7:model=Qwen/Qwen2.5-VL-7B-Instruct:temperature=0.1:prompt_template=basic:response_template=plain --parallel 50
264264
stop_server
265265

266266
start_server vllm "reducto/RolmOCR" --max-model-len 8192

0 commit comments

Comments
 (0)