Skip to content

Releases: VikParuchuri/marker

Support word, powerpoint, excel, html, epub + math improvements

28 Feb 23:55
b985880
Compare
Choose a tag to compare

Support xlsx, docx, pptx, html, epub

Marker now has support for additional document formats. You have to run pip install marker-pdf[full] to install all the dependencies.

Improved text detection

OCR should now work better due to an improved text detection model.

Inline math improvements

  • Better inline math detection with an improved model.
  • Inline math lines are now inference.
  • --redo-inline-math option to enable the highest quality math detection

Misc improvements

  • Support for the claude model
  • Improve benchmarking scripts
  • Merge lines better with new text detection model

What's Changed

New Contributors

Full Changelog: v1.5.5...v1.6.0

Fix LLM layout relabeling bug

19 Feb 23:04
141da8c
Compare
Choose a tag to compare

When text blocks were relabeled by the LLM in --use_llm mode, sometimes lines wouldn't merge into the blocks properly, resulting in missing text.

What's Changed

Full Changelog: v1.5.4...v1.5.5

Download models from r2; fix inline math

19 Feb 17:53
434c0ce
Compare
Choose a tag to compare

Download models from R2

  • Models will now download from R2, which is fast and reliable

Inline math improvements

  • Fix issues with inline math and escape characters

What's Changed

Full Changelog: v1.5.3...v1.5.4

v1.5.3

19 Feb 03:43
27d2b9e
Compare
Choose a tag to compare

Windows fixes

  • Fix issue with streamlit app and permissions
  • Fix torch classes issue

Memory leak fix

Fixed memory leak with repeatedly reusing the same converter.

Convert.py enhancements

  • disable tqdm bars when converting multiple files

What's Changed

New Contributors

Full Changelog: v1.5.2...v1.5.3

Fix LLM service issue

14 Feb 01:49
0af86b1
Compare
Choose a tag to compare

Fix issue with initializing the LLM service with no default specified.

Fix OCR issue

14 Feb 01:06
9721c17
Compare
Choose a tag to compare

Fix issue with OCRing documents with a mix of good and bad pages.

Inline math; speed up LLM calls; allow local models

13 Feb 21:09
b87d088
Compare
Choose a tag to compare

Inline math

Marker will handle inline math if --use_llm is set. This makes reading scientific papers a lot nicer! The feature has been optimized for speed.

image

Local LLMs

We now support Ollama - when you're passing the --use_llm flag, you can select the Ollama inference service like this:

marker_single FILEPATH --use_llm --llm_service marker.services.ollama.OllamaService

You can set the options --ollama_base_url and --ollama_model. By default, it will use llama3.2-vision.

Batch LLM calls

LLM calls are now batched across processors for a significant speedup if you're passing --use_llm.

Misc fixes

  • Biology PDFs now work a lot better - leading line numbers are stripped
  • Improved OCR heuristics
  • Updated the examples

What's Changed

Full Changelog: v1.4.0...v1.5.0

LLM fixes; new benchmarks

11 Feb 16:41
a357967
Compare
Choose a tag to compare

New benchmarks

Overall

Benchmark against llamaparse, docling, mathpix (see README for how to run benchmarks). Marker performs favorably against alternatives in speed, llm as judge scoring, and heuristic scoring.

image

Table

Benchmark tables against gemini flash:

image

Update gemini model

  • Use the new genai library
  • Update to gemini flash 2.0

Misc bugfixes

  • Fix bug with OCR heuristics not being aggressive enough
  • Fix bug with empty tables
  • Ensure references get passed through in llm processors

What's Changed

Full Changelog: v1.3.5...v1.4.0

Bump gemini version

06 Feb 02:08
0aa40a5
Compare
Choose a tag to compare

When using the optional llm mode, there appears to be a bug with gemini flash 1.5. This release bumps the version to gemini flash 2.0, which appears to resolve the bug.

Fix pytorch bug

31 Jan 03:00
dba5b4c
Compare
Choose a tag to compare

There was a bug with pytorch 2.6 and MPS that caused errors in inference - this has been fixed.