Authorizations
Cookies
Body
Optional file URL (http/https). If provided, the server will download and process it.
The maximum number of pages in the PDF to convert.
The page range to parse, comma separated like 0,5-10,20. This will override max_pages if provided. Example: '0,2-4' will process pages 0, 2, 3, and 4.
Note: This parameter has been deprecated, and will be ignored in the current version. The languages to use if OCR is needed, comma separated. Must be either the names or codes from https://github.com/datalab-to/surya/blob/master/surya/languages.py. Any other inputs will be ignored.
Force OCR on all pages of the PDF. Defaults to False. This can lead to worse results if you have good text in your PDFs (which is true in most cases).
Format the lines in the output. Defaults to False. If set to True, the lines will be formatted to detect inline math and styles.
Whether to paginate the output. Defaults to False. If set to True, each page of the output will be separated by a horizontal rule that contains the page number (2 newlines, {PAGE_NUMBER}, 48 - characters, 2 newlines).
Strip existing OCR text from the PDF and re-run OCR. If force_ocr is set, this will be ignored. Defaults to False.
Disable image extraction from the PDF. If use_llm is also set, then images will be automatically captioned. Defaults to False.
Disable inline math recognition in OCR.
Significantly improves accuracy by using an LLM to enhance tables, forms, inline math, and layout detection. Will increase latency. Defaults to False.
Which output mode to use - fast has the lowest latency and preserves the most positional information. Accurate is the slowest, and preserves the least.
The output format for the text. Can be 'json', 'html', 'markdown', or 'chunks'. Defaults to 'markdown'. You can comma separate multiple formats, like markdown,html.
Skip the cache and re-run the inference. Defaults to False. If set to True, the cache will be skipped and the inference will be re-run.
Save the checkpoint after processing. Defaults to False. This is only useful if you're applying custom rules iteratively.
An optional prompt that marker will use to improve the output, and align it to specific requirements.
The schema to use for structured extraction (only used with structured extraction endpoint). The ideal way to generate this is to create a Pydantic schema, then convert to JSON with .model_dump_json().
The schema to use for document segmentation. Should be a JSON string containing segment names and descriptions for identifying page ranges of different document sections.
Additional configuration options for marker. This should be a JSON string with key-value pairs. For example, '{"key": "value"}'. This supports these keys: ['disable_links', 'keep_pageheader_in_output', 'keep_pagefooter_in_output', 'filter_blank_pages', 'drop_repeated_text', 'layout_coverage_threshold', 'merge_threshold', 'height_tolerance', 'gap_threshold', 'image_threshold', 'min_line_length', 'level_count', 'default_level', 'no_merge_tables_across_pages', 'force_layout_block']
Optional workflow step data ID. If provided, this request will be associated with the specified workflow step execution.
Input PDF, word document, powerpoint, or image file, uploaded as multipart form data. Images must be png, jpg, or webp format.
Response
Successful Response
The ID of the request. This ID can be used to check the status of the request.
The URL to check the status of the request and get results.
Whether the request was successful.
If the request was not successful, this will contain an error message.
A dictionary of the versions of the libraries used in the request.