Datalab implements various limits to ensure fair usage and maintain service quality for all users. This guide covers all the limits you need to know about, how they work, and what to do if you need higher limits.
For files exceeding 200 MB, you have several options:
File Splitting: Break large PDFs into smaller chunks
Compression: Reduce file size before upload
Enterprise Limits: Contact support for increased limits
Batch Processing: Process files in segments
Copy
Ask AI
# Split a large PDF into 50-page chunksimport PyPDF2def split_pdf(input_file, pages_per_chunk=50): pdf_reader = PyPDF2.PdfReader(input_file) total_pages = len(pdf_reader.pages) for start in range(0, total_pages, pages_per_chunk): pdf_writer = PyPDF2.PdfWriter() end = min(start + pages_per_chunk, total_pages) for page in range(start, end): pdf_writer.add_page(pdf_reader.pages[page]) output_file = f"chunk_{start//pages_per_chunk + 1}.pdf" with open(output_file, 'wb') as output: pdf_writer.write(output)
When you exceed rate limits, you’ll receive a 429 error and will need to wait up to 60 seconds. For most accounts, this is capped to 200 documents per minute. Custom rate limits are available for enterprise plan users. More information can be found here.