Need to Extract Text from Large PDF Files (1000+ Pages) in Bubble – Any Solution?

Hi everyone,

I’m working on a Bubble project where I need to extract text content from very large PDF files – some of them have 200, 500, even 1000+ pages.

I’ve tried a few APIs (like PDF.co, ConvertAPI, etc.) but ran into issues like timeouts, malformed JSON, or limits on file size/pages.

Has anyone successfully implemented a reliable solution for converting large PDF files to plain text in Bubble?
I’m open to plugins, API integrations, or workflow tricks – just need the full text content extracted.

Appreciate any suggestions, examples, or tools that worked for you!

Thanks in advance :folded_hands:

I don’t have issue with pdf.co and we are processing large pdf. Is the time out was from pdf.co or from api connector?

As an alternative to using something off the shelf, you could develop your own serverless function (eg Cloudflare worker, AWS lambda - all free) with the help of chatGPT, Gemini etc. It could be pretty straightforward for something bespoke like this.