PDF Pals v1.5

This release integrates with GPT 4 Vision and supports more AI service providers, including Local LLM Inference Server.

TL;DR:

  • New: Supports more AI service providers: Mistral AI, Perplexity and Together AI

  • New: Supports custom OpenAI-compatible servers: OpenAI proxy, LocalAI and LM Studio Inference Server...

  • New: Incorporates ShotSolve into PDF Pals: take a screenshot of your PDF and ask AI about it

  • New: Switch to the new embedding model text-embedding-3: better & cheaper

  • New: Added support for the new GPT 4 Turbo model (gpt-4-0125-preview)

  • New: Allow searching within the current chat
  • Fix: Fixed the bug where the Upgrade window keeps showing

More AI service providers

Use Mistral AI and Perplexity on macOS

In this version, I've added support for 3 new AI service providers: Mistral AI, Perplexity & Together AI.

Note that PDF Pals will always use OpenAI's embedding model or a local AI model to process your documents.

To set up, go to Settings > Models, click (+) button then fill the form. Pretty straightforward isn't it?

Set up Mistral AI and Perplexity on macOS

Use PDF Pals with Local Inference Servers

In this version, I also added the ability to use a custom OpenAI-compatible inference server. It's still in beta though, if you found an issue, please reach out. I would priotize this feature.

To start, follow this setup guide.

Set up local LLMs on macOS

ShotSolve: easily chat with screenshots

Earlier this year, I started a new project called ShotSolve. It's a free app that allows you to take a screenshot and quickly ask GPT-4 Vision about it.

In this version, I've incorporated its features back into PDF Pals. You can now take a screenshot of the PDF quickly via the "Capture" toolbar button, then let GPT-4 Vision handle your request.

Note that it requires an OpenAI API key as this feature utilizes the GPT 4 Vision model.

Here is the demo video

Other improvements

Search within conversation

I've added the ability to search within the current chat. You can manually switch the search scope (screenshot below), or to use the keyboard shortcut Command + Shift + F

Cheaper and better document analysis

In this version, I've switched to the new embedding model from OpenAI (text-embedding-3). It's 5x cheaper than the previous model (text-embedding-ada), which allows you to process a 63,000 pages document for just $1.

Enjoy cheaper and better document analysis!

New GPT 4 Turbo model

I've added support for the improved GPT 4 Turbo model gpt-4-0125-preview. OpenAI fixed the issue with lazyness in this version. If you use PDF Pals to gain insights from programming ebooks, you should use to this model.

Other minor bug fixes & improvements

And that's it

See you in the next update 👋

PS: If you enjoy PDF Pals, please share a testimonial. Much appreciated 🙏

If you are new here, PDF Pals is a native macOS app that allows you to chat with local PDFs instantly. Download now.