Claude is a family of foundational AI models that can be used in a variety of applications.
You can talk directly with Claude to brainstorm ideas, analyze images, create and process long documents.
Claude can help with use cases including summarization, search, creative and collaborative writing, Q&A, coding, and more.
Early customers report that Claude is much less likely to produce harmful outputs, easier to converse with, and more steerable - so you can get your desired output with less effort. Claude can also take direction on personality, tone, and behavior.
This plugin uses an external service to provide streaming capabilities.
Informing you that Prompt Caching has been added to this plugin.
Prompt caching allows reducing costs by up to 90% and latency by up to 85% for long prompts:
Use case
Latency w/o caching (time to first token)
Latency w/ caching (time to first token)
Cost reduction
Chat with a book (100,000 token cached prompt) [1]
11.5s
2.4s (-79%)
-90%
Many-shot prompting (10,000 token prompt) [1]
1.6s
1.1s (-31%)
-86%
Multi-turn conversation (10-turn convo with a long system prompt) [2]