Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.wcr.is/llms.txt

Use this file to discover all available pages before exploring further.

WeCareRemote gives you the flexibility to choose the AI model that best fits your use case. Rather than locking you into a single provider, the platform supports a broad range of leading models — from frontier models like GPT-4o and Claude 3.5 Sonnet to locally hosted options for privacy-sensitive environments. You can set a platform-wide default or specify a model per request.
If you are unsure where to start, GPT-4o (OpenAI) and Claude 3.5 Sonnet (Anthropic) deliver strong results across a wide range of tasks including multi-language conversation, document retrieval, and nuanced role-aware responses.

Supported providers

OpenAI models are widely used and well-suited for general-purpose conversation, summarization, and document Q&A.
ModelNotes
gpt-4oLatest flagship model; fast and highly capable
gpt-4o-miniLightweight, cost-efficient variant of GPT-4o
gpt-4-turboHigh-performance model with large context window
gpt-3.5-turboFast and affordable for simpler tasks
Anthropic’s Claude models are known for their nuanced, thoughtful responses and strong instruction-following.
ModelNotes
claude-3-5-sonnet-20241022Anthropic’s most capable model
claude-3-haiku-20240307Fastest and most compact Claude model
claude-3-opus-20240229Most powerful Claude 3 model
Google AI models include the Gemini family, available through Google AI Studio or Vertex AI.
ModelNotes
gemini-1.5-proLong context, strong reasoning and multimodal support
gemini-1.5-flashFast, efficient variant of Gemini 1.5
gemini-proPrevious-generation general-purpose model
Groq provides ultra-fast inference for open-weight models, making it ideal for low-latency interactions.
ModelNotes
llama-3.1-70b-versatileLarge Llama 3.1 model via Groq
llama-3.1-8b-instantCompact Llama 3.1 for fast responses
mixtral-8x7b-32768Mixtral MoE model with large context window
gemma2-9b-itGoogle Gemma 2 via Groq
Ollama lets you run open-weight models locally on your own hardware. No data leaves your environment.
Example modelsNotes
llama3Meta’s Llama 3
mistralMistral 7B
phi3Microsoft Phi-3
Any Ollama-supported modelSee ollama.com/library
Ollama is the recommended option for privacy-sensitive use cases. Because models run locally, conversation data never leaves your organization’s infrastructure.
AWS Bedrock provides managed access to a curated set of foundation models through Amazon’s infrastructure.Models available via Bedrock include Anthropic Claude, Meta Llama, Amazon Titan, Mistral, and others depending on your region and access configuration.
Access OpenAI GPT models through your organization’s Azure OpenAI resource, with enterprise-grade compliance and data residency controls.The models available depend on the deployments configured in your Azure portal.
OpenRouter is an aggregator that provides access to models from many providers through a single API, including models from OpenAI, Anthropic, Google, Meta, Mistral, and more.
DeepSeek offers capable models particularly well-suited for reasoning and coding tasks.
Connect any API that is compatible with the OpenAI API format — including self-hosted models, fine-tuned endpoints, or third-party providers not listed above.

How to select a model

You can choose which model the assistant uses in two ways: Platform default — Your administrator configures the default model for the platform. All conversations use this model unless you override it per request. See Configure the AI assistant for details. Per request — When interacting with the assistant through the API, pass a model parameter in your request body to use a specific model for that interaction only. This lets you run different models for different tasks without changing the platform default.