AI providers
The AI Assistant extension supports multiple AI providers:
- ChatGPT (OpenAI)
- Anthropic Claude
- Google Gemini
You can either set a single provider/model to be used across extensions or explicitly choose different providers/models for specific tasks. To do this, the extension offers multi-level provider configuration with a clear priority order:
- Global AI provider settings: located in Stores -> Configuration -> Mirasvit extensions -> Developer -> AI configuration. These defaults apply to all Mirasvit extensions that use AI. Lowest priority.
- Module AI provider settings: located in Stores -> Configuration -> Mirasvit extensions -> AI Assistant. Use these when you want a different provider or model specifically for the AI Assistant module. Higher priority than Global.
- Per-item AI provider settings: located inside a prompt, automation rule, or prompt popup. These settings override all others and are applied when that item runs. Highest priority.
Connecting AI providers
To connect an AI provider, fill out the following fields:
- ChatGPT (OpenAI)
- Anthropic Claude
- Google Gemini
Use these settings to configure OpenAI provider.
-
Enable OpenAI provider:
- Yes: use global AI settings from Core module.
- No: use Assistant module provider settings below.
-
Open API key: paste the API key into the field OpenAI Secret Key in order to use the artificial intellect in your store.
noteThe AI Assistant extension accesses ChatGPT via an API. You need to sign up for a ChatGPT account in order to obtain the OpenAI secret key.
Generate the OpenAI secret API key on page platform.openai.com/account/api-keys. Click Create new secret key and copy the key.
New secret key Using ChatGPT API costs money, which are billed from your ChatGPT account. The total pricing depends on the number of tokens generated by this language model. Refer to openai.com/api/pricing/ for more info on prices.
-
Default OpenAI model: choose a most suitable language models for you to utilize. AI Assistant works via various language models.
- GPT-3 (text-davinci-003): legacy model, is suited for any language task. It is configured for longer output of better quality and consistent instruction-following. Its input can be up to 4 097 tokens. It's trained on data before Jun 2021.
- GPT-3.5 (gpt-3.5-turbo): is optimized for chat. It is considered by OpenAI as a most capable GPT-3.5 model, and its price is 1/10th the cost of text-davinci-003. Maximal support input is 16 385 tokens. It was trained on data existing before Sep 2021.
- GPT-4 (gpt-4): is more capable for complex tasks and gives better results on large texts. Maximal support input is 8 192 tokens. It was trained on data existing before Sep 2021.
- GPT-4 Turbo (gpt-4-turbo-preview): is improved GPT-4 model featuring improved instruction following, JSON mode and more. Maximal support input is 128 000 tokens. It was trained on data existing before Apr 2023.
- GPT-4o (gpt-4o): is improved GPT-4 Turbo model. It is 2 times faster and up to 50% cheaper than GPT-4 Turbo model. Maximal support input is 128 000 tokens. It was trained on data existing before Oct 2023.
- GPT-4o mini (gpt-4o-mini): is a fast, affordable small model for focused tasks. It accepts both text and image inputs, and produces text outputs (including Structured Outputs). Maximal support input is 128 000 tokens. It was trained on data existing before Oct 2023.
- GPT-4.1 (gpt-4.1): high-intelligent model suited for complex tasks and processing complex-structured content. Maximal support input is 1 047 576 tokens. It was trained on data existing before Jun 2024.
- GPT-4.1 mini (gpt-4.1-mini): provides a balance between intelligence, speed, and cost that makes it an attractive model for many use cases. Maximal support input is 1 047 576 tokens. It was trained on data existing before Jun 2024.
- GPT-4.1 nano (gpt-4.1-nano): is the fastest, most cost-effective GPT-4.1 model. Maximal support input is 1 047 576 tokens. It was trained on data existing before Jun 2024.
- GPT-5 (gpt-5): is the latest, high-intelligent model, most powerful model. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
- GPT-5 mini (gpt-5-mini): is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
- GPT-5 nano (gpt-5-nano): is the fastest, cheapest version of GPT-5. It's great for summarization and classification tasks. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
- GPT-5.1 (gpt-5.1): is an enhanced iteration of GPT-5 with improved reasoning stability, better long-context understanding, and more consistent instruction following. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
- GPT-5.2 (gpt-5.2): is a refined, next-generation version of GPT-5.1 that delivers stronger logical reasoning and improved factual accuracy. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
- GPT-5.2 Pro (gpt-5.2-pro): is the most advanced and powerful GPT-5-series model, designed for maximum intelligence, depth of reasoning, and reliability on the most demanding tasks. Maximal support input is 400 000 tokens. It was trained on data existing before May 2024.
Use these settings to configure Claude provider.
- Enable Claude provider:
- Yes: use global AI settings from Core module.
- No: use Assistant module provider settings below.
- Claude API key: generate an Anthropic API key. You'll need to create an account if you don't have one.
- Default Claude model: choose a most suitable language models for you to utilize:
- Claude Haiku 3.5: the fastest and most affordable model in the 3.5 line; ideal for classification, data extraction, and high-throughput tasks. Max input: 200,000 tokens. Trained on data up to July 2024.
- Claude Sonnet 3.7: "hybrid thinking" (can respond briefly or do more elaborate step-by-step reasoning); versatile for code, analysis, and planning. Max input: 200,000 tokens. Trained on data up to November 2024.
- Claude Sonnet 4: noticeably smarter than 3.7, especially for coding and reasoning; a dependable workhorse for assistants and heavier workloads. Max input: 200,000 tokens (up to 1M in beta with a special header).
- Claude Opus 4: flagship model for complex/long-horizon tasks and agentic scenarios, particularly strong at coding. Max input: 200,000 tokens.
- Claude Opus 4.1: tailored for specialized, difficult problems with advanced reasoning. Max input: 200,000 tokens.
- Claude Haiku 4.5: fastest and most affordable model in the 4.5 line; optimized for low-latency, high-throughput tasks such as classification, summarization, and data extraction. Max input: 200,000 tokens. Trained on data up to November 2024.
- Claude Sonnet 4.5: balanced, high-intelligence model with improved reasoning and coding performance; suitable for analysis, planning, software development, and advanced assistants. Max input: 200,000 tokens (up to 1M tokens in beta with a special header). Trained on data up to November 2024.
- Claude Opus 4.5: flagship model for complex, long-horizon, and agentic tasks; excels in deep reasoning, large codebases, and research workflows. Max input: 200,000 tokens. Trained on data up to November 2024.
Use these settings to configure Gemini provider.
- Enable Gemini provider:
- Yes: use global AI settings from Core module.
- No: use Assistant module provider settings below.
- Gemini API key: generate an Google AI Studio API key. You'll need to create an account if you don't have one.
- Default Gemini model: choose a most suitable language models for you to utilize:
- Gemini 2.0 Flash Lite: the "small" fast option for low-latency, low-cost tasks. Max input: 1,048,576 tokens; output 8,192.
- Gemini 2.0 Flash: a versatile workhorse of the second generation with a 1M context and high speed. Max input: 1,048,576; output 8,192.
- Gemini 2.5 Flash Lite: the fastest 2.5-Flash; optimized for cost/speed in large-scale pipelines. Max input: 1,048,576; output 65,536.
- Gemini 2.5 Flash: best price/performance in the 2.5 series; good for high-volume, "thinking", and agentic use cases. Max input: 1,048,576; output 65,536.
- Gemini 2.5 Pro: the most capable Gemini for complex reasoning and coding; supports "thinking", grounding, and more. Max input: 1,048,576**; output 65,536.
- Gemini 3 Flash: fast, efficient third-generation model optimized for low-latency and high-throughput workloads; suitable for real-time assistants, summarization, classification, and large-scale pipelines. Max input: 1,048,576 tokens; output 65,536.
- Gemini 3 Pro: most capable Gemini 3 model, designed for advanced reasoning, complex coding, and agentic workflows; delivers improved accuracy, stronger “thinking” capabilities, and reliable long-context performance. Max input: 1,048,576 tokens; output 65,536.