Location: SensAI > AI Settings
The AI Settings page controls how SensAI responds, maintains context, and manages costs. This is also where you connect your OpenAI API key.

OpenAI API Key
↑ Back to topPurpose:
Connect SensAI to OpenAI so the AI can generate responses.
Setup Instructions:
Paste your secret key starting with sk-.
Important:
Keep this key private. Do not share it publicly or include it in screenshots.
AI Model
↑ Back to topDescription:
Select which OpenAI model SensAI will use to generate responses.
Options:
- GPT-3.5 Turbo (Recommended): Fast, cost-effective, high-quality
- GPT-4: More accurate, but ~10× higher cost
- GPT-4 Turbo: Latest model, high quality, similar cost to GPT-4
Recommended Use:
- GPT-3.5 Turbo → Most websites (balanced speed, cost, and accuracy)
- GPT-4 / GPT-4 Turbo → Complex content or maximum accuracy required
Response Length
↑ Back to topPurpose:
Controls how detailed AI answers are.
Options:
- Short (256 tokens): Quick, concise answers
- Medium (512 tokens): Balanced responses (recommended)
- Long (1024 tokens): Detailed explanations
- Very Long (2048 tokens): Comprehensive, in-depth responses
Guidelines:
- Short: Simple product info, basic questions
- Medium: Default for most sites
- Long / Very Long: Tutorials, complex topics, technical content
Note: Longer responses take more time and increase costs.
Conversation Memory
↑ Back to topPurpose:
Determines how many previous messages the AI remembers in a conversation.
Options:
- Short (3 messages): Cost-effective, good for simple Q&A
- Medium (5 messages): Balanced (recommended)
- Long (8 messages): Best for complex conversations
How it works:
The AI uses the last N messages to maintain context and provide coherent responses.
Cost Consideration:
Longer memory = more tokens = higher cost per response.
Cache Expiration
↑ Back to topPurpose:
Stores AI responses temporarily to save time and reduce API usage.
Options:
- 24 hours – Fresh content, moderate savings
- 1 week (168 hours) – Recommended, good balance
- 2–4 weeks (336–672 hours) – Maximum savings for stable content
How it works:
If someone asks a question that has already been answered within the cache period, the AI serves the cached response instantly without making a new API call.
When to use shorter cache:
- Content changes frequently
- Prices or availability update often
- Time-sensitive information
When to use longer cache:
- Content is stable
- Cost savings are important
- High volume of repeated questions