TL;DR: Alysium's model selector takes a plain-language description of what your agent does (minimum 3 characters) and returns 3–5 matched model categories — labels like "Deep Reasoning" or "Fast & Affordable" — with ranked options inside each. No benchmark research required. There's a 5-second cooldown and 20-request session cap. After selection, fine-tune with temperature, response length, top-p, and stop sequences.
Choosing an AI model used to mean researching benchmark tables, comparing context windows, and guessing at pricing implications. Alysium's model selector skips all of that.
Here's how it works and how to use it effectively.
How the Model Selector Works
Open any agent in the builder and navigate to the AI Model section. Type a description of what your agent does — at least 3 characters, but a sentence works better. Alysium uses semantic matching to find models whose capabilities align with your use case and returns 3–5 categories with ranked options inside each.
Categories use plain-language labels rather than cryptic version strings. "Deep Reasoning" for models suited to complex analysis. "Fast & Affordable" for high-volume lightweight tasks. Inside each category, models are ranked by fit for your described use case. Pick the category that matches your priority, choose from the ranked options.
There's a 5-second cooldown between searches and a 20-request cap per session — enough to try several phrasings. If the first results don't feel right, try describing your use case differently.
When to Choose Deep Reasoning
Deep reasoning models are for agents where quality of analysis matters more than speed or cost.
Complex advisory agents. A consultant's methodology agent synthesizing multiple framework documents to help a client reason through a specific situation needs more than factual retrieval. If the agent is helping someone make a decision — not just answering a factual question — reasoning capability matters.
Graduate research support. Agents built to help graduate students work through methodology questions, evaluate evidence quality, or design research instruments. These require holding and applying multiple concepts simultaneously.
Writing feedback agents. Agents configured to give substantive multi-point feedback on drafts. Deep reasoning produces more coherent feedback than a fast model that summarizes obvious issues.
Legal case analysis. Law school case analysis companions that need to reason through holdings, distinctions, and applications across multiple documents.
When to Choose Fast & Affordable
Fast, lightweight models suit agents where the interaction pattern is high-volume and questions are relatively factual.
FAQ agents. Your small business answers 50 customer questions a day about hours, pricing, and booking. "Our hours are 9–6pm Monday–Saturday" doesn't require deep reasoning to produce accurately. Fast and affordable handles this at significantly lower credit cost.
Course tutors for factual recall. Students asking "what was in the reading" or "can you explain this concept again" at volume. Factual retrieval and explanation — not synthesis or analysis.
Employee onboarding knowledge bases. New employee FAQ handled by a fast model saves credits without any perceptible quality difference for the interaction type.
Fine-Tuning with Model Parameters
After selecting a model, four parameters let you adjust its behavior:
Temperature controls how creative vs precise responses feel. Lower values (0.1–0.4) produce more consistent, factual responses — better for FAQ agents. Higher values (0.6–0.8) produce more varied, natural-sounding responses — better for coaching or creative agents.
Response length cap limits how long responses can be. Useful for widget contexts where long responses create a poor reading experience.
Top-p controls output diversity. Default values work for most creators; adjust only when response diversity is a specific concern.
Stop sequences define text strings that end a response when generated. Advanced use case — most creators don't need this.
All parameter changes apply immediately with no redeployment required.
The Practical Decision
For most knowledge FAQ agents: start with Fast & Affordable. Test with real questions. If responses feel thin or miss nuance on complex questions, try a reasoning category. The cost difference is meaningful at volume; the quality difference is real for complex tasks.
For advisory or analysis agents where the answer quality is what buyers pay for: start with Deep Reasoning. The credit cost is higher but appropriate when the interaction value justifies it.
Try it now. Open your agent builder and describe your use case in the model selector.
Using Multiple Models Across Your Agents
One of the model selector's most useful applications is running different models on different agents in the same workspace. Your agents don't all need to run the same model, and they shouldn't — they have different use cases and different quality requirements.
A consultant might have three agents in their workspace: a client methodology agent (Deep Reasoning — handles complex advisory questions), a general FAQ agent for prospects (Fast & Affordable — handles pricing and process questions), and a specialized case analysis agent for a specific framework (Deep Reasoning — needs to reason across multiple documents). Three different agents, two different model categories, credit costs matched to interaction value.
The model selector makes this easy. Open each agent, describe what it does, pick the category that fits, choose from the ranked options. Each agent remembers its model configuration independently.
Testing Model Quality Before Committing
When you switch models, Alysium doesn't immediately lock you in. Test the agent in the preview mode with real questions before publishing to your website. Compare the response quality between a fast model and a reasoning model on the same question. For factual FAQ responses, you often won't notice a difference. For complex advisory questions, the gap can be substantial.
The 5-second cooldown and 20-request session cap on the model selector mean you can try 5–10 different use case descriptions in a session to explore the category options. Use that budget to find the best match for each agent type you build.
When the Default Model Is Fine
For first-time builders: if you're not sure which model to pick, the default Alysium assigns is a reasonable starting point. Build the agent, configure the knowledge base and instructions, and test with real questions. If the responses feel adequate, leave the model alone. If they feel vague or miss nuance on complex questions, revisit the model selector and try a reasoning category.
Over-optimizing model selection before you know how your agent is used is premature. Deploy first. Adjust based on real conversation data.
The Role of Temperature in Response Feel
Temperature is the parameter most creators benefit from adjusting first, before any other parameter. The difference between temperature 0.3 and temperature 0.7 is perceptible in conversation quality — one produces responses that feel precise and consistent, the other produces responses that feel more natural and varied.
For a customer FAQ agent: lower temperature. "Your hours are 9–6pm Monday through Saturday, and we're closed on Sundays" should be the same answer every time someone asks about hours. Consistency is the virtue.
For a coaching companion: higher temperature. "When clients explore this question, it often surfaces..." sounds different every time the model generates it at higher temperature, which reads as more natural in a conversational context. Consistency is less important when the interaction style is exploratory rather than factual.
The practical guidance: set temperature, save, ask five test questions, and notice how the responses feel. If they feel robotic, go higher. If they feel unpredictably variable, go lower. The right setting is the one that makes the responses feel like your agent — not a generic AI and not an inconsistent one.
Stop Sequences: Only If You Need Them
Stop sequences are text strings that tell the model to stop generating when it produces them. This is an advanced use case most creators don't need.
Where they're useful: if you've configured an agent to generate structured output in a specific format — for example, always ending with a bracketed resource list — a stop sequence can ensure the response terminates cleanly rather than adding additional prose after the format ends.
For conversational agents, standard response length capping is sufficient. Stop sequences are for specific structured output patterns where you need precise control over where the response ends.
Model Choices by Audience Type
The abstract categories become clearer when you match them to what specific audiences actually build:
Coaches and consultants almost always want a balanced mid-range model — conversational enough to feel warm, capable enough to reason about complex client situations. "Coaching session companion" or "methodology advisor" as your use case description typically surfaces these. Avoid pure "speed" models for coaching agents — the brevity can feel abrupt when someone is asking about a real challenge they're working through.
Educators building course tutors benefit from reasoning-capable models. When a student asks "I don't understand concept X," a reasoning model can trace the misunderstanding rather than restating the definition from the document. For multiple-choice quiz practice or recall exercises, a faster lower-cost model is fine.
Small businesses with customer FAQ agents usually need Fast & Affordable models most. The questions are predictable ("what are your hours," "do you offer refunds"), the answers are in the knowledge base, and speed matters more than nuance. Reserve the deeper models for agents doing something more complex — contract clause analysis, technical troubleshooting, scenario walkthroughs.
Content creators selling AI products through AgentHub should test their agent from a buyer's perspective before publishing. The model you chose during build might feel appropriately fast to you but slow to a first-time buyer expecting instant responses. Check your analytics for conversation dropout patterns — if conversations consistently end after two messages, response latency might be a contributing factor.
When to Change Your Model
Change your model when: your analytics show low helpfulness ratings clustered around certain question types (the model may be under-reasoning), your responses are consistently too long or too short despite parameter adjustments, or you've added significant new knowledge base content that changes the complexity level of likely questions.
Models can be swapped anytime in the AI Model panel. Changes take effect immediately — no redeployment, no reconfiguration. The knowledge base, instructions, and widget settings are all model-agnostic and carry forward unchanged.
Try it now. Open your agent builder — AI Model is in the left navigation. Describe your use case and explore the matched categories.
Frequently Asked Questions
Related Articles
Ready to build?
Turn your expertise into an AI agent — today.
No code. No engineers. Just your knowledge, packaged as an AI that works around the clock.
Get started free