64 lines
2.1 KiB
Markdown

# Testing Your MCP with Large Language Models
After completing [[connect-llm|connecting your MCP server]], you're ready to begin testing. While [[put-into-llm|quick start]] covered basic testing, this article details advanced configuration options visible below the input box.
## Model Selection
Switch between different LLMs here. OpenMCP tracks models at the message level, enabling mixed-model testing scenarios.
> Can't find your model? See [[connect-llm|connecting MCP servers]] to add new models.
## System Prompts
Manage and create system prompts through this module:
![](./images/system-prompt.png)
Prompts are stored in `~/.openmcp/nedb/systemPrompt.db` (NeDB format for easy inspection/backup).
## Prompt Generation
Invoke server-provided prompt functions, with results inserted as rich text:
![](./images/prompt.png)
## Resource Access
Call server resource functions - outputs are inserted as formatted content:
![](./images/resource.png)
:::warning Data Persistence
OpenMCP doesn't manage resource persistence! Empty resources after restart indicate the MCP server lacks storage implementation - this is expected behavior.
:::
## Parallel Tool Execution
When enabled (default), models may call multiple tools in one response (e.g., three parallel web searches):
![](./images/parallel-tool-call.png)
Disable for sequential execution.
> Note: Some providers (like Gemini) may force-disable this if they don't fully support OpenAI's parallel calling spec.
## Temperature Control
Controls output randomness:
- Recommended range: 0.6-0.7 for general tasks
- Default: 0.6 (balanced creativity/consistency)
## Context Window
Determines how many prior messages (default: 20) are sent with each request. Count includes:
- User queries
- Model responses
- Tool calls/results
:::danger Minimum Threshold
Values below 20 often break tool call sequencing, causing 400 errors. Start fresh conversations if this occurs.
:::
## Server Timeout
Default 30-second timeout adjustable via Settings → General (global configuration, in seconds). Increase for long-running operations.