-
-
Notifications
You must be signed in to change notification settings - Fork 6
Model Configuration
VaultAI supports fine-tuning AI responses through advanced model configuration parameters. These settings allow you to control the creativity, randomness, and length of generated content.
Controls the randomness of AI responses.
- 0.0: Deterministic, always chooses most likely response
- 0.3-0.5: Balanced, good for factual content
- 0.7-0.9: Creative, good for writing and brainstorming
- 1.0: Maximum creativity and randomness
Default: 0.7
Limits the AI to consider only the top K most probable tokens at each step.
- 1-10: Very focused responses
- 20-30: Balanced variety (recommended)
- 40: Maximum vocabulary diversity
Default: 40
Nucleus sampling - considers tokens that make up the top P probability mass.
- 0.1-0.3: Very focused
- 0.7-0.9: Balanced (recommended)
- 1.0: Considers all tokens
Default: 0.9
Maximum length of generated responses.
- 512: Short responses
- 1024: Medium responses
- 2048: Long responses (recommended)
- 4096-8192: Very long responses
Default: 2048
- Open Settings β Community Plugins β VaultAI
- Navigate to Model Configuration section
- Adjust the sliders or enter values directly
- Settings apply immediately to new conversations
- Temperature: 0.3
- Top-K: 20
- Top-P: 0.8
- Max Tokens: 1024
- Temperature: 0.9
- Top-K: 40
- Top-P: 0.95
- Max Tokens: 2048
- Temperature: 0.5
- Top-K: 30
- Top-P: 0.9
- Max Tokens: 2048
- Temperature: 0.8
- Top-K: 35
- Top-P: 0.9
- Max Tokens: 1024
- Start Conservative: Begin with lower temperature (0.5-0.7) and adjust based on results
- Experiment: Different tasks benefit from different settings
- Balance Parameters: High temperature with low top-K can produce inconsistent results
- Token Limits: Higher token limits allow for more detailed responses but use more API quota
- Save Presets: Create custom presets for different types of work
- Lower temperature (try 0.3-0.5)
- Reduce top-K (try 20)
- Lower top-P (try 0.8)
- Increase temperature (try 0.7-0.9)
- Increase top-K (try 30-40)
- Increase top-P (try 0.95)
- Increase max output tokens
- Break complex requests into smaller parts
- Reduce max output tokens
- Use more focused prompts
- Consider using lower temperature for factual queries
For power users, VaultAI also supports:
- Model Selection: Choose between different Gemini models
- Context Window: Adjust how much conversation history to include
- Streaming: Enable real-time response streaming
- Custom System Prompts: Set behavior guidelines for the AI
See Advanced Settings for more details.
Need Help? Check out our Troubleshooting Guide or join the Community Discussions.
Intelligent AI Writing Assistant for Obsidian
π Home β’ β‘ Getting Started β’ π― Deep Editor β’ β¨ Custom Prompts β’ β¨οΈ Shortcuts β’ π οΈ Installation β’ π‘ Tips & Tricks
οΏ½ Report Issues β’ π‘ Feature Requests β’ οΏ½οΈ Discussions β’ β Troubleshooting
Made with β€οΈ by Neo β’ Transforming note-taking with AI