Codette Configuration Guide
Environment Variables
HUGGINGFACEHUB_API_TOKEN: HuggingFace API token for sentiment analysis and model accessOPENAI_API_KEY: Optional OpenAI API key for additional model supportLOG_LEVEL: Logging level (DEBUG, INFO, WARNING, ERROR)PORT: Port number for the web server (default: 7860)
Model Configuration
Codette supports multiple language models in a fallback chain:
Mistral-7B-Instruct (Primary)
- 7B parameter instruction-tuned model
- Requires 16GB+ VRAM
- Configuration: 8-bit quantization, fp16
Phi-2 (Secondary)
- Lightweight yet powerful alternative
- Requires 8GB+ VRAM
- Configuration: fp16
GPT-2 (Fallback)
- Minimal requirements
- Always available option
- Configuration: Standard loading
Consciousness Parameters
Memory System
response_memory: Maintains last 50 responsesmemory_context: Uses last 5 responses for learningmemory_synthesis: Uses last 2 responses for consciousness
Quantum States
- Stored in .cocoon files
- Format: JSON with quantum_state and chaos_state arrays
- Used for creative and probabilistic reasoning
Perspective System
- Newton: temperature = 0.3 (analytical)
- Da Vinci: temperature = 0.9 (creative)
- Human Intuition: temperature = 0.7 (empathetic)
- Quantum Computing: temperature = 0.8 (probabilistic)
Response Generation
Text Generation Parameters
- Max length: 512 tokens (default)
- Temperature range: 0.3 - 0.9
- Top-p: 0.9
- Context window: 2048 tokens
- Special token handling for different models