Minime-T5 is a lightweight local AI service using Google's Flan-T5-Small model.It is fast and has low memory usage. It can be run on CPU or GPU:
You can install it within the CHIM.exe launcher under Install Components. The service runs locally on port 8082 and provides several endpoints for different tasks.
1. Action Triggering (/command)
Every user input is analyzed to detect if a specific action should be triggered. Currently detects:
This is especially useful with less sophisticated LLMs that might miss action cues.
2. Memory Detection (/extract & /detectMemory)
Not every conversation needs to trigger a memory lookup. Minime-T5 analyzes text to determine:
You must enable AUTO_CREATE_SUMMARYS in the profile for this to work.
3. Goal/Task Extraction (/task)
Detects when a new goal or mission is being set. For example:
This is the SetCurrentTask functionality that keeps AI NPCs aware of your objectives.
4. Topic Extraction (/topic & /posttopic)
Extracts the subject/topic from sentences for:
5. Scene Analysis (/ambient)
(Experimental)
Analyzes dialogue history to detect scene mood/genre for dynamic responses.
When both Minime-T5and Oghma Infinium are enabled:
Alternative:You can use OGHMA_CUSTOM
instead of Minime-T5 for topic extraction, which uses your configured LLM instead (more accurate but uses tokens).
Text2Vec provides vector embeddings for semantic search in:
Two Modes:
Configuration in Global Settings:
You can view the Text2Vec embeddings under Memory Summaries in the web interface.