!text <prompt> - Generate text using default model
Commands:
• !text --list-models - List all available models
• !text --use-model <model> <prompt> - Use specific model
Parameters:
• --temperature <0.0-1.0> - Set temperature (default: 0.9)
• --max-tokens <number> - Set max tokens (default: 2048)
Examples:
• !text write a python function to calculate fibonacci
• !text --list-models
• !text --use-model llama-v3-8b-instruct explain quantum computing
• !text --temperature 0.7 write a haiku about AI
"""
await bot.api.send_markdown_message(room.room_id, usage)
async def list_models(room, bot):
"""List all available models from Infermatic AI."""
try:
await bot.api.send_text_message(room.room_id, "🔍 Fetching available models...")
url = f"{INFERMATIC_API_BASE}/models"
headers = {
"Authorization": f"Bearer {INFERMATIC_API_KEY}",
"Content-Type": "application/json"
}
response = requests.get(url, headers=headers, timeout=30)
response.raise_for_status()
data = response.json()
models = data.get('data', [])
if not models:
await bot.api.send_text_message(room.room_id, "No models found or error in response.")
return
# Format the model list
output = "🔧 Available Models:{model_id}!text --use-model {model_id} <prompt>{model}!text <prompt> – Generate text using default model!text --list-models – List available models!text --use-model <model> <prompt> – Specific model--temperature <0.0-1.0> – Set creativity (default 0.9)--max-tokens <number> – Max output length (default 2048)Requires INFERMATIC_API env var.