AI Setup Guide

Configure your preferred AI provider for metadata generation.

Choose Your AI Provider

Loki Studio supports multiple AI providers for generating titles, descriptions, tags, and thumbnail text. Choose the option that best fits your needs:

🧠

Grimnir (Recommended)

Pro Feature

Built-in LLM that runs entirely on your machine. No internet required after download. Best quality/privacy balance.

  • One-time download, use forever
  • No API costs
  • 100% offline capable
  • Multiple model choices
🦙

Ollama / LM Studio

Free for All Users

Use external local LLM servers. Great if you already run Ollama or prefer managing models separately.

  • Free to use
  • Choose any model
  • Local processing
  • Requires separate install
☁️

OpenAI / Claude

Pro Feature

Cloud AI for highest quality. Use GPT-4 or Claude with your own API keys. Best for premium content.

  • Highest quality output
  • Always up-to-date models
  • Requires API key
  • Pay-per-use costs

Grimnir Setup (Built-in LLM)

Grimnir is Loki Studio's integrated AI engine, powered by the Skuld module. It runs state-of-the-art language models directly on your GPU.

Step 1: Select Grimnir Mode

  1. Go to Application Settings (Yggdrasil tab)
  2. Click the LLM Providers sub-tab
  3. In the LLM Mode section, select "Grimnir (Built-in LLM)"

Step 2: Download a Model

Grimnir supports multiple models optimized for different use cases:

Model Speed Quality Best For
qwen3:8b Fast 6.8/10 BEST overall - All languages
nemotron:30b Medium 6.1/10 Excellent descriptions (20GB+ VRAM)
gemma2:9b Fast 6.1/10 Consistent multilingual
gpt-oss:20b Slow 5.5/10 Good vocabulary variety
qwen2.5-coder:7b Fast 5.5/10 Tech/coding content only
deepseek-r1:7b Slow 4.0/10 Not recommended
llama3.2:3b Very Fast 3.5/10 Not recommended
  1. Click a model button to start downloading
  2. Wait for the download to complete (size varies by model)
  3. The status will change to "Ready: [model-name]" in green

Step 3: Configure Max Tokens

The Max Tokens setting controls how much "thinking time" the AI gets. For reasoning models:

  • Lower values (4K-8K): Faster responses, shorter content
  • Higher values (16K-32K): More detailed, better quality
  • Maximum (200K): Full reasoning capability for complex content
Recommendation: Start with qwen3:8b at 16K tokens. Adjust based on your results and generation speed.

Ollama Setup

Ollama is a free, open-source tool for running LLMs locally. It's available to all Loki Studio users.

Step 1: Install Ollama

  1. Download Ollama from ollama.ai
  2. Run the installer
  3. Ollama runs as a background service on port 11434

Step 2: Download a Model

Open a terminal/command prompt and run:

ollama pull qwen3:8b

Or install directly from Loki Studio using the quick-install buttons.

Step 3: Configure in Loki Studio

  1. Go to Application Settings > LLM Providers
  2. Select "Bring Your Own (Third-Party)" mode
  3. Choose Ollama (Local) from the provider dropdown
  4. Base URL should be http://localhost:11434
  5. Select your model from the dropdown (click Refresh if needed)

LM Studio Setup

LM Studio provides a graphical interface for managing local LLMs with an OpenAI-compatible API.

Step 1: Install LM Studio

  1. Download from lmstudio.ai
  2. Install and launch the application
  3. Download your preferred model through the LM Studio interface

Step 2: Start the Local Server

  1. In LM Studio, go to the Local Server tab
  2. Select your model
  3. Click Start Server
  4. Note the URL (usually http://localhost:1234)

Step 3: Configure in Loki Studio

  1. Go to Application Settings > LLM Providers
  2. Select "Bring Your Own (Third-Party)" mode
  3. Choose LM Studio (Local) from the provider dropdown
  4. Enter the LM Studio server URL

OpenAI / Claude Setup

For the highest quality AI output, you can use cloud-based models from OpenAI or Anthropic. This is a Pro feature.

OpenAI (GPT-4) Setup

  1. Create an account at platform.openai.com
  2. Navigate to API Keys in your account settings
  3. Create a new API key and copy it
  4. In Loki Studio, go to Application Settings > LLM Providers
  5. Select "Bring Your Own (Third-Party)" mode
  6. Choose OpenAI (GPT-4, GPT-3.5)
  7. Paste your API key
  8. Select your preferred model (GPT-4 recommended for quality)

Claude (Anthropic) Setup

  1. Create an account at console.anthropic.com
  2. Navigate to API Keys
  3. Create a new API key and copy it
  4. In Loki Studio, go to Application Settings > LLM Providers
  5. Select "Bring Your Own (Third-Party)" mode
  6. Choose Claude (Anthropic)
  7. Paste your API key
Cost Note: Cloud AI providers charge per token. Monitor your usage to avoid unexpected bills. A typical video's metadata generation costs $0.01-0.05 with GPT-4.

Language Recommendations

Different models perform better with different languages:

  • All Languages: qwen3:8b (recommended for all content)
  • English Only: Any model works well
  • Japanese/Korean/Chinese: qwen3:8b or gemma2:9b
  • Russian/Hindi: qwen3:8b
  • Tech/Coding Content: qwen2.5-coder:7b (English only)
  • High VRAM (20GB+): nemotron:30b for premium quality

Troubleshooting

Model download fails or is stuck

Check your internet connection and available disk space. Models can be 4-20GB. Try clicking the download button again to resume.

Ollama not connecting

Ensure Ollama is running (check system tray on Windows). Verify the URL is correct: http://localhost:11434. Try restarting Ollama.

OpenAI/Claude API errors

Verify your API key is correct and has available credits. Check that you've selected a valid model. Ensure your account isn't rate-limited.

Generation is very slow

Try a smaller model or reduce Max Tokens. Ensure your GPU drivers are up to date. For Grimnir, verify CUDA is being used (check device setting).

Related Topics

Need More Help?

Can't find what you're looking for? Join our Discord community for help. He personally responds to every question.

Join Discord
Buy me a coffee