Ollama Skills
Run large language models locally. Pull, run, create, and manage AI models on your own hardware.
4 skills
Ollama REST API Integration
Integrate Ollama's REST API into your applications — chat completions, streaming responses, embeddings, and model management endpoints for building local AI-powered features.
Local Embeddings & Semantic Search
Set up local embedding models with Ollama for semantic code search, RAG pipelines, and documentation indexing — all running privately without cloud API dependencies.
GPU & Performance Optimization
Optimize Ollama inference performance — GPU layer allocation, batch processing, context window tuning, concurrent requests, and hardware-specific configuration for fast local AI.
Model Management & Organization
Master Ollama model lifecycle — pulling, listing, copying, removing models, managing storage, and organizing custom model variants for efficient local AI development.