Sys-Arc-Visl/scripts/setup-ollama-models.ps1

38 lines
1.5 KiB
PowerShell

# Check if Ollama is installed
if (-not (Get-Command ollama -ErrorAction SilentlyContinue)) {
Write-Host "Error: Ollama is not installed or not in your PATH." -ForegroundColor Red
Write-Host "Please install Ollama from https://ollama.com/"
exit 1
}
Write-Host "Checking local Ollama service status..." -ForegroundColor Cyan
try {
$status = Invoke-RestMethod -Uri "http://localhost:11434/api/tags" -ErrorAction Stop
Write-Host "Ollama service is running!" -ForegroundColor Green
} catch {
Write-Host "Error: Could not connect to Ollama at http://localhost:11434" -ForegroundColor Red
Write-Host "Please ensure 'ollama serve' is running in another terminal."
exit 1
}
# Define recommended models
$models = @(
@{ Name = "moondream"; Desc = "Tiny, fast vision model (1.7GB). Best for low-end hardware." },
@{ Name = "llava-phi3"; Desc = "High-performance small vision model (2.3GB). Good balance." },
@{ Name = "llama3"; Desc = "Fast standard LLM for text logic (4.7GB)." }
)
Write-Host "`nReady to pull recommended models:" -ForegroundColor Yellow
foreach ($m in $models) {
Write-Host " - $($m.Name): $($m.Desc)"
}
$confirm = Read-Host "`nDo you want to pull these models now? (Y/n)"
if ($confirm -eq 'n') { exit }
foreach ($m in $models) {
Write-Host "`nPulling $($m.Name)..." -ForegroundColor Cyan
ollama pull $m.Name
}
Write-Host "`nAll models ready! You can now select them in the KV-Graph settings." -ForegroundColor Green