canvas-website/worker
Jeff Emmett 30e2219551 feat: add Ollama private AI integration with model selection
- Add Ollama as priority AI provider (FREE, self-hosted)
- Add model selection UI in Settings dialog
- Support for multiple models: Llama 3.1 70B, Devstral, Qwen Coder, etc.
- Ollama server configured at http://159.195.32.209:11434
- Models dropdown shows quality vs speed tradeoffs
- Falls back to RunPod/cloud providers when Ollama unavailable

Models available:
- llama3.1:70b (Best quality, ~7s)
- devstral (Best for coding agents)
- qwen2.5-coder:7b (Fast coding)
- llama3.1:8b (Balanced)
- llama3.2:3b (Fastest)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-11-26 14:47:07 -08:00
..
shapes pin object, fix fathom, and a bunch of other things 2025-11-11 22:32:36 -08:00
AutomergeDurableObject.ts feat: add Ollama private AI integration with model selection 2025-11-26 14:47:07 -08:00
assetUploads.ts fix asset upload rendering errors 2025-03-19 18:30:15 -07:00
types.ts automerge, obsidian/quartz, transcribe attempt, fix AI APIs 2025-09-21 11:43:06 +02:00
worker.ts fix: implement real-time Automerge sync across clients 2025-11-16 19:45:36 -07:00