#

ollama

3 articles tagged with this topic.

Studio 8H in a Database — How the Doodle Cast Audience Got Built
AI Engineering

Studio 8H in a Database — How the Doodle Cast Audience Got Built

News-desk comedy needs an audience. Sitcom laugh tracks were wrong. Dense festival crowds were wrong. Weekend Update was right — and getting there took nine reaction types, a multi-take stem library, a density dial, a multi-model LLM planner, and one weirdly specific trick to make the goodbye bleed into the outro.

WillWin: predicting five championships nightly with one local LLM
AI Engineering

WillWin: predicting five championships nightly with one local LLM

WillWin is a personal technical experiment — a single open-weights 32B model running on a local RTX 5090 reads public data sources and publishes nightly probability estimates for the 2026 World Cup, Eurovision, F1, Tour de France, and the Oscars. Not a prediction service, not betting advice.

How We Built the TheDoodleCast AI Chatbot
AI Engineering

How We Built the TheDoodleCast AI Chatbot

How we built a real-time AI chatbot for TheDoodleCast using Ollama, a local RTX 5090, Cloudflare Tunnel, and a custom streaming UI — without paying per token.