| | Last week I built a local pipeline where a state machine + LLM watches my security cam and yells at Amazon drivers peeing on my house. State machine is the magic: it flips the system from passive (just watching) to active (video/audio ingest + ~1s TTS out) only when a trigger hits. Keeps things deterministic and way more reliable than letting the LLM run solo. LLM handles the fuzzy stuff (vision + reasoning) while the state machine handles control flow. Together it’s solid. Could just as easily be swapped to spot trespassing, log deliveries, or recognize gestures. TL;DR: gave my camera a brain and a mouth + a state machines to keep it focused. Repo in comments to see how it’s wired up. submitted by /u/Weary-Wing-6806 |
https://x.com/viccpoes/status/2054278218719637925 submitted by /u/Total-Resort-3120 [link] [comments]
Large language models (LLMs) now power everything from customer service bots to autonomous coding agents.
Amazon’s Finance Technology (FinTech) teams build and operate systems for Amazon teams to manage regulatory…
As AI coding agents become deeply embedded in developer workflows, defenders must evolve their definition…
Musk’s lawyers questioned Altman over allegations of deception and his network of financial investments, but…
A technology that surpasses the limitations of existing sensors, which fail to distinguish between water…