| | Last week I built a local pipeline where a state machine + LLM watches my security cam and yells at Amazon drivers peeing on my house. State machine is the magic: it flips the system from passive (just watching) to active (video/audio ingest + ~1s TTS out) only when a trigger hits. Keeps things deterministic and way more reliable than letting the LLM run solo. LLM handles the fuzzy stuff (vision + reasoning) while the state machine handles control flow. Together it’s solid. Could just as easily be swapped to spot trespassing, log deliveries, or recognize gestures. TL;DR: gave my camera a brain and a mouth + a state machines to keep it focused. Repo in comments to see how it’s wired up. submitted by /u/Weary-Wing-6806 |
People like my img2img workflow so it wasn't much work to adapt it to just…
submitted by /u/ThetaCursed [link] [comments]
Step up to the oche and hit the bull’s-eye with this automatic darts scoring system…
made a short video with LTX-2 using an iCloRA Flow to recreate a Space Jam…
The composition of objects and their parts, along with object-object positional relationships, provides a rich…
Today, we’re announcing structured outputs on Amazon Bedrock—a capability that fundamentally transforms how you can…