Skip to main content

How Object Detection Actually Works on a Security Camera

If you’ve ever owned a motion-triggered security camera, you know the failure mode: a tree branch sways, a moth lands on the lens, a cloud passes overhead, and your phone buzzes for the fortieth time today. You stop looking at the alerts. The camera is technically working and practically useless.

Object detection is what fixes that. Here’s what’s actually happening when a modern camera says “person detected” instead of “something moved.”

Motion vs. objects #

A motion-only camera is doing something simple: comparing this frame to the last one and asking did enough pixels change? That’s it. It can’t tell a person from a squirrel from a headlight glare on the garage door. Anything that crosses some threshold of pixel-change is an alert.

Object detection asks a fundamentally different question: what is in those pixels?

What’s under the hood #

Every few frames, the camera (or a small computer it’s feeding) runs the image through a neural network — a model that’s been trained on millions of labeled photos of people, cars, dogs, packages, and so on. The model’s job is to draw a box around anything it recognizes and stick a label on it with a confidence score.

So instead of “motion at 11:42pm,” you get something like:

  • person, 94% in the driveway
  • car, 88% at the curb
  • dog, 71% near the bushes

Your alert rule then becomes a sentence a normal human would write: tell me when there’s a person in the driveway between 10pm and 6am. Squirrels, swaying branches, and your neighbor’s cat stop making your phone buzz.

Where the detection runs #

This is the part that matters for privacy.

The cloud-camera approach (Ring, Nest, the usual suspects) is to upload your footage to a vendor’s servers and run the detection there. That works, but every clip your camera records lives, at least briefly, on someone else’s hard drive — and the labels and timestamps live there longer.

The local approach is to run the same kind of model on a small piece of hardware in your own house. A Coral TPU the size of a thumb drive can do real-time detection on several camera streams at once. Software like Frigate ties it together: cameras feed in, the model runs locally, recordings stay on your NAS, and the only thing that leaves the house is a notification to your phone — and only when you said it should.

Same intelligence. None of the surveillance side-effects.

The honest limits #

Object detection is good, not magic.

  • Lighting matters. Night footage is harder than daytime. Infrared helps but flattens detail.
  • Categories are coarse. A typical model knows “car” but not “your car” — though some setups can be trained to recognize specific vehicles or familiar faces, all on-device.
  • False positives shrink, but don’t go to zero. A mannequin in a window, a person on a billboard across the street, a very confident raccoon — these still trip things up occasionally. The difference is occasionally instead of constantly.

The goal isn’t perfection. It’s getting your alerts down to the handful that actually matter, so you start paying attention to them again.

Putting it together #

A camera setup that uses local object detection ends up looking pretty unremarkable from the outside: cameras on the eaves, a small box in a closet, an app on your phone. What’s different is what’s happening inside — your footage staying on your hardware, a model you control deciding what’s worth telling you about, and zero monthly fees flowing to a company whose business model depends on your data.

If that’s the kind of setup you’ve been trying to picture, that’s exactly what Security Cameras & Local Surveillance is. Or if you’ve already got cameras and want to graduate them past the “everything triggers an alert” stage, drop me a line.

— Dana