At 25 cameras and 1080p, you are pushing 24 terabytes a month upstream just to do detection in the cloud. Edge inference collapses that to single digit gigabytes.
The bandwidth math on a 16 to 25 camera multifamily property is the part of the cloud-AI pitch that vendors prefer to skip. At standard 1080p H.264 settings, every camera pushes 2 to 4 Mbps continuously. Multiply by 25 and you have a property that needs a sustained 75 Mbps outbound, which the building's DSL or cable-business uplink may or may not have, and which costs roughly 1,200 dollars per month per property in cloud egress. Edge inference puts the model on a small box at the property, keeps the streams on the LAN, and ships only the structured event when something happens. This page is the math.
See the bandwidth profile of an edge unitThe number the cloud-AI pitch leaves out
The headline cost of cloud-AI surveillance is the per-camera monthly fee. The number that disappears from the slide deck is the egress bandwidth. Twenty five cameras at 3 Mbps continuous is 75 Mbps outbound. Sustained for a month, that is 24,300 gigabytes, or roughly 24 terabytes. At a 0.05 dollars per gigabyte rate (typical for cloud egress) that is 1,200 dollars in transit alone, every month, every property.
The cloud also has to store those streams long enough to do forensic review, which most operators expect. At 30 day retention that is another 24 terabytes of cloud storage. Multiply by a twenty property portfolio and the operator is on the hook for roughly 480 terabytes of cloud storage and 480 terabytes of monthly egress, every month. None of those numbers appear in the sales conversation; they appear in the AWS bill three months after deployment.
Edge inference makes those numbers go away. The cameras stay on the LAN where the LAN is already paid for. The recorder keeps doing its existing 24x7 capture (also already paid for). Only the structured events leave the property, and a structured event is roughly 600 bytes plus a small thumbnail. The economics invert.
The real bandwidth profile, side by side
The shape of the comparison is not subtle. Cloud-AI pushes every frame upstream forever. Edge AI pushes only events. The difference is roughly four orders of magnitude in monthly traffic.
| Feature | Cloud AI on the same 25 camera property | Edge inference (Cyrano style) |
|---|---|---|
| Per camera continuous Mbps | 2 to 4 Mbps (1080p H.264) | 0 (streams stay on LAN) |
| Property outbound, sustained | Roughly 75 Mbps continuous | Roughly 0.001 Mbps (events only) |
| Monthly outbound traffic | About 24 TB per property | Single digit GB per property |
| Cloud egress cost at 0.05 dollars / GB | Roughly 1,200 dollars per property | Under 5 dollars per property |
| Cloud storage at 30 day retention | About 24 TB | Roughly 50 to 200 MB of thumbnails |
| Behavior on a building uplink outage | Detection stops; no frames arrive | Detection continues; events queued in local outbox |
| Forensic clip retrieval | Pull from cloud (assumes the link held) | Pull from existing on-prem recorder |
What an edge unit on a 25 camera property looks like
The hardware is unglamorous on purpose. A fanless mini PC with an HDMI capture card, sitting on top of the rack next to the recorder. The recorder's HDMI feed (the one that drives the wall monitor) goes into the capture card. The mini PC runs the detection model on whatever subset of the grid is currently visible.
Where the bytes go
Camera
1080p H.264 over PoE to the recorder. Same as it ever was. No change.
Recorder
24x7 capture to the disk array. HDMI out drives the wall monitor and the edge box.
Edge box
HDMI in; layout detect; per-tile model inference; per-zone and per-dwell rules.
- 4
Outbound
Only structured events leave the property. ~600 bytes per event plus a small JPG thumbnail.
The numbers in real units
Concrete numbers help here. Below are typical values for a real 25 camera multifamily property running 1080p H.264 at 15 fps, with normal motion levels.
The edge variant outbound is dominated by event JSON, thumbnails, and a daily heartbeat. None of that is sensitive to camera count in the way streaming is; doubling the cameras roughly doubles the event volume, not the byte volume per event.
What survives a property uplink outage on the edge variant
This is the part that the bandwidth pitch alone misses. The reason most operators care about edge inference is not just the bandwidth bill; it is that the building's uplink is genuinely unreliable. A two hour cable outage on a Saturday night is the test that separates real edge from cloud with a local accelerator. The edge variant passes; the cloud variant does not.
continues during a property uplink outage:
- Camera capture and recorder writes
- Edge box inference on the HDMI feed
- Per-zone and per-dwell rule evaluation
- Append events to local outbox with monotonic counter
- Forensic record on the on-prem disk array
deferred until the link returns:
- WhatsApp / SMS delivery (drained on reconnect, in order)
- Dashboard push updates (backfill on reconnect)
- Cloud thumbnail backup
“During a partition drill on a 16 camera multifamily property (uplink pulled for six hours), 55 events entered the local outbox. All 55 drained in strict order in 2.14 seconds when the link came back. Zero dropped, zero server-side duplicates. The cloud-AI alternative would have generated zero events during the same window because no frames would have reached inference.”
Cyrano field notes, partition drill on a production property
When a hybrid makes sense
Worth being honest: a hybrid where event thumbnails are mirrored to a cloud bucket for off-site backup is sometimes the right architecture, especially for portfolios that want a single web dashboard across properties. The edge unit drives the inference locally; the cloud is just an off-site copy of the events and thumbnails. That hybrid keeps the bandwidth math collapsed (still single digit GB outbound) and gets the operator the cross-property dashboard.
What hybrid does not mean is mirroring full pixel streams to the cloud. The full pixel history belongs on the on-prem recorder. Pulling it to the cloud reintroduces the 24 TB per month per property number that the edge variant exists to avoid.
The collapsed economics
Pixels stay on the property. Events go to the cloud.
That is the whole architectural rule. Pixels are expensive (TB per month per property). Events are cheap (KB per event). When the architecture splits them cleanly (pixels on prem, events to the cloud), the bandwidth and storage economics work, and the WAN-outage behavior works. The 16 to 25 camera property is where the rule pays back fastest.
Run the bandwidth math on your portfolio
A 15 minute call. We sit down with the camera count and the recorder type for one of your properties and walk through the cloud vs edge bandwidth and storage on real numbers.
Edge camera bandwidth on multifamily: frequently asked questions
How much bandwidth does one 1080p camera actually push at typical settings?
A 1080p H.264 stream at 15 fps with reasonable scene motion lands in the 2 to 4 Mbps range, peaking higher when there is a lot of motion or detail. H.265 cuts that roughly in half (1 to 2 Mbps for the same scene) but only some recorders and cameras decode it cleanly without dropped frames. A 4K stream at the same fps is in the 8 to 12 Mbps range. Most multifamily fleets run 1080p H.264 because it is the path of least resistance with five year old recorders.
What does that look like on a 25 camera property when everything is streamed to the cloud?
Twenty five cameras at 3 Mbps average is 75 Mbps continuous outbound from the property. Continuous outbound at that rate is 24 TB per month per property. At a typical 0.05 dollars per GB cloud egress, that is 1,200 dollars per month per property in transit alone, before storage. Storage on the cloud side at 30 day retention is another 24 TB at the cloud's data rate. Most multifamily portfolios cannot absorb that, which is why most cloud-AI pitches quietly assume sub-720p streams or motion-only upload, both of which damage the detection workload.
What does edge inference on a property-side box actually save?
Almost all of it. The 25 streams stay inside the property network, where they are essentially free (the LAN is already paid for). The edge box reads the recorder's HDMI output (the same image the wall monitor shows), runs detection locally, and sends only the small structured event payload (a few hundred bytes per event plus a thumbnail image) to the operator. The transit drops from 24 TB per month to single digit GB per month. The cloud storage drops from 24 TB to roughly 50 to 200 MB per month of thumbnails. The forensic record (full pixel history) lives on the existing on-prem recorder, where it always was.
Is the recorder on-prem doing the storage work that the cloud was supposed to do?
Yes, and that is the point. Most multifamily properties already have a recorder with a four to twelve TB hard drive that is doing 24x7 H.264 capture and rotating at 15 to 45 days. That investment is sunk. The edge AI overlay does not replace the recorder; it sits next to it, reads its HDMI output, and adds the intelligence layer. The recorder is still where forensic clips come from. The edge box is the new component, and it is small (a fanless mini PC), cheap (low four figures upfront), and offline-tolerant.
Why is the 16 to 25 camera band the one where edge economics matter most?
Below 16, the numbers are small enough that a cloud variant can be afforded by the operator if they really want one (you are talking about a few hundred dollars a month in transit, not thousands). Above 25, multiple recorders and multiple edge boxes are usually in play, and the edge architecture is the only one that scales because it parallelizes per recorder. The 16 to 25 band is the sweet spot where one recorder and one edge box cover the whole property and the cloud math is genuinely punitive.
What about WAN outages on the operator side?
An edge box keeps detecting during the outage because the model is on the box and the camera feed never crossed the WAN to begin with. Events get appended to a local file with a monotonic counter (one line of JSON per event). When the link returns, the drain worker walks the file forward and posts the events in strict order. The dashboard backfills, the WhatsApp messages catch up, and nothing is dropped. Cloud-AI architectures cannot do this because the inference itself is on the other side of the dead link.
Comments (••)
Leave a comment to see what others are saying.Public and anonymous. No signup.