160 points by aegis_camera about 18 hours ago | 147 comments | View on ycombinator
psyclobe about 17 hours ago |
0xbadcafebee about 17 hours ago |
- This is a benchmark for "home security" workflows. I.e., extremely simple tasks that even open weight models from a year ago could handle.
- They're only comparing recent Qwen models to SOTA. Recent Qwen models are actually significantly slower than older Qwen models, and other open weight model families.
- Specific tasks do better with specific models. Are you doing VL? There's lots of tiny VL models now that will be faster and more accurate than small Qwen models. Are you doing multiple languages? Qwen supports many languages but none of them well. Need deep knowledge? Any really big model today will do, or you can use RAG. Need reasoning? Qwen (and some others) love to reason, often too much. They mention Qwen taking 435ms to first token, which is slow compared to some other models.
Yes, Qwen 3.5 is very capable. But there will never be one model that does everything the best. You get better results by picking specific models for specific tasks, designing good prompts, and using a good harness.
And you definitely do not need an M5 mac for all of this. Even a capable PC laptop from 2 years ago can do all this. Everyone's really excited for the latest toys, and that's fine, but please don't let people trick you into thinking you need the latest toys. Even a smartphone can do a lot of these tasks with local AI.
aegis_camera about 18 hours ago |
hparadiz about 18 hours ago |
simonw about 14 hours ago |
https://github.com/SharpAI/DeepCamera/blob/c7e9ddda012ad3f8e...
jjcm about 16 hours ago |
This is the classic issue in tech right now - it's becoming easier to build the systems, but the compliance/legal hurdles are still real, slow, and human. Even if the monitoring is best in class (which I'd argue it likely is - this is a fantastic application of AI), if the compliance isn't there it wont be a real product.
infecto about 17 hours ago |
loloquwowndueo about 17 hours ago |
undefined about 16 hours ago |
aimarketintel about 12 hours ago |
For example, your local Qwen model can call an MCP server that queries 9 public APIs (GitHub, HN, arXiv, npm, etc.) and returns structured JSON. The model does the reasoning locally, but gets real-time data from the web — best of both worlds.
Richard_Jiang about 9 hours ago |
Havoc about 18 hours ago |
alcazar about 18 hours ago |
jamesponddotco about 15 hours ago |
carlgreene about 17 hours ago |
gos9 about 13 hours ago |
nubg about 17 hours ago |
goldenarm about 18 hours ago |
wrcwill about 15 hours ago |
the analysis is very suspicious: “gpt 5 mini had api failures due to wrong temp setting”? wtf?
whatever you used to slop your benchmark didt even take the time to set the temp to 1 (which the docs say is required)
tristor about 17 hours ago |
llm_nerd about 17 hours ago |
Seems like trying to make a need from the tools. My security system front page shows me every event that happened at my house, and I don't have to interrogate it on every happenstance, and I don't see what the value of that is.
bigyabai about 18 hours ago |
Why would you run this on your M5 instead of a dedicated machine for it? A Jetson Orin would be faster at prefill and decode, as well as cheaper for home installation.
aplomb1026 about 17 hours ago |
rodchalski about 16 hours ago |
ieie3366 about 18 hours ago |
DGAP about 17 hours ago |
dmonterocrespo about 15 hours ago |
still-learning about 15 hours ago |
Machine hardware evolution is slowing down, pretty soon you can buy one big ass server that will last potentially decades as it would be purpose built for ai.
Things like 'context based home security' yeah thats just, automatic, free, part of the ai system.
Everyone will talk to the ai through their phones and it'll be connected to the house, it'll have lineage info of the family may be passed down through generations etc, and it'll all be 100% owned, offline, for the family; a forever assistant just there.