| ▲ | state_less 9 hours ago |
| Interesting, the pacing seemed very slow when conversing in english, but when I spoke to it in spanish, it sounded much faster. It's really impressive that these models are going to be able to do real time translation and much more. The Chinese are going to end up owning the AI market if the American labs don't start competing on open weights. Americans may end up in a situation where they have some $1000-2000 device at home with an open Chinese model running on it, if they care about privacy or owning their data. What a turn of events! |
|
| ▲ | mbac32768 5 hours ago | parent | next [-] |
| sitting here in the US, reading that China is strongly urging the adoption of Linux and pushing for open CPU architectures like RISC-V and also self-hosted open models are we the baddies?? |
| |
| ▲ | BobbyJo 3 hours ago | parent [-] | | If there is a walled garden, and you aren't in it, you'll probably push for the walls to come down. No moral basis needed. | | |
| ▲ | h4ny 41 minutes ago | parent | next [-] | | Could you elaborate on what you mean by "moral basis" in your comment? | | |
| ▲ | BobbyJo 12 minutes ago | parent [-] | | I mean China's push for open weights/source/architecture probably has more to do with them wanting legal access to markets than it does with those things being morally superior. |
| |
| ▲ | vachina 3 hours ago | parent | prev [-] | | silent majority of HN |
|
|
|
| ▲ | tedivm 9 hours ago | parent | prev | next [-] |
| This is exactly what I do. I have two 3090s at home, with Qwen3 on it. This is tied into my Home Assistant install, and I use esp32 devices as voice satellites. It works shockingly well. |
| |
| ▲ | mrandish 8 hours ago | parent | next [-] | | I run Home Assistant on an RPi4 and have an ESP32-based Core2 with mic (https://shop.m5stack.com/products/m5stack-core2-esp32-iot-de...), along with a 16GB 4070 Ti Super in an always-on Windows system I only use for occasional gaming and serving media. I'd love to set up something like you have. Can you recommended a starting place, or ideally, a step-by-step tutorial? I've never set up any AI system. Would you say setting up such a self-hosted AI is at a point now where an AI novice can get an AI system installed and integrated with an existing Home Assistant install in a couple hours? | | |
| ▲ | Implicated 6 hours ago | parent [-] | | I mean - the AI itself will help you get all that setup. Claude code is your friend. I run proxmox on an old Dell R710 in my closet that hosts my homeassistant (amongst others) VM and then I've setup my "gaming" PC (which hasn't done any gaming in quite some time) to dual boot (Windows or Deb/Proxmox) and just keep it booted into Deb as another proxmox node. That PC also has a 4070 Super that I have setup to passthru to a VM and on that VM I've got various services utilizing the GPU. This includes some that are utilized by my hetzner bare metal servers for things like image/text embeddings as well as local LLM use (though, rather minimal due to VRAM constraints) and some image/video object detection stuff with my security cameras (slowly working on a remote water gun turret to keep the racoons from trying to eat the kittens that stray cats keep having in my driveway/workshop). Install claude code (or, opencode, it's also good) - use Opus (get the max plan) and give it a directory that it can use as it's working directory (don't open it in ~/Documents and just start doing things) and prompt it with something as simple as this: "I have an existing home assistant setup at home and I'd like to determine what sort of self-hosted AI I could setup and integrate with that home assistant install - can you help me get started? Please also maintain some notes in .md files in this working directory with those note files named and organized as you see appropriate so that we can share relevant context and information with future sessions. (example: Hardware information, local urls, network layout, etc) If you're unsure of something, ask me questions. Do not perform any destructive actions without first confirming with me." Plan mode. _ALWAYS_ use plan mode to get the task setup, if there's something about the plan you don't like, say no and give it notes - it will return with a new plan. Eventually agree to the plan when it's right - then work through that plan not in plan mode, but if it gets off the plan, get back in plan mode to get the/a plan set and then again let it go and just steer it in regular mode. |
| |
| ▲ | servercobra 8 hours ago | parent | prev | next [-] | | Ooo interesting, I'd love to hear more about the esp32's as voice satellites! | | |
| ▲ | sho_hn 8 hours ago | parent [-] | | I assume it's very similar to what Home Assistant's backing commercial entity Nabu Casa sells with the "Home Assistant Voice PE" device, which is also esp32-based. The code is open and uses the esphome framework so it's fairly easy to recreate on custom HW you have laying around. |
| |
| ▲ | tomatoman 9 hours ago | parent | prev | next [-] | | Seems interesting setup, do you have it documented anywhere, thinking of building one! | |
| ▲ | state_less 8 hours ago | parent | prev | next [-] | | That's great to hear. I was mostly impressed with Qwen3 coder on my 4090, but am hobbled by the small memory footprint of the single card. What motherboard are you using with your 3090s? Like the others, I too am curious about those esp32s and what software you run on them. Keep up the good hacking - it's been fun to play with this stuff! | |
| ▲ | bityard 8 hours ago | parent | prev | next [-] | | Can you tell me about these voice satellites? | | |
| ▲ | nitroedge 3 hours ago | parent [-] | | He is referring the M5 Atom's I believe. I strongly recommend the ESP32 S3 box now, you can fire up Bobbas special firmware for it, search on Github, and its a blast with Home Assistant. |
| |
| ▲ | mudkipdev 4 hours ago | parent | prev | next [-] | | Do you also add custom tools to turn on/off the lights? | |
| ▲ | crorella 8 hours ago | parent | prev [-] | | omg, this is something I've had in mind for quite some time, I even bought some i2s devices to test it out.
Do you have some pointers on how to do it? |
|
|
| ▲ | nerdsniper 8 hours ago | parent | prev | next [-] |
| When has the average American ever been willing to spend a $1,000-2,000 premium for privacy-respecting tech? They already save $20-200 to buy IoT cameras which provide all audio and video from inside their home directly to the government without a warrant (Ring vs Reolink/etc). |
| |
| ▲ | state_less 7 hours ago | parent | next [-] | | To be fair, it isn't $1000-2000 extra, it's the new laptop/pc you just bought that is powerful enough (now, or in the near future) to run these open weight models. | |
| ▲ | echelon 7 hours ago | parent | prev [-] | | Ease of use is a major issue. What percentage of the people that you know are able to install python and dependencies plus the correct open weights models? I'd wager most of your parents can't do it. Most "normies" wouldn't even know what a local model even is, let alone how to install a GPU. | | |
| ▲ | nerdsniper 5 hours ago | parent | next [-] | | Wiredpancake got flagged to death but they’re right. MacWhisper provides a great example of good value for dead-simple user-friendly on-device processing. | |
| ▲ | wiredpancake 5 hours ago | parent | prev [-] | | That sounds like all software. We can make better software. |
|
|
|
| ▲ | Aurornis 5 hours ago | parent | prev | next [-] |
| > Americans may end up in a situation where they have some $1000-2000 device at home with an open Chinese model running on it, if they care about privacy or owning their data. I think HN vastly overestimates the market for something like this. Yes, there are some people who would spend $2,000 to avoid having prompts go to any cloud service. However, most people don’t care. Paying $20 per month for a ChatGPT subscription is a bargain and they automatically get access to new versions as they come. I think the at-home self hosting hobby is interesting, but it’s never going to be a mainstream thing. |
| |
| ▲ | reilly3000 3 hours ago | parent | next [-] | | There is going to be a big market for private AI appliances, in my estimation at least. Case in point: I give Gmail OAuth access to nobody. I nearly got burned once and I really don’t want my entire domain nuked. But I want to be able to have an LLM do things only LLMs can do with my email. “Find all emails with ‘autopay’ in the subject from my utility company for the past 12 months, then compare it to the prior year’s data.” GPT-OSS-20b tried its best but got the math obviously wrong. Qwen happily made the tool calls and spat out an accurate report, and even offered to make a CSV for me. Surely if you can’t trust npm packages or MS to not hand out god tokens to any who asks nicely, you shouldn’t trust a random MCP server with your credentials or your model. So I had Kilocode build my own. For that use case, local models just don’t quite cut it. I loaded $10 into OpenRouter, told it what I wanted, and selected GPT5 because it’s half off this week. 45 minutes, $0.78, and a few manual interventions later I had a working Gmail MCP that is my very own. It gave me some great instructions on how to configure an OAuth app in GCP, and I was able to get it running queries within minutes from my local models. There is a consumer play for a ~$2499-$5000 box that can run your personal staff of agents on the horizon. We need about one more generation of models and another generation of low-mid inference hardware to make it commercially feasible to turn a profit. It would need to pay for itself easily in the lives of its adopters. Then the mass market could open up. A more obvious path goes through SMBs who care about control and data sovereignty. If you’re curious, my power bill is up YoY, but there was a rate hike, definitely not my 4090;). | |
| ▲ | CJefferson an hour ago | parent | prev | next [-] | | The reason people will pay $2,000 for a private at home AI is porn. | |
| ▲ | malnourish 3 hours ago | parent | prev | next [-] | | The sales case for having LLMs at the edge is to run inference everywhere on everything. Video games won't go to the cloud for every AI call, but they will use on-device models that will run on the next iteration of hardware. | |
| ▲ | rafterydj 4 hours ago | parent | prev [-] | | I agree the market is niche atm, but I can't help but disagree with your outlook long term. Self hosted models don't have the problems ChatGPT subscribers are facing with models seemingly performing worse over time, they don't need to worry about usage quotas, they don't need to worry about getting locked out of their services, etc. All of these things have a dark side, though; but it's likely unnecessary for me to elaborate on that. |
|
|
| ▲ | moffkalast 7 hours ago | parent | prev | next [-] |
| There is some irony about buying Chinese hardware to run American software on it for the past decade(s), and now the exact reverse. |
| |
| ▲ | fragmede 3 hours ago | parent [-] | | Is that irony? Regardless, it's hilarious. All of the upvotes to you! |
|
|
| ▲ | 7 hours ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | an hour ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | 8 hours ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | bilbo0s 9 hours ago | parent | prev [-] |
| Americans may end up in a situation where they have some $1000-2000 device at home with an open Chinese model running on it Wouldn't worry about that, I'm pretty sure the government is going to ban running Chinese tech in this space sooner or later. And we won't even be able to download it. Not saying any of the bans will make any kind of sense, but I'm pretty sure they're gonna say this is a "strategic" space. And everything else will follow from there. Download Chinese models while you can. |
| |
| ▲ | Sanzig 9 hours ago | parent | next [-] | | When DeepSeek first hit the news, an American senator proposed adding it to ITAR so they could send people to prison for using it. Didn't pass, thankfully. | | |
| ▲ | thoroughburro 8 hours ago | parent [-] | | If it does in the future, do we just hope it won’t be retroactive? Is this water boiling yet? | | |
| ▲ | dingnuts 8 hours ago | parent [-] | | ex post facto law is explicitly banned in the US Constitution | | |
| ▲ | Light_Hope 7 hours ago | parent | next [-] | | For criminal concerns regarding retroactive ITAR additions, yes. However, significant civil financial penalties if congress so wished could still be constitutional as the ex post facto clause has been held to apply exclusively to criminal matters starting in Calder v. Bull [1]. [1] https://www.oyez.org/cases/1789-1850/3us386 | |
| ▲ | CamperBob2 7 hours ago | parent | prev | next [-] | | Dogs can't play basketball, either, but we've sure been getting dunked on a lot lately. | | |
| ▲ | jondwillis 4 hours ago | parent [-] | | History is littered with unconstitutional, enforced laws, as well. Watched a lot of Ken Burns docs this weekend while sick. “The West” has quite a few examples. |
| |
| ▲ | 4 hours ago | parent | prev [-] | | [deleted] |
|
|
| |
| ▲ | whimsicalism 9 hours ago | parent | prev | next [-] | | government hardly has the capacity to ban foreign weights | | |
| ▲ | quotemstr 5 hours ago | parent [-] | | The danger is that lawmakers, confused about the difference between foreign weights and foreign APIs, accidentally ban both. | | |
| ▲ | CamperBob2 2 hours ago | parent [-] | | There will be no confusion whatsoever, and no accidents. They don't write those bills themselves. Whatever a given bill does is precisely what its authors, who are almost never elected by any constituency on Earth, intend for it to do. |
|
| |
| ▲ | wkat4242 4 hours ago | parent | prev [-] | | Eh this is the internet. There's always a way. They couldn't ban piracy either. | | |
|