Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What are people using to cost efficiently use this? I was using a Google Ultra sub which gave enough but that’s gone now.

ChatGPT $20/month is alright but I got locked out for a day after a couple hours. Considering the GitHub pro plus plan.



Run Qwen3-coder-next locally. That's what I'm doing (using LMstudio). It's actually a surprisingly capable model. I've had it working on some LLVM-IR manipulation and microcode generation for a kind of VLIW custom processor. I've been pleasantly surprised that it can handle this (LLVM is not easy) - there are also verilog code that define the processor's behavior that it reads to determine the microcode format and expected processor behavior. When I do hit something that it seems to struggle with I can go over to antigravity and get some free Gemini 3 flash usage.


What kind of hardware do you run it on?


Framework Desktop (AMD Strix Halo with 128GB). Runs it around 27 tok/sec which is quite acceptable.


Same here


Qwen3 Coder Next in llama.cpp on my own machine. I'm an AI hater, but I need to experiment with it occasionally, I'm not going to pay someone rent for something they trained on my own GitHub, Stack overflow, and Reddit posts.


FWIW the lockout probably wasn't related... maybe the content you were working on or your context window management somehow triggered something?


You could try minimax 2.5 via openrouter.


MiniMax has an incredibly affordable coding plan for $10/month. It has a rolling five hour limit of 100 prompts. 100 prompts doesn't sound like much, but in typical AI company accounting fashion, 1 prompt is not really 1 prompt. I have yet to come even close to hitting the limit with heavy use.


Kimi code with the .99 Cent plan is not to bad if you're savy




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: