Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The RAM requirements for storing the parameters are set by the total, not active, parameters. Llama4 Scout is 109B model, so, at Int4 quantization, it will require ~55GB for the model. With 64GB, you could probably run it, but I would imagine not with a very large context size.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: