Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
MoonGhost
8 months ago
|
parent
|
context
|
favorite
| on:
The Llama 4 herd
I think the problem is with positional encoding. If model cannot clearly separate tokens in context window they overlap which leads to mess. That encoding matters and actual position does not.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: