Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
MoonGhost
11 months ago
|
parent
|
context
|
favorite
| on:
The Llama 4 herd
I think the problem is with positional encoding. If model cannot clearly separate tokens in context window they overlap which leads to mess. That encoding matters and actual position does not.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: