No technical report published yet, unlikely code or weights will be either given VC funding.
by Chance-Device
1 subcomments
It’s probably something like deepseek’s native sparse attention with content based granularity. They might not be publishing anything because it’s not such a strong value proposition and doing so would lead to commentary that would tank their investment opportunities.
by flowerthoughts
3 subcomments
For Claude Code, I feel 1M is enough. I've had a compaction once, but that was because I was forcing Claude to do something it clearly had a hard time understanding.
For general chat bots where the user doesn't understand what a context window is, what do you do about context? Latest few messages and then a memory tool? Compaction?
by roger_
0 subcomment
Have they published?
by Bombthecat
0 subcomment
I believe it, when I see it.
by Alifatisk
0 subcomment
Waiting for the paper and model card. I believe it when I see it.