FRESH Hacker News
Home
Show HN: Duplicate 3 layers in a 24B LLM, logical deduction .22→.76. No training
248 points by xlayn