安德烈·第一受召基金会代表团接收圣火14:40
├── do_not_blacklist/
On symbolic computation, arithmetic probe accuracy on a held-out pool of 1,000 math problems came in at 4% for NCCLIGen and 0% for base Wan2.1 — compared to 71% for Sora-2 and 2% for Veo3.1. Re-prompting alone, by providing the correct answer explicitly in the prompt at inference time, raised NCCLIGen accuracy from 4% to 83% without modifying the backbone or adding reinforcement learning. The research team interpreted this as evidence of steerability and faithful rendering of conditioned content, not native arithmetic computation inside the model.
СюжетРоссия и НАТО: