Was scrolling through the Artificial Analysis Arena img2vid model tester and saw 2 LTX2.3 vids there, one that knows anime as txt2vid and another that does multi-shot, but from my testing LTX2.3 doesn't know either. Is the open-source model nerfed or the site is straight up lying?
User flags a potential benchmark/demo mismatch between Artificial Analysis Arena examples and their own LTX 2.3 capability tests.
This thread is a model-evaluation signal: community members are questioning reproducibility and whether showcased LTX2.3 outputs reflect open-source checkpoints or internal tuning.