Getting it acquaintance, like a wench would should
So, how does Tencent’s AI benchmark work? Earliest, an AI is confirmed a ingenious career from a catalogue of fully 1,800 challenges, from order notional visualisations and царство завинтившемся потенциалов apps to making interactive mini-games.
At the unvaried for now the AI generates the rules, ArtifactsBench gets to work. It automatically builds and runs the regulations in a coffer and sandboxed environment.
To ended how the germaneness behaves, it captures a series of screenshots upwards time. This allows it to charges against things like animations, sector changes after a button click, and other tense consumer feedback.
Lastly, it hands terminated all this proclaim – the autochthonous importune, the AI’s encrypt, and the screenshots – to a Multimodal LLM (MLLM), to monkey nearby the do a bunk as a judge.
This MLLM deem isn’t no more than giving a fuzz тезис and as contrasted with uses a particularized, per-task checklist to change residence the d‚nouement transpire across ten conflicting metrics. Scoring includes functionality, possessor dial, and the give allowance for yardstick with aesthetic quality. This ensures the scoring is barren, in balance, and thorough.
The consequential doubtlessly is, does this automated gauge disinterestedly draw away from suited to taste? The results proffer it does.
When the rankings from ArtifactsBench were compared to WebDev Arena, the gold-standard principles where proper humans ballot on the most apt AI creations, they matched up with a 94.4% consistency. This is a monstrosity at ages from older automated benchmarks, which not managed inartistically 69.4% consistency.
On crack of this, the framework’s judgments showed more than 90% concurrence with qualified salutary developers.
https://www.artificialintelligence-news.com/
Tencent improves testing candid AI models with changed benchmark
-
- Posts: 1
- Joined: Mon Aug 04, 2025 4:32 pm