Zoom Claims Top Score on AI Benchmark Amid Skepticism; Zencoder Releases Free AI Coding Tool

Zoom announced a record score on a major AI test, drawing skepticism, while Zencoder launched a free tool comparing Claude and OpenAI models for coding.

Zoom’s Controversial AI Benchmark Achievement

According to VentureBeat, Zoom Video Communications announced last week that it achieved “the highest score ever recorded on one of artificial intelligence’s most demanding tests.” The claim generated “ripples of surprise, skepticism” from critics who suggest the company “copied off its neighbors,” though specific details about the test or the nature of the criticism were not provided in the source.

Zencoder Launches Free AI Orchestration Tool

VentureBeat reports that Silicon Valley startup Zencoder released a free desktop application on Monday called Zenflow. The tool is described as an “AI orchestration tool that pits Claude against OpenAI’s models to catch coding errors.” According to Zencoder, the application “will fundamentally change how software engineers interact with artificial intelligence,” aiming to move the industry “beyond the freewheeling era of ‘vibe coding’ toward a more” systematic approach.

New Benchmark for PhD-Level Mathematical Reasoning

A paper on arXiv (arXiv:2512.13978v1) introduces a new benchmark for evaluating frontier large language models on “PhD-Level Mathematical Reasoning” using a textbook on randomized algorithms in theoretical computer science. The abstract notes that “rapid advancement of large language models (LLMs) has led to significant breakthroughs in automated mathematical reasoning and scientific discovery,” citing work by Georgiev et al. demonstrating that “AI systems can explore” new territory in mathematical reasoning.