Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Stanford and Berkeley proposed LLM-as-a-Verifier, while also setting new records on the Terminal-Bench and SWE-Bench leaderboards.
ME News Report, April 14 (UTC+8), according to 1M AI News monitoring, when AI programming agents handle a single task multiple times, they often produce different solutions, some of which may be correct or incorrect. If the best solution can be automatically selected, the overall success rate can surpass that of a single run. The challenge is how to select the best: using another model as a judge to score (i.e., LLM-as-a-Judge) is the current mainstream approach, but the scoring granularity is too coarse, often giving different solutions the same score, making it hard to distinguish the better one. Stanford AI Laboratory and Berkeley Sky Computing Laboratory, in collaboration with NVIDIA, proposed LLM-as-a-Verifier to improve this selection process. Instead of only considering the final score given by the judge, it reads the probability distribution across each scoring level to calculate a continuous reward value. The judge is also asked to evaluate multiple times and average the results to eliminate random bias, and the overall assessment is broken down into three independent dimensions (whether the task requirements are met, whether the output format is correct, and whether there are error signals) for separate verification. In experiments, Gemini 2.5 Flash was used as the verifier, achieving a single verification accuracy of 74.7%, compared to only 57.0% for the traditional judge; after 16 repetitions, the Verifier reached 77.4%, while the Judge was at 70.2%. The traditional Judge had a 26.5% tie rate, whereas the Verifier had a 0% tie rate under all configurations. Practical results: on Terminal-Bench 2, running GPT-5.4 five times on the same task, the success rate of a randomly selected solution was 81.8%, which increased to 86.4% after using the Verifier for selection. On SWE-Bench Verified, selecting one solution each from Claude Opus 4.5, Claude Opus 4.6, and Gemini 3 Flash (total 3 solutions), the success rate increased from 76.1% to 77.8%. As of the release date on April 9, both results are top-ranked. The framework has been open-sourced. (Source: BlockBeats)