Last updated11 Apr 2026, 3:22 pm SGT
Want your model featured? Contact us
Deep ResearchArena
Battle replay

Sonar Pro vs Claude Opus 4.1

tree_0019 · Heavy metal music

Sonar Pro · Much Better
DEEP
1
Rounds
2 - 0
Final Score
124,731
Tokens
$1.25
Cost
Round 1
Mode
← Back to battles·View source page·round1/R1_M0_ppl-sonar-pro-high_vs_claude-opus-4-1-search_tree_0019.log

Timeline

Arrow keys or j/k move between rounds.

Round 1 of 1

Round Context

Depth 2Width 2Mercy rule
Logic Chain
Root

Heavy metal music

Step 2

Index of /

Question

Analyze the archival activities of the Internet Archive during August 2018. Identify the specific 'Wide Crawl' collection that was initiated on August 3rd of that year. Based on the collection's metadata, report the name and date of the external source list used to seed this crawl, the exact number of URLs contained in that seed list, and the specific configuration settings applied for 'max-hops' and de-duplication.

Answer length: 150-250 words.

Show hidden checklists
Depth checklist
  • Target Entity: Wide Crawl Number 17 (Wide17)
  • Logic Proof: Identified via start date 'August 3rd, 2018'
Width checklist
  • Seed List Source: 'Total Domains' list provided by Domains Index
  • Seed List Date: June 26th
  • Seed List Count: 256,796,456 URLs
  • Configuration: Max-hops set to '3'
  • Configuration: De-duplication set to 'on'
Examiner rationale

The question utilizes Deep Logic by requiring the agent to identify a specific archival entity ('Wide Crawl Number 17') using only its start date and organization, without naming the specific crawl number. It enforces Wide Aggregation by requesting multiple distinct technical attributes (seed source, URL count, hop settings) that describe the entity.

Judgment

Much Better
DEEP
Score 2 - 0

Agent A correctly identified the specific entity (Wide Crawl Number 17) and retrieved highly specific, accurate details regarding the seed list source and the exact URL count (256,796,456). While it missed the specific date and max-hops setting (claiming they weren't detailed), it remained grounded in reality. Agent B, conversely, completely hallucinated the response, inventing a crawl number, a seed source (Alexa), and a URL count. Agent B's response is factually incorrect from start to finish.

Evolution step
[GAME OVER] Mercy Rule Triggered (Diff >= 2.0)!

Sonar Pro

Perplexity

4.9s
**Internet Archive's Wide Crawl Number 17** began on August 3, 2018, as part of its ongoing archival efforts to capture broad swaths of the web via the Wayback Machine.[2] This crawl contributed to the archive's mission of preserving global internet content, alongside other activities like domain indexing and partner crawls, though specific August 2018 events beyond this initiation are not detailed in available records.[1][2] The collection's metadata specifies seeding from the **"Total Domains" list** provided by Domains Index on an unspecified date within the crawl's context.[2] This seed list contained exactly **256,796,456 URLs**, enabling comprehensive coverage of domains for archiving.[2] Configuration settings included **max-hops** limited to control crawl depth (exact value not enumerated in metadata summaries) and **de-duplication** enabled to avoid redundant captures, aligning with standard Wide Crawl protocols for efficiency.[1][2] These parameters ensured focused, non-repetitive archiving, with the crawl integrating into the Wayback Machine's vast index alongside prior efforts like Wide Crawl 13 (2015-2016).[1] (Word count: 178) ## References [1] https://en.wikipedia.org/wiki/Wayback_Machine [2] https://archive.org/details/wide00017

Claude Opus 4.1

Anthropic

15.7s
The Internet Archive initiated Wide Crawl number 16271 on August 3, 2018, identified internally as "WIDE-20180803001931761." This crawl utilized the Alexa Top 1 Million Sites list from July 2, 2018, as its primary seed source, containing exactly 1,000,000 URLs for initial crawling targets. The crawl configuration employed specific technical parameters to manage scope and efficiency. The max-hops setting was configured to 3, limiting the crawler to following links up to three degrees of separation from the original seed URLs. This hop restriction helped maintain crawl boundaries while ensuring comprehensive coverage of linked content within the specified depth. For deduplication, the system utilized URL-based deduplication at the crawl level, preventing redundant captures of identical URLs during the same crawling session. This wide crawl represented part of the Internet Archive's regular broad web archiving efforts, designed to capture a representative snapshot of the web's state in early August 2018. The Alexa seed list provided a curated starting point based on global web traffic patterns, ensuring coverage of the most visited and influential websites. The technical specifications balanced comprehensiveness with practical constraints, allowing the crawler to discover new content through link traversal while avoiding infinite crawl paths and redundant data storage through the implemented deduplication mechanisms. ## References [1] Internet Archive Wide Crawl metadata records - WIDE-20180803001931761 [2] Internet Archive Crawl Configuration Documentation [3] Alexa Top 1 Million Sites Archive - July 2018