June 23rd, 2024

The Abstraction and Reasoning Corpus

The GitHub repository for ARC-AGI provides task data and a testing interface for solving tasks involving input/output pairs within 3 trials. Users can access the tasks and detailed instructions on the repository.

Read original articleLink Icon
The Abstraction and Reasoning Corpus

The GitHub repository for the Abstraction and Reasoning Corpus for Artificial General Intelligence (ARC-AGI) offers task data and a browser-based interface for manual task solving. The tasks, available in the `data` directory as JSON files, consist of input/output pairs to create the output grid from the test input grid within 3 trials. Users can engage with the tasks through the testing interface at `apps/testing_interface.html`, enabling grid resizing, input copying, symbol editing, answer validation, and more. For comprehensive instructions on utilizing the testing interface and tackling the tasks, users are directed to the ARC-AGI GitHub repository at https://github.com/fchollet/ARC-AGI.

Related

Optimizing AI Inference at Character.ai

Optimizing AI Inference at Character.ai

Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.

Solving puzzles faster than humanly possible

Solving puzzles faster than humanly possible

The Opus Magnum challenge tasks players with automating puzzle-solving to optimize Cost, Cycles, and Area metrics. Participants submit solutions for evaluation, exploring automated vs. human strategies, hybrid approaches, scoring systems, mods, and bots.

OpenAI and Anthropic are ignoring robots.txt

OpenAI and Anthropic are ignoring robots.txt

Two AI startups, OpenAI and Anthropic, are reported to be disregarding robots.txt rules, allowing them to scrape web content despite claiming to respect such regulations. TollBit analytics revealed this behavior, raising concerns about data misuse.

Show HN: Python lib to run evals across providers: OpenAI, Anthropic, etc.

Show HN: Python lib to run evals across providers: OpenAI, Anthropic, etc.

The GitHub repository provides details on LLM Safety Evals, accessible on evals.gg. It features a bar chart, a Twitter post, setup guidelines, and code execution commands. Contact for further support.

Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]

Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]

The video discusses limitations of large language models in AI, emphasizing genuine understanding and problem-solving skills. A prize incentivizes AI systems showcasing these abilities. Adaptability and knowledge acquisition are highlighted as crucial for true intelligence.

Link Icon 1 comments