PH30vPBZLwZXlSFrALk6AT507Qn70LSPLW5a89vsRhDdkje xaPGvNE2UrhOBy8Gkaasn FVRuDWlPhEPntzw02gxSAEPygt7djS

DeepMind’s latest research at ICLR 2023

Research towards AI models that can generalise, scale, and accelerate science Next week marks the start of the 11th International Conference on Learning Representations (ICLR), taking place 1-5 May in Kigali, Rwanda. This will be the first major artificial intelligence (AI) conference to be hosted in Africa and the first in-person event since the start …

DeepMind’s latest research at ICLR 2023 Read More »

1 JAIj8bNlYsJ0OM8BnAiNzA scaled.webp

Unraveling Large Language Model Hallucinations

ContentsIntroductionLLM Training PipelinePretrainingPost-Training: Supervised Fine-TuningPost-training: Reinforcement Learning with Human FeedbackWhy Hallucinations?Model InterrogationUsing Web SearchConclusion Introduction In a YouTube video titled Deep Dive into LLMs like ChatGPT, former Senior Director of AI at Tesla, Andrej Karpathy discusses the psychology of Large Language Models (LLMs) as emergent cognitive effects of the training pipeline. This article is inspired by his explanation …

Unraveling Large Language Model Hallucinations Read More »

REkFCC8KEOAocMWBwcHOxKM6K2zRs qpMeUhnmHYkkGSbPPCLRhPDluhoZzx2k6 b4XvgZmhUqeuko9BXZZIPLmGR1q4BycDjLuD

An early warning system for novel AI risks

Responsibility & Safety Published 25 May 2023 Authors Toby Shevlane New research proposes a framework for evaluating general-purpose models against novel threats To pioneer responsibly at the cutting edge of artificial intelligence (AI) research, we must identify new capabilities and novel risks in our AI systems as early as possible. AI researchers already use a …

An early warning system for novel AI risks Read More »

unnamed 39

Vision Transformers (ViT) Explained: Are They Better Than CNNs?

Contents1. Introduction2. The Transformer2.1. The Self-Attention Mechanism2.2 The Multi-Headed Self-Attention3. The Vision Transformer4. The Result4.1 What does the ViT model learn?5. So, is ViT the future of Computer Vision?References 1. Introduction Ever since the introduction of the self-attention mechanism, Transformers have been the top choice when it comes to Natural Language Processing (NLP) tasks. Self-attention-based …

Vision Transformers (ViT) Explained: Are They Better Than CNNs? Read More »

kYAs9KTHdhYZE0BeKMKlphVqU3eQS8oXP GNrrWBjFbl8r4YFv2FWlRbe6x9L4Q L eKZeE7E GtKVJTLXvW zGTTzplSJCplN0

AlphaDev discovers faster sorting algorithms

Impact Published 7 June 2023 Authors Daniel J. Mankowitz and Andrea Michi New algorithms will transform the foundations of computing Digital society is driving increasing demand for computation, and energy use. For the last five decades, we relied on improvements in hardware to keep pace. But as microchips approach their physical limits, it’s critical to …

AlphaDev discovers faster sorting algorithms Read More »

1eP7D1tLIEpQ0b6YYugyvAA scaled

The Dangers of Deceptive Data–Confusing Charts and Misleading Headlines

“You don’t have to be an expert to deceive someone, though you might need some expertise to reliably recognize when you are being deceived.” When my co-instructor and I start our quarterly lesson on deceptive visualizations for the data visualization course we teach at the University of Washington, he emphasizes the point above to our …

The Dangers of Deceptive Data–Confusing Charts and Misleading Headlines Read More »

Flash Family meta.2e16d0ba.fill 1200x600

Start building with Gemini 2.0 Flash and Flash-Lite

Since the launch of the Gemini 2.0 Flash model family, developers are discovering new use cases for this highly efficient family of models. Gemini 2.0 Flash offers stronger performance over 1.5 Flash and 1.5 Pro, plus simplified pricing that makes our 1 million token context window more affordable. Today, Gemini 2.0 Flash-Lite is now generally …

Start building with Gemini 2.0 Flash and Flash-Lite Read More »

clint adair BW0vK FA3eg unsplash scaled

Enhancing RAG: Beyond Vanilla Approaches

Retrieval-Augmented Generation (RAG) is a powerful technique that enhances language models by incorporating external information retrieval mechanisms. While standard RAG implementations improve response relevance, they often struggle in complex retrieval scenarios. This article explores the limitations of a vanilla RAG setup and introduces advanced techniques to enhance its accuracy and efficiency. ContentsThe Challenge with Vanilla …

Enhancing RAG: Beyond Vanilla Approaches Read More »

6tSxHgEgSLR8FSELf3If1M1QBbXTtpsfH6w2ocuruWGnFDTdogbyNA8sHOyKpFYCja4hT7fGCVwl2xyI9biVB1bFNcnTxvYptuVd

MuZero, AlphaZero, and AlphaDev: Optimizing computer systems

As part of our aim to build increasingly capable and general artificial intelligence (AI) systems, we’re working to create AI tools with a broader understanding of the world. This can allow useful knowledge to be transferred between many different types of tasks. Using reinforcement learning, our AI systems AlphaZero and MuZero have achieved superhuman performance …

MuZero, AlphaZero, and AlphaDev: Optimizing computer systems Read More »

Scroll to Top