Powered by RND
PodcastsTechnologiesLarge Language Model (LLM) Talk
Écoutez Large Language Model (LLM) Talk dans l'application
Écoutez Large Language Model (LLM) Talk dans l'application
(48 139)(250 169)
Sauvegarde des favoris
Réveil
Minuteur

Large Language Model (LLM) Talk

Podcast Large Language Model (LLM) Talk
AI-Talk
AI Explained breaks down the world of AI in just 10 minutes. Get quick, clear insights into AI concepts and innovations, without any complicated math or jargon....

Épisodes disponibles

5 sur 36
  • OpenAI-o1
    OpenAI's o1 is a generative pre-trained transformer (GPT) model, designed for enhanced reasoning, especially in science and math. It uses a 'chain of thought' approach, spending more time "thinking" before answering, making it better at complex tasks. While not a successor to GPT-4o, o1 excels in scientific and mathematical benchmarks, and is trained with a new optimization algorithm. Different versions like o1-preview and o1-mini are available. Limitations include high computational cost, occasional "fake alignment," and a hidden reasoning process, and potential replication of training data.
    --------  
    10:33
  • GPT-4o
    GPT-4o is a multilingual, multimodal model that can process and generate text, images, and audio and represents a significant advancement over previous models like GPT-4 and GPT-3.5. GPT-4o is faster and more cost-effective, has improved performance in multiple areas, and natively supports voice-to-voice. GPT-4o's knowledge is limited to what was available up to October 2023. It has a context length of 128k tokens. The cost of training GPT-4 was more than $100 million, and it has 1 trillion parameters.
    --------  
    16:04
  • Kimi k1.5
    Kimi k1.5 is a multimodal LLM trained with reinforcement learning (RL). Key aspects include: long context scaling to 128k, improving performance with increased context length; improved policy optimization using a variant of online mirror descent; and a simplistic framework that enables planning and reflection without complex methods. It uses a reference policy in its off-policy RL approach, and long2short methods such as model merging and DPO to transfer knowledge from long-CoT to short-CoT models, achieving state-of-the-art reasoning performance. The model is jointly trained on text and vision data.
    --------  
    22:27
  • DeepSeek-R1
    DeepSeek-R1 is a language model focused on enhanced reasoning, employing reinforcement learning (RL) and building upon the DeepSeek-V3-Base model. It uses Group Relative Policy Optimization (GRPO) to reduce computational costs by eliminating the need for a separate critic model, which is commonly used in other algorithms such as PPO. The model uses a multi-stage training pipeline including an initial fine-tuning with cold-start data, followed by reasoning-oriented RL, and supervised fine-tuning (SFT) using rejection sampling, and a final RL stage. A rule-based reward system avoids reward hacking. DeepSeek-R1 also employs a language consistency reward during RL to address language mixing. The model's reasoning capabilities are then distilled into smaller models. DeepSeek-R1 achieves performance comparable to, and sometimes surpassing, OpenAI's o1 series on various reasoning, math, and coding tasks.
    --------  
    26:56
  • Claude-3
    Claude 3 is a family of large multimodal AI models developed by Anthropic, with a focus on safety, interpretability, and user alignment. The models, which include Opus, Sonnet, and Haiku, excel in reasoning, math, coding, and multilingual understanding. They are designed to be helpful, honest, and harmless assistants and can process text, audio, and visual inputs. Claude 3 models use Constitutional AI principles, aiming for more ethical and reliable responses. They have improved abilities in long context comprehension, and have shown strong performance in various tests, often outperforming previous Claude models and sometimes matching or exceeding GPT models in some benchmarks.
    --------  
    16:29

Plus de podcasts TechnologiesPlus de podcasts Technologies

À propos de Large Language Model (LLM) Talk

AI Explained breaks down the world of AI in just 10 minutes. Get quick, clear insights into AI concepts and innovations, without any complicated math or jargon. Perfect for your commute or spare time, this podcast makes understanding AI easy, engaging, and fun—whether you're a beginner or tech enthusiast.
Site web du podcast

Écoutez Large Language Model (LLM) Talk, Underscore_ ou d'autres podcasts du monde entier - avec l'app de radio.fr

Obtenez l’app radio.fr
 gratuite

  • Ajout de radios et podcasts en favoris
  • Diffusion via Wi-Fi ou Bluetooth
  • Carplay & Android Auto compatibles
  • Et encore plus de fonctionnalités
Applications
Réseaux sociaux
v7.5.1 | © 2007-2025 radio.de GmbH
Generated: 1/30/2025 - 6:33:00 PM