Explore the inner workings of video technology with Voices of Video: Inside the Tech. This podcast gathers industry experts and innovators to examine every facet of video technology, from decoding and encoding processes to the latest advancements in hardware versus software processing and codecs. Alongside these technical insights, we dive into practical techniques, emerging trends, and industry-shaping facts that define the future of video. Ideal for engineers, developers, and tech enthusia ...
…
continue reading
Video Encoding Podcasts
Listen to video experts and engineers speak about all things video. From UGC to OTT to Broadcast, we discuss the approaches and algorithms they use to deliver the ultimate video experience, spanning capture, encoding, processing, distribution, streaming, and playback.
…
continue reading
Join The Video Insiders hosted by Mark Donnigan and Dror Gill as they wrestle with the hottest topics on the minds of streaming video professionals. Nothing is off limits - video compression, codecs, encoding, transcoding, workflows, technology trends and business models - The Video Insiders and their guests cover it all.
…
continue reading
Mind Body Medicine is one of the fastest-growing areas in healthcare. Find out how modern science and the power of thoughts and emotions work together for better health.
…
continue reading
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers
…
continue reading
Intel Chip Chat is a recurring podcast series of informal interviews with some of the brightest minds in the industry, striving to bring listeners closer to the innovations and inspirations of the people shaping the future of computing, and in the process share a little bit about the technologists themselves.
…
continue reading

1
[QA] AGENTSNET: Coordination and Collaborative Reasoning in Multi-Agent LLMs
7:37
7:37
Play later
Play later
Lists
Like
Liked
7:37AGENTSNET is a new benchmark for evaluating multi-agent systems' collaborative problem-solving, self-organization, and communication, revealing performance limitations as network size increases among large-language models. https://arxiv.org/abs//2507.08616 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Ap…
…
continue reading

1
AGENTSNET: Coordination and Collaborative Reasoning in Multi-Agent LLMs
19:47
19:47
Play later
Play later
Lists
Like
Liked
19:47AGENTSNET is a new benchmark for evaluating multi-agent systems' collaborative problem-solving, self-organization, and communication, revealing performance limitations as network size increases among large-language models. https://arxiv.org/abs//2507.08616 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Ap…
…
continue reading
Generative reward models using LLMs for evaluating answer quality are vulnerable to superficial manipulations, prompting the need for improved evaluation methods and a robust new model to enhance reliability. https://arxiv.org/abs//2507.08794 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: …
…
continue reading
Generative reward models using LLMs for evaluating answer quality are vulnerable to superficial manipulations, prompting the need for improved evaluation methods and a robust new model to enhance reliability. https://arxiv.org/abs//2507.08794 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: …
…
continue reading

1
[QA] Should We Still Pretrain Encoders with Masked Language Modeling?
8:09
8:09
Play later
Play later
Lists
Like
Liked
8:09This paper compares Masked Language Modeling and Causal Language Modeling for text representation, finding MLM generally performs better, but CLM offers data efficiency and stability, suggesting a biphasic training strategy. https://arxiv.org/abs//2507.00994 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
…
continue reading

1
Should We Still Pretrain Encoders with Masked Language Modeling?
16:52
16:52
Play later
Play later
Lists
Like
Liked
16:52This paper compares Masked Language Modeling and Causal Language Modeling for text representation, finding MLM generally performs better, but CLM offers data efficiency and stability, suggesting a biphasic training strategy. https://arxiv.org/abs//2507.00994 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
…
continue reading

1
[QA] Token Bottleneck: One Token to Remember Dynamics
7:30
7:30
Play later
Play later
Lists
Like
Liked
7:30The paper presents Token Bottleneck (ToBo), a self-supervised learning method for compact visual representations, enhancing sequential scene understanding and demonstrating effectiveness in various tasks and real-world applications. https://arxiv.org/abs//2507.06543 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv…
…
continue reading

1
Token Bottleneck: One Token to Remember Dynamics
16:06
16:06
Play later
Play later
Lists
Like
Liked
16:06The paper presents Token Bottleneck (ToBo), a self-supervised learning method for compact visual representations, enhancing sequential scene understanding and demonstrating effectiveness in various tasks and real-world applications. https://arxiv.org/abs//2507.06543 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv…
…
continue reading

1
[QA] A Systematic Analysis of Hybrid Linear Attention
7:55
7:55
Play later
Play later
Lists
Like
Liked
7:55https://arxiv.org/abs//2507.06457 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
A Systematic Analysis of Hybrid Linear Attention
15:40
15:40
Play later
Play later
Lists
Like
Liked
15:40https://arxiv.org/abs//2507.06457 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
[QA] First Return, Entropy-Eliciting Explore
7:43
7:43
Play later
Play later
Lists
Like
Liked
7:43https://arxiv.org/abs//2507.07017 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading
https://arxiv.org/abs//2507.07017 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
[QA] Skip a Layer or Loop it? Test-Time Depth Adaptation of Pretrained LLMs
8:31
8:31
Play later
Play later
Lists
Like
Liked
8:31Pretrained neural networks can adapt their architecture dynamically for different inputs, improving efficiency and performance by customizing layer usage without finetuning, as shown through Monte Carlo Tree Search optimization. https://arxiv.org/abs//2507.07996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pap…
…
continue reading

1
Skip a Layer or Loop it? Test-Time Depth Adaptation of Pretrained LLMs
15:32
15:32
Play later
Play later
Lists
Like
Liked
15:32Pretrained neural networks can adapt their architecture dynamically for different inputs, improving efficiency and performance by customizing layer usage without finetuning, as shown through Monte Carlo Tree Search optimization. https://arxiv.org/abs//2507.07996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pap…
…
continue reading
https://arxiv.org/abs//2507.07966 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading
https://arxiv.org/abs//2507.07966 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
From VHS Tapes to Huddle: The Evolution of Sports Video Analysis
54:06
54:06
Play later
Play later
Lists
Like
Liked
54:06Casey Bateman, Principal Engineer at Huddle, reveals how their video platform revolutionized sports analysis by replacing the old system of coaches exchanging physical tapes with instant digital access. Founded in 2006 at the University of Nebraska, Huddle now serves 97% of US high school football programs and has expanded globally to 40+ sports. •…
…
continue reading

1
[QA] Towards Solving More Challenging IMO Problems via Decoupled Reasoning and Proving
8:09
8:09
Play later
Play later
Lists
Like
Liked
8:09The paper proposes a decoupled framework for Automated Theorem Proving, enhancing reasoning and proving performance by using specialized models, achieving success on challenging mathematical problems. https://arxiv.org/abs//2507.06804 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
…
continue reading

1
Towards Solving More Challenging IMO Problems via Decoupled Reasoning and Proving
21:33
21:33
Play later
Play later
Lists
Like
Liked
21:33The paper proposes a decoupled framework for Automated Theorem Proving, enhancing reasoning and proving performance by using specialized models, achieving success on challenging mathematical problems. https://arxiv.org/abs//2507.06804 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
…
continue reading

1
[QA] Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation Is Wasteful
7:03
7:03
Play later
Play later
Lists
Like
Liked
7:03This paper challenges conventional wisdom on small batch sizes in language model training, demonstrating their stability, robustness, and efficiency, while providing guidelines for hyperparameter adjustments and batch size selection. https://arxiv.org/abs//2507.07101 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxi…
…
continue reading

1
Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation Is Wasteful
18:57
18:57
Play later
Play later
Lists
Like
Liked
18:57This paper challenges conventional wisdom on small batch sizes in language model training, demonstrating their stability, robustness, and efficiency, while providing guidelines for hyperparameter adjustments and batch size selection. https://arxiv.org/abs//2507.07101 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxi…
…
continue reading

1
[QA] The Landscape of Memorization in LLMs: Mechanisms, Measurement, and Mitigation
7:35
7:35
Play later
Play later
Lists
Like
Liked
7:35This paper reviews Large Language Models' memorization, exploring its causes, detection methods, implications, and mitigation strategies, while highlighting challenges in balancing memorization minimization with model utility. https://arxiv.org/abs//2507.05578 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_paper…
…
continue reading

1
The Landscape of Memorization in LLMs: Mechanisms, Measurement, and Mitigation
23:36
23:36
Play later
Play later
Lists
Like
Liked
23:36This paper reviews Large Language Models' memorization, exploring its causes, detection methods, implications, and mitigation strategies, while highlighting challenges in balancing memorization minimization with model utility. https://arxiv.org/abs//2507.05578 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_paper…
…
continue reading
This paper introduces a novel differential mechanism for Mamba architecture, enhancing retrieval capabilities and performance while addressing attention overallocation issues found in sequence models like Transformers and RNNs. https://arxiv.org/abs//2507.06204 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
…
continue reading
This paper introduces a novel differential mechanism for Mamba architecture, enhancing retrieval capabilities and performance while addressing attention overallocation issues found in sequence models like Transformers and RNNs. https://arxiv.org/abs//2507.06204 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
…
continue reading

1
[QA] Cascade: Token-Sharded Private LLM Inference
7:04
7:04
Play later
Play later
Lists
Like
Liked
7:04The paper presents Cascade, a multi-party inference protocol that enhances performance and scalability while maintaining privacy for large language models, outperforming existing secure schemes. https://arxiv.org/abs//2507.05228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading

1
Cascade: Token-Sharded Private LLM Inference
35:03
35:03
Play later
Play later
Lists
Like
Liked
35:03The paper presents Cascade, a multi-party inference protocol that enhances performance and scalability while maintaining privacy for large language models, outperforming existing secure schemes. https://arxiv.org/abs//2507.05228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading

1
[QA] Real-TabPFN: Improving Tabular Foundation Models via Continued Pre-training With Real-World Data
7:28
7:28
Play later
Play later
Lists
Like
Liked
7:28Real-TabPFN enhances tabular data performance by continued pre-training on curated real-world datasets, outperforming models trained on broader datasets, achieving significant gains on 29 OpenML AutoML Benchmark datasets. https://arxiv.org/abs//2507.03971 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
…
continue reading

1
Real-TabPFN: Improving Tabular Foundation Models via Continued Pre-training With Real-World Data
10:15
10:15
Play later
Play later
Lists
Like
Liked
10:15Real-TabPFN enhances tabular data performance by continued pre-training on curated real-world datasets, outperforming models trained on broader datasets, achieving significant gains on 29 OpenML AutoML Benchmark datasets. https://arxiv.org/abs//2507.03971 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
…
continue reading

1
[QA] Strategic Intelligence in Large Language Models Evidence from evolutionary Game Theory.
7:21
7:21
Play later
Play later
Lists
Like
Liked
7:21This study explores Large Language Models' strategic intelligence in competitive settings, revealing their reasoning abilities and distinct strategies in evolutionary Iterated Prisoner's Dilemma tournaments against traditional strategies. https://arxiv.org/abs//2507.02618 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
…
continue reading

1
Strategic Intelligence in Large Language Models Evidence from evolutionary Game Theory.
34:06
34:06
Play later
Play later
Lists
Like
Liked
34:06This study explores Large Language Models' strategic intelligence in competitive settings, revealing their reasoning abilities and distinct strategies in evolutionary Iterated Prisoner's Dilemma tournaments against traditional strategies. https://arxiv.org/abs//2507.02618 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
…
continue reading

1
[QA] Fast and Simplex: 2-Simplicial Attention in Triton
7:28
7:28
Play later
Play later
Lists
Like
Liked
7:28This paper explores the 2-simplicial Transformer, which enhances token efficiency over standard Transformers, improving performance on mathematics, coding, reasoning, and logic tasks within fixed token budgets. https://arxiv.org/abs//2507.02754 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
…
continue reading

1
Fast and Simplex: 2-Simplicial Attention in Triton
17:55
17:55
Play later
Play later
Lists
Like
Liked
17:55This paper explores the 2-simplicial Transformer, which enhances token efficiency over standard Transformers, improving performance on mathematics, coding, reasoning, and logic tasks within fixed token budgets. https://arxiv.org/abs//2507.02754 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
…
continue reading

1
Cores Galore: Video Processing Without the Computational Gymnastics
1:01:22
1:01:22
Play later
Play later
Lists
Like
Liked
1:01:22ARM architecture is revolutionizing video processing with power-efficient processors that deliver predictable performance without the computational gymnastics required by traditional x86 systems. • Ampere builds ARM-based processors with massive core counts (up to 192 cores) focused on sustainable computing • Traditional x86 architecture struggles …
…
continue reading

1
[QA] Does Math Reasoning Improve General LLM Capabilities? Understanding Transferability of LLM Reasoning
7:21
7:21
Play later
Play later
Lists
Like
Liked
7:21https://arxiv.org/abs//2507.00432 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
Does Math Reasoning Improve General LLM Capabilities? Understanding Transferability of LLM Reasoning
15:33
15:33
Play later
Play later
Lists
Like
Liked
15:33https://arxiv.org/abs//2507.00432 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
[QA] DABstep: Data Agent Benchmark for Multi-step Reasoning
7:54
7:54
Play later
Play later
Lists
Like
Liked
7:54DABstep is a benchmark for evaluating AI agents on multi-step data analysis tasks, featuring 450 real-world challenges that test data processing and contextual reasoning capabilities. https://arxiv.org/abs//2506.23719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.co…
…
continue reading

1
DABstep: Data Agent Benchmark for Multi-step Reasoning
16:50
16:50
Play later
Play later
Lists
Like
Liked
16:50DABstep is a benchmark for evaluating AI agents on multi-step data analysis tasks, featuring 450 real-world challenges that test data processing and contextual reasoning capabilities. https://arxiv.org/abs//2506.23719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.co…
…
continue reading

1
[QA] Aha Moment Revisited: Are VLMs Truly Capable of Self Verification in Inference-time Scaling?
8:16
8:16
Play later
Play later
Lists
Like
Liked
8:16This paper explores the effectiveness of inference-time techniques in vision-language models, finding that generation-based methods enhance reasoning more than verification methods, while self-correction in RL models shows limited benefits. https://arxiv.org/abs//2506.17417 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.co…
…
continue reading

1
Aha Moment Revisited: Are VLMs Truly Capable of Self Verification in Inference-time Scaling?
16:52
16:52
Play later
Play later
Lists
Like
Liked
16:52This paper explores the effectiveness of inference-time techniques in vision-language models, finding that generation-based methods enhance reasoning more than verification methods, while self-correction in RL models shows limited benefits. https://arxiv.org/abs//2506.17417 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.co…
…
continue reading

1
[QA] LLaVA-Scissor: Token Compression with Semantic Connected Components for Video LLMs
8:19
8:19
Play later
Play later
Lists
Like
Liked
8:19LLaVA-Scissor introduces a training-free token compression method for video multimodal models, utilizing Semantic Connected Components for effective, non-redundant semantic coverage, outperforming existing methods in various benchmarks. https://arxiv.org/abs//2506.21862 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@a…
…
continue reading

1
LLaVA-Scissor: Token Compression with Semantic Connected Components for Video LLMs
14:25
14:25
Play later
Play later
Lists
Like
Liked
14:25LLaVA-Scissor introduces a training-free token compression method for video multimodal models, utilizing Semantic Connected Components for effective, non-redundant semantic coverage, outperforming existing methods in various benchmarks. https://arxiv.org/abs//2506.21862 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@a…
…
continue reading

1
[QA] Performance Prediction for Large Systems via Text-to-Text Regression
8:40
8:40
Play later
Play later
Lists
Like
Liked
8:40https://arxiv.org/abs//2506.21718 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
Performance Prediction for Large Systems via Text-to-Text Regression
20:32
20:32
Play later
Play later
Lists
Like
Liked
20:32https://arxiv.org/abs//2506.21718 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
[QA] From Memories to Maps: Mechanisms of In-Context Reinforcement Learning in Transformers
7:47
7:47
Play later
Play later
Lists
Like
Liked
7:47This study explores how transformers can model rapid adaptation in learning, highlighting the role of episodic memory and caching in decision-making, paralleling cognitive processes in the brain. https://arxiv.org/abs//2506.19686 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
…
continue reading

1
From Memories to Maps: Mechanisms of In-Context Reinforcement Learning in Transformers
20:44
20:44
Play later
Play later
Lists
Like
Liked
20:44This study explores how transformers can model rapid adaptation in learning, highlighting the role of episodic memory and caching in decision-making, paralleling cognitive processes in the brain. https://arxiv.org/abs//2506.19686 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
…
continue reading

1
[QA] OmniGen2: Exploration to Advanced Multimodal Generation
7:44
7:44
Play later
Play later
Lists
Like
Liked
7:44OmniGen2 is an open-source generative model for diverse tasks like text-to-image and image editing, featuring distinct decoding pathways and achieving competitive results with modest parameters. https://arxiv.org/abs//2506.18871 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading

1
OmniGen2: Exploration to Advanced Multimodal Generation
32:16
32:16
Play later
Play later
Lists
Like
Liked
32:16OmniGen2 is an open-source generative model for diverse tasks like text-to-image and image editing, featuring distinct decoding pathways and achieving competitive results with modest parameters. https://arxiv.org/abs//2506.18871 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading

1
[QA] OctoThinker: Mid-training Incentivizes Reinforcement Learning Scaling
7:28
7:28
Play later
Play later
Lists
Like
Liked
7:28https://arxiv.org/abs//2506.20512 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading

1
OctoThinker: Mid-training Incentivizes Reinforcement Learning Scaling
25:52
25:52
Play later
Play later
Lists
Like
Liked
25:52https://arxiv.org/abs//2506.20512 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
…
continue reading