Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo

Video Encoding Podcasts

show episodes
 
Artwork

1
Voices of Video

NETINT Technologies

icon
Unsubscribe
icon
icon
Unsubscribe
icon
Weekly
 
Explore the inner workings of video technology with Voices of Video: Inside the Tech. This podcast gathers industry experts and innovators to examine every facet of video technology, from decoding and encoding processes to the latest advancements in hardware versus software processing and codecs. Alongside these technical insights, we dive into practical techniques, emerging trends, and industry-shaping facts that define the future of video. Ideal for engineers, developers, and tech enthusia ...
  continue reading
 
Artwork
 
Listen to video experts and engineers speak about all things video. From UGC to OTT to Broadcast, we discuss the approaches and algorithms they use to deliver the ultimate video experience, spanning capture, encoding, processing, distribution, streaming, and playback.
  continue reading
 
Artwork

1
The Video Insiders

The Video Insiders

icon
Unsubscribe
icon
icon
Unsubscribe
icon
Monthly
 
Join The Video Insiders hosted by Mark Donnigan and Dror Gill as they wrestle with the hottest topics on the minds of streaming video professionals. Nothing is off limits - video compression, codecs, encoding, transcoding, workflows, technology trends and business models - The Video Insiders and their guests cover it all.
  continue reading
 
Artwork

1
Arxiv Papers

Igor Melnyk

icon
Unsubscribe
icon
icon
Unsubscribe
icon
Daily+
 
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers
  continue reading
 
Artwork

1
Intel Chip Chat

Intel Corporation

icon
Unsubscribe
icon
icon
Unsubscribe
icon
Monthly
 
Intel Chip Chat is a recurring podcast series of informal interviews with some of the brightest minds in the industry, striving to bring listeners closer to the innovations and inspirations of the people shaping the future of computing, and in the process share a little bit about the technologists themselves.
  continue reading
 
Loading …
show series
 
AGENTSNET is a new benchmark for evaluating multi-agent systems' collaborative problem-solving, self-organization, and communication, revealing performance limitations as network size increases among large-language models. https://arxiv.org/abs//2507.08616 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Ap…
  continue reading
 
AGENTSNET is a new benchmark for evaluating multi-agent systems' collaborative problem-solving, self-organization, and communication, revealing performance limitations as network size increases among large-language models. https://arxiv.org/abs//2507.08616 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Ap…
  continue reading
 
Generative reward models using LLMs for evaluating answer quality are vulnerable to superficial manipulations, prompting the need for improved evaluation methods and a robust new model to enhance reliability. https://arxiv.org/abs//2507.08794 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: …
  continue reading
 
Generative reward models using LLMs for evaluating answer quality are vulnerable to superficial manipulations, prompting the need for improved evaluation methods and a robust new model to enhance reliability. https://arxiv.org/abs//2507.08794 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: …
  continue reading
 
This paper compares Masked Language Modeling and Causal Language Modeling for text representation, finding MLM generally performs better, but CLM offers data efficiency and stability, suggesting a biphasic training strategy. https://arxiv.org/abs//2507.00994 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
  continue reading
 
This paper compares Masked Language Modeling and Causal Language Modeling for text representation, finding MLM generally performs better, but CLM offers data efficiency and stability, suggesting a biphasic training strategy. https://arxiv.org/abs//2507.00994 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
  continue reading
 
The paper presents Token Bottleneck (ToBo), a self-supervised learning method for compact visual representations, enhancing sequential scene understanding and demonstrating effectiveness in various tasks and real-world applications. https://arxiv.org/abs//2507.06543 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv…
  continue reading
 
The paper presents Token Bottleneck (ToBo), a self-supervised learning method for compact visual representations, enhancing sequential scene understanding and demonstrating effectiveness in various tasks and real-world applications. https://arxiv.org/abs//2507.06543 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv…
  continue reading
 
Pretrained neural networks can adapt their architecture dynamically for different inputs, improving efficiency and performance by customizing layer usage without finetuning, as shown through Monte Carlo Tree Search optimization. https://arxiv.org/abs//2507.07996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pap…
  continue reading
 
Pretrained neural networks can adapt their architecture dynamically for different inputs, improving efficiency and performance by customizing layer usage without finetuning, as shown through Monte Carlo Tree Search optimization. https://arxiv.org/abs//2507.07996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pap…
  continue reading
 
Casey Bateman, Principal Engineer at Huddle, reveals how their video platform revolutionized sports analysis by replacing the old system of coaches exchanging physical tapes with instant digital access. Founded in 2006 at the University of Nebraska, Huddle now serves 97% of US high school football programs and has expanded globally to 40+ sports. •…
  continue reading
 
The paper proposes a decoupled framework for Automated Theorem Proving, enhancing reasoning and proving performance by using specialized models, achieving success on challenging mathematical problems. https://arxiv.org/abs//2507.06804 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
The paper proposes a decoupled framework for Automated Theorem Proving, enhancing reasoning and proving performance by using specialized models, achieving success on challenging mathematical problems. https://arxiv.org/abs//2507.06804 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
This paper challenges conventional wisdom on small batch sizes in language model training, demonstrating their stability, robustness, and efficiency, while providing guidelines for hyperparameter adjustments and batch size selection. https://arxiv.org/abs//2507.07101 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxi…
  continue reading
 
This paper challenges conventional wisdom on small batch sizes in language model training, demonstrating their stability, robustness, and efficiency, while providing guidelines for hyperparameter adjustments and batch size selection. https://arxiv.org/abs//2507.07101 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxi…
  continue reading
 
This paper reviews Large Language Models' memorization, exploring its causes, detection methods, implications, and mitigation strategies, while highlighting challenges in balancing memorization minimization with model utility. https://arxiv.org/abs//2507.05578 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_paper…
  continue reading
 
This paper reviews Large Language Models' memorization, exploring its causes, detection methods, implications, and mitigation strategies, while highlighting challenges in balancing memorization minimization with model utility. https://arxiv.org/abs//2507.05578 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_paper…
  continue reading
 
This paper introduces a novel differential mechanism for Mamba architecture, enhancing retrieval capabilities and performance while addressing attention overallocation issues found in sequence models like Transformers and RNNs. https://arxiv.org/abs//2507.06204 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
This paper introduces a novel differential mechanism for Mamba architecture, enhancing retrieval capabilities and performance while addressing attention overallocation issues found in sequence models like Transformers and RNNs. https://arxiv.org/abs//2507.06204 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
The paper presents Cascade, a multi-party inference protocol that enhances performance and scalability while maintaining privacy for large language models, outperforming existing secure schemes. https://arxiv.org/abs//2507.05228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
The paper presents Cascade, a multi-party inference protocol that enhances performance and scalability while maintaining privacy for large language models, outperforming existing secure schemes. https://arxiv.org/abs//2507.05228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
Real-TabPFN enhances tabular data performance by continued pre-training on curated real-world datasets, outperforming models trained on broader datasets, achieving significant gains on 29 OpenML AutoML Benchmark datasets. https://arxiv.org/abs//2507.03971 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
  continue reading
 
Real-TabPFN enhances tabular data performance by continued pre-training on curated real-world datasets, outperforming models trained on broader datasets, achieving significant gains on 29 OpenML AutoML Benchmark datasets. https://arxiv.org/abs//2507.03971 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
  continue reading
 
This study explores Large Language Models' strategic intelligence in competitive settings, revealing their reasoning abilities and distinct strategies in evolutionary Iterated Prisoner's Dilemma tournaments against traditional strategies. https://arxiv.org/abs//2507.02618 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
  continue reading
 
This study explores Large Language Models' strategic intelligence in competitive settings, revealing their reasoning abilities and distinct strategies in evolutionary Iterated Prisoner's Dilemma tournaments against traditional strategies. https://arxiv.org/abs//2507.02618 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
  continue reading
 
This paper explores the 2-simplicial Transformer, which enhances token efficiency over standard Transformers, improving performance on mathematics, coding, reasoning, and logic tasks within fixed token budgets. https://arxiv.org/abs//2507.02754 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
  continue reading
 
This paper explores the 2-simplicial Transformer, which enhances token efficiency over standard Transformers, improving performance on mathematics, coding, reasoning, and logic tasks within fixed token budgets. https://arxiv.org/abs//2507.02754 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
  continue reading
 
ARM architecture is revolutionizing video processing with power-efficient processors that deliver predictable performance without the computational gymnastics required by traditional x86 systems. • Ampere builds ARM-based processors with massive core counts (up to 192 cores) focused on sustainable computing • Traditional x86 architecture struggles …
  continue reading
 
https://arxiv.org/abs//2507.00432 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
  continue reading
 
https://arxiv.org/abs//2507.00432 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
  continue reading
 
DABstep is a benchmark for evaluating AI agents on multi-step data analysis tasks, featuring 450 real-world challenges that test data processing and contextual reasoning capabilities. https://arxiv.org/abs//2506.23719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.co…
  continue reading
 
DABstep is a benchmark for evaluating AI agents on multi-step data analysis tasks, featuring 450 real-world challenges that test data processing and contextual reasoning capabilities. https://arxiv.org/abs//2506.23719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.co…
  continue reading
 
This paper explores the effectiveness of inference-time techniques in vision-language models, finding that generation-based methods enhance reasoning more than verification methods, while self-correction in RL models shows limited benefits. https://arxiv.org/abs//2506.17417 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.co…
  continue reading
 
This paper explores the effectiveness of inference-time techniques in vision-language models, finding that generation-based methods enhance reasoning more than verification methods, while self-correction in RL models shows limited benefits. https://arxiv.org/abs//2506.17417 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.co…
  continue reading
 
LLaVA-Scissor introduces a training-free token compression method for video multimodal models, utilizing Semantic Connected Components for effective, non-redundant semantic coverage, outperforming existing methods in various benchmarks. https://arxiv.org/abs//2506.21862 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@a…
  continue reading
 
LLaVA-Scissor introduces a training-free token compression method for video multimodal models, utilizing Semantic Connected Components for effective, non-redundant semantic coverage, outperforming existing methods in various benchmarks. https://arxiv.org/abs//2506.21862 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@a…
  continue reading
 
This study explores how transformers can model rapid adaptation in learning, highlighting the role of episodic memory and caching in decision-making, paralleling cognitive processes in the brain. https://arxiv.org/abs//2506.19686 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
  continue reading
 
This study explores how transformers can model rapid adaptation in learning, highlighting the role of episodic memory and caching in decision-making, paralleling cognitive processes in the brain. https://arxiv.org/abs//2506.19686 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
  continue reading
 
OmniGen2 is an open-source generative model for diverse tasks like text-to-image and image editing, featuring distinct decoding pathways and achieving competitive results with modest parameters. https://arxiv.org/abs//2506.18871 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
OmniGen2 is an open-source generative model for diverse tasks like text-to-image and image editing, featuring distinct decoding pathways and achieving competitive results with modest parameters. https://arxiv.org/abs//2506.18871 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
Loading …
Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play