ViewTube

ViewTube
Sign inSign upSubscriptions
Filters

Upload date

Type

Duration

Sort by

Features

Reset

52,625 results

IBM Technology
What is Mixture of Experts?

Want to play with the technology yourself? Explore our interactive demo → https://ibm.biz/BdK8fn Learn more about the ...

7:58
What is Mixture of Experts?

49,099 views

1 year ago

Maarten Grootendorst
A Visual Guide to Mixture of Experts (MoE) in LLMs

In this highly visual guide, we explore the architecture of a Mixture of Experts in Large Language Models (LLM) and Vision ...

19:44
A Visual Guide to Mixture of Experts (MoE) in LLMs

49,123 views

1 year ago

AI Papers Academy
Introduction to Mixture-of-Experts | Original MoE Paper Explained

In this video we go back to the extremely important Google paper which introduced the Mixture-of-Experts (MoE) layer with ...

4:41
Introduction to Mixture-of-Experts | Original MoE Paper Explained

11,620 views

1 year ago

Chris Hay
MoE Models Don't Work Like You Think - Inside GPT-OSS

Many people think that mixture of expert models have domain experts, i.e. math experts, code experts, language experts.

18:28
MoE Models Don't Work Like You Think - Inside GPT-OSS

3,838 views

3 weeks ago

bycloud
1 Million Tiny Experts in an AI? Fine-Grained MoE Explained

To try everything Brilliant has to offer—free—for a full 30 days, visit https://brilliant.org/bycloud/ . You'll also get 20% off an annual ...

12:29
1 Million Tiny Experts in an AI? Fine-Grained MoE Explained

55,281 views

1 year ago

Cerebras
Mixture of Experts Explained: How to Build, Train & Debug MoE Models in 2025

Mixture-of-Experts (MoE) models now power leading AI systems like GPT-4, Qwen3, DeepSeek-v3, and Gemini 1.5. But behind ...

4:32
Mixture of Experts Explained: How to Build, Train & Debug MoE Models in 2025

1,577 views

6 months ago

SaM Solutions
Mixture-of-Experts (MoE) LLMs: The Future of Efficient AI Models

Imagine having a whole team of specialists at your disposal, each an expert in a different field, and a smart coordinator who ...

6:01
Mixture-of-Experts (MoE) LLMs: The Future of Efficient AI Models

302 views

5 months ago

Organic Mechanisms
How To Create And Use A Pharmacophore In MOE | MOE Tutorial

Molecular Operating Environment (MOE) tutorial covering how to create and use a pharmacophore in MOE. When docking ...

4:12
How To Create And Use A Pharmacophore In MOE | MOE Tutorial

7,316 views

3 years ago

LLM Implementation
How 120B+ Parameter Models Run on One GPU (The MoE Secret)

How is it possible for a 120 billion parameter AI model to run on a single consumer GPU? This isn't magic—it's the result of ...

6:47
How 120B+ Parameter Models Run on One GPU (The MoE Secret)

1,590 views

5 months ago

No Hype AI
How Did They Do It? DeepSeek V3 and R1 Explained

DeepSeek: The First Open-Weight Reasoning Model! In this video, I'll break down DeepSeek's two flagship models— V3 and R1 ...

11:15
How Did They Do It? DeepSeek V3 and R1 Explained

47,359 views

1 year ago

Marktechpost AI
NVIDIA Releases Nemotron 3: Hybrid Mamba Transformer Models With Latent MoE .....

NVIDIA Nemotron 3 is an open family of hybrid Mamba Transformer MoE models, designed for agentic AI with long context and ...

5:30
NVIDIA Releases Nemotron 3: Hybrid Mamba Transformer Models With Latent MoE .....

604 views

1 month ago

Matt Williams
What are the different types of models - The Ollama Course

Dive into the world of Ollama and discover the various types of AI models at your fingertips. This comprehensive guide breaks ...

6:49
What are the different types of models - The Ollama Course

38,614 views

1 year ago

bycloud
The REAL AI Architecture That Unifies Vision & Language

Get started now with open source & privacy focused password manager by Proton! https://proton.me/pass/bycloudai In this video, ...

10:13
The REAL AI Architecture That Unifies Vision & Language

44,653 views

7 months ago

Paper With Video
[2024 Best AI Paper] MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts

This video was created using https://paperspeech.com. If you'd like to create explainer videos for your own papers, please visit the ...

11:36
[2024 Best AI Paper] MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts

71 views

1 year ago

Cerebras
Daria Soboleva   Training and Serving MoE Models Efficiently

... models efficiently before I start a quick intro about myself um I am researching LMS at Cerebras one of my recent gigs is MOE ...

9:18
Daria Soboleva Training and Serving MoE Models Efficiently

212 views

1 month ago

AI with Lena Hall
Transformers vs MoE vs RNN vs Hybrid: Intuitive LLM Architecture Guide

Most developers default to transformers without understanding the alternatives. This video breaks down the intuition behind four ...

16:56
Transformers vs MoE vs RNN vs Hybrid: Intuitive LLM Architecture Guide

19,219 views

4 months ago

Mehdi Hosseini Moghadam
Mixture of Experts (MoE) Explained: How GPT-4 & Switch Transformer Scale to Trillions!

What You'll Learn In this comprehensive tutorial, we dive deep into Mixture of Experts (MoE) - the revolutionary architecture that ...

12:59
Mixture of Experts (MoE) Explained: How GPT-4 & Switch Transformer Scale to Trillions!

210 views

7 months ago

New Machina
What is LLM Mixture of Experts ?

You'll also learn about real-world MoE models like Mixtral, and DeepSeek, which achieve state-of-the-art performance while ...

5:41
What is LLM Mixture of Experts ?

4,927 views

11 months ago

AI Papers Academy
The Era of Hierarchical Reasoning Models

Check out Emergent here: https://emergent.1stcollab.com/aipapersacademy Can AI models learn to reason more like humans?

8:56
The Era of Hierarchical Reasoning Models

9,048 views

5 months ago

Matthew Berman
This Tiny Model is Insane... (7m Parameters)

Build your first app today with Mocha: https://www.getmocha.com?utm_source=matthew_berman Download Humanities Last ...

13:53
This Tiny Model is Insane... (7m Parameters)

124,810 views

3 months ago

IBM Technology
What are Transformers (Machine Learning Model)?

Learn more about Transformers → http://ibm.biz/ML-Transformers Learn more about AI → http://ibm.biz/more-about-ai Check out ...

5:51
What are Transformers (Machine Learning Model)?

693,323 views

3 years ago

Learn Meta-Analysis
Change this setting in LM Studio to run MoE LLMs faster.

I changed 2 settings in LM Studio and I increased my t/s by about 4x. My 8gb gpu (rtx 4060) now runs GPT OSS 120b at 20t/s and ...

8:45
Change this setting in LM Studio to run MoE LLMs faster.

13,158 views

4 months ago

IBM Technology
Granite 4.0: Small AI Models, Big Efficiency

Ready to become a certified watsonx AI Assistant Engineer? Register now and use code IBMTechYT20 for 20% off of your exam ...

11:16
Granite 4.0: Small AI Models, Big Efficiency

35,101 views

3 months ago

Next Tech and AI
Small Language Models Under 4GB: What Actually Works?

Never get stuck without AI again. Run three Small Language Models (SLMs)—also called Local LLMs—TinyLlama, Gemma-3 and ...

5:51
Small Language Models Under 4GB: What Actually Works?

6,134 views

5 months ago

AI Papers Academy
Fast Inference of Mixture-of-Experts Language Models with Offloading

However, MoE models usually have a large memory footprint since we need to load the weights of all experts. This makes it hard ...

11:59
Fast Inference of Mixture-of-Experts Language Models with Offloading

1,941 views

2 years ago