ViewTube

ViewTube
Sign inSign upSubscriptions
Filters

Upload date

Type

Duration

Sort by

Features

Reset

26 results

Ready Tensor
RunPod Serverless Deployment Tutorial: Deploy Your Fine-Tuned LLM with vLLM

In this video, we walk through how to deploy a fine-tuned large language model from Hugging Face to a RunPod Serverless ...

12:42
RunPod Serverless Deployment Tutorial: Deploy Your Fine-Tuned LLM with vLLM

50 views

6 days ago

Ready Tensor
SageMaker LLM Deployment Tutorial: Serve Fine-Tuned Models with vLLM

In this video, you'll learn how to deploy a fine-tuned large language model from Hugging Face to AWS SageMaker using the vLLM ...

6:23
SageMaker LLM Deployment Tutorial: Serve Fine-Tuned Models with vLLM

0 views

6 days ago

Universe of AI
GLM-4.7-Flash: 42x Cheaper Than Claude, Actually Good at Coding!

GLM-4.7-Flash is crushing coding benchmarks at a fraction of the cost of Claude and GPT-5. In this video, I break down the ...

8:19
GLM-4.7-Flash: 42x Cheaper Than Claude, Actually Good at Coding!

11,058 views

5 days ago

Nate Herk | AI Automation
Master 95% of Claude Code in 36 Mins (as a beginner)

Full courses + unlimited support: https://www.skool.com/ai-automation-society-plus/about All my FREE resources: ...

36:58
Master 95% of Claude Code in 36 Mins (as a beginner)

53,494 views

4 days ago

DevOps in Action
🚀 7 Ways to Run/Deploy Any LLMs Locally - Simple Methods | Docker Compose (CPU + GPU Hybrid Mode)

7 Ways to Run/Deploy Any LLMs Locally - Simple Methods | Docker Compose (CPU + GPU Hybrid Mode) In this video: Learn ...

28:14
🚀 7 Ways to Run/Deploy Any LLMs Locally - Simple Methods | Docker Compose (CPU + GPU Hybrid Mode)

0 views

16 hours ago

OpenShift
Deploy a model with vLLM and Llama Stack on MCP servers

Intel's Alex Sin demonstrates how Model Context Protocol (MCP) servers running with Llama Stack on Red Hat OpenShift AI ...

8:25
Deploy a model with vLLM and Llama Stack on MCP servers

184 views

6 days ago

Leon van Zyl
Why Pay $200 for Claude Code When This Costs $6?

Access ALL video resources & get personalized help in my community: ...

9:25
Why Pay $200 for Claude Code When This Costs $6?

44,918 views

5 days ago

Henry Un
Run local MLX LLM  supports multiple Users' Requests and provided response in parallel.

To run multiple user requests in parallel with a local MLX LLM model in Apple Silicon (M GPU), use the vLLM server, and Page ...

19:52
Run local MLX LLM supports multiple Users' Requests and provided response in parallel.

0 views

7 days ago

KI erklärt in 5 Minuten
vLLM Boosting mit LMCache Germ

LMCache vs. vLLM: Architektur für effizienten persistenten KV-Cache Der bereitgestellte Text vergleicht vLLM, eine ...

7:21
vLLM Boosting mit LMCache Germ

0 views

4 days ago

Sam Witteveen
Open Responses - The NEW Standard API for Open Models

In this video, I look at the Open Responses Standard that's been released by OpenAI to support open models with their ...

15:53
Open Responses - The NEW Standard API for Open Models

8,898 views

5 days ago

Julian Goldie SEO
NEW GLM 4.7 Flash Update is INSANE!

Want to make money and save time with AI? Get AI Coaching, Support & Courses ...

8:43
NEW GLM 4.7 Flash Update is INSANE!

2,695 views

5 days ago

DevOps in Action
GLM-4.7 Flash: How to use GLM 4.7 Flash for free - The New King of Local AI Coding? (30B MoE) 🚀

Zhipu AI just dropped GLM-4.7-Flash, and it's shaking up the open-source AI world. In this video, we break down why this 30B ...

10:24
GLM-4.7 Flash: How to use GLM 4.7 Flash for free - The New King of Local AI Coding? (30B MoE) 🚀

101 views

1 day ago

AI Podcast Series. Byte Goose AI.
[Unsloth Fine-Tuning] Unsloth 3.3x Faster Embedding Fine-tuning with 20% Less Memory LoRA / QLoRA.

Usually, fine-tuning these models is a resource hog. But the team at Unsloth has just changed the game. We're diving into their ...

5:24
[Unsloth Fine-Tuning] Unsloth 3.3x Faster Embedding Fine-tuning with 20% Less Memory LoRA / QLoRA.

76 views

2 days ago

Donato Capitella
Kimi-K2(1T)/GLM 4.7(355B) on a 4-Node Strix Halo Cluster - 512GB of Unified Memory

In this video, I demonstrate running large-scale Mixture-of-Experts (MoE) models on a 4-node cluster of AMD Strix Halo systems.

9:36
Kimi-K2(1T)/GLM 4.7(355B) on a 4-Node Strix Halo Cluster - 512GB of Unified Memory

6,255 views

2 days ago

Devs Kingdom
KLing Motion + Higgsfield: Best Free AI Influencer that changes everything

Creating Quality Content with AI Influencer used to be a complex and expensive task, Now it is free, HD and easy with Higgsfield.

8:32
KLing Motion + Higgsfield: Best Free AI Influencer that changes everything

45 views

2 days ago

Self-Hosted AI Automation
Self Hosted LLM + n8n Market Overview JSON → WordPress + Social Posting

An end-to-end n8n automation that generates a crypto market overview (JSON), publishes a WordPress post, and distributes the ...

2:15
Self Hosted LLM + n8n Market Overview JSON → WordPress + Social Posting

4 views

7 days ago

LOUIS PYTHON
AI Agents with WeDLM Python Script! #WeDLM #Generative

Ready to supercharge your AI agents and generative AI projects? Discover the incredible power of integrating Python with the ...

3:04
AI Agents with WeDLM Python Script! #WeDLM #Generative

8 views

6 days ago

Prashant Lakhera
Kickoff & Overview: From Software & DevOps Engineer → Generative AI Engineer

Kickoff & Overview Session: From Software & DevOps Engineer → Generative AI Engineer (4-Month Hands-On Journey) I'm ...

58:00
Kickoff & Overview: From Software & DevOps Engineer → Generative AI Engineer

83 views

7 days ago

Tech With Machines
INFERENCE ENGINES In AI Explained Simply

inferenceengine #aiagents #ai #machinelearning #deeplearning #knowledgegraph #expertsystems #logicprogramming An ...

5:45
INFERENCE ENGINES In AI Explained Simply

0 views

2 days ago

Curious Enough
LLM vs vLLM: Efficiency and Scaling Explained

While a **Large Language Model (LLM)** functions as the core intelligence capable of predicting text and answering prompts, ...

5:49
LLM vs vLLM: Efficiency and Scaling Explained

126 views

3 days ago