All
Search
Images
Videos
Shorts
Maps
News
More
Shopping
Flights
Travel
Notebook
Report an inappropriate content
Please select one of the options below.
Not Relevant
Offensive
Adult
Child Sexual Abuse
Length
All
Short (less than 5 minutes)
Medium (5-20 minutes)
Long (more than 20 minutes)
Date
All
Past 24 hours
Past week
Past month
Past year
Resolution
All
Lower than 360p
360p or higher
480p or higher
720p or higher
1080p or higher
Source
All
Dailymotion
Vimeo
Metacafe
Hulu
VEVO
Myspace
MTV
CBS
Fox
CNN
MSN
Price
All
Free
Paid
Clear filters
SafeSearch:
Moderate
Strict
Moderate (default)
Off
Filter
24:05
ORPO: NEW DPO Alignment and SFT Method for LLM
4.9K views
Mar 24, 2024
YouTube
Discover AI
0:28
The Truth About LLM Alignment: SFT, RLHF, and DPO
277 views
2 months ago
YouTube
Ryan Banze
LLM Reinforcement Learning Fine-Tuning DeepSeek Method GRPO
40 views
11 months ago
git.ir
1:20:54
LLM Alignment (RLHF, DPO, ORPO) + Hands-on Project
9.7K views
3 months ago
YouTube
BrainOmega
LLM Fine-Tuning Mastery: Basic to Advanced & Cloud Deploy
8 months ago
git.ir
10:38
Stop Using RLHF: How to Align & Control LLMs (DPO Guide)
330 views
3 months ago
YouTube
Shane | LLM Implementation
39:41
ORPO Explained: Superior LLM Alignment Technique vs. DPO/RLHF
3K views
Apr 9, 2024
YouTube
AI Anytime
12:55
DPO Coding | Direct Preference Optimization (DPO) Code impleme
…
403 views
1 year ago
YouTube
AILinkDeepTech
0:14
Aligning LLMs with Human Preferences
3 views
1 month ago
YouTube
The AI Opus
6:26
NEW WizardLM-2 8x22B: Fine-tune & Stage-DPO align
2.5K views
Apr 15, 2024
YouTube
Discover AI
41:28
LLMs | Alignment of Language Models: Contrastive Learning | Le
…
1.6K views
Sep 26, 2024
YouTube
LCS2
39:15
Advanced LLM Post-Training: SFT, DPO, Reinforcement Learning w/
…
151 views
3 months ago
YouTube
Youth AI Initiative
58:07
Aligning LLMs with Direct Preference Optimization
34.1K views
Feb 8, 2024
YouTube
DeepLearningAI
왜 안전한 AI를 만들기 어려울까? | RLHF, DPO로 만드는 윤리적 AI | Gu
…
601 views
2 months ago
linkedin.com
40:55
Fast Fine Tuning and DPO Training of LLMs using Unsloth
5.9K views
Mar 25, 2024
YouTube
AI Anytime
28:53
Fine-tuning LLMs on Human Feedback (RLHF + DPO)
21.6K views
Mar 3, 2025
YouTube
Shaw Talebi
12:30
How does DPO improve the LLM's performance? | Simple Explanation
198 views
Jan 29, 2025
YouTube
MLWorks
5:08
LLM Alignment Methods - DPO vs IPO vs KTO vs PCL
1.6K views
Jan 27, 2024
YouTube
Fahd Mirza
3:36:14
LLM Fine-Tuning Crash Course: Finetune model on PDFs, Instructi
…
7.6K views
3 months ago
YouTube
Sunny Savita
13:23
An update on DPO vs PPO for LLM alignment
3.9K views
Jul 22, 2024
YouTube
Nathan Lambert
1:54
【大模型入门】大模型全流程学习一图搞定!LLM、预训练、SFT监督微
…
3.6K views
4 months ago
bilibili
我十六条
21:15
Direct Preference Optimization (DPO) - How to fine-tune LLMs dir
…
30.7K views
Jun 21, 2024
YouTube
Serrano.Academy
22:35
Train LLM Easily with Llama Factory LORA, SFT, DPO, etc. | GUI Traini
…
752 views
Feb 11, 2024
YouTube
Code Port
3:28
Enhancing Song Generation in LLMs using DPO-based Multi-Pref
…
7 views
2 months ago
YouTube
Quang Phạm Việt
0:37
সবাই ভাবে হাল্ক শক্তিশালী… কিন্তু তার
…
1 month ago
YouTube
The islamic story
4:58
构建大语言模型,DPO训练方法,原理和实现
16K views
Nov 1, 2023
bilibili
蓝斯诺特
1:36
15/2/26 DINNER CONTRIBUTION FOR MAHA SHIVARATHIRI BY DR.
…
2 views
4 weeks ago
YouTube
ANBIN VAMSAM ARAKKATTALAI for DESTIT…
59:40
Direct Preference Optimization (DPO) in 1 hour
2.3K views
5 months ago
YouTube
Zachary Huang
8:34
This tool will change the way apps are created and connected to APIs.
915 views
4 months ago
YouTube
TutorialTec
24:56
Make AI Think Like YOU: A Guide to LLM Alignment
2.5K views
Nov 12, 2024
YouTube
Adam Lucek
See more videos
More like this
Feedback