大型語言模型(LLMs)與強化學習(RL)培訓
Large Language Models (LLMs) 是高級類型的神經網路,旨在根據接收到的輸入來理解和生成類似人類的文本。Reinforcement Learning (RL) 是一種機器學習,其中代理通過在環境中執行操作來學習做出決策,以最大化累積獎勵。
這種以講師為主導的現場培訓(在線或遠端)面向希望全面瞭解 Large Language Models (LLMs) 和 Reinforcement Learning (RL) 的中級數據科學家。
在培訓結束時,參與者將能夠:
- 瞭解變壓器模型的元件和功能。
- 針對特定任務和應用程式優化和微調 LLM。
- 瞭解強化學習的核心原則和方法。
- 瞭解強化學習技術如何提高 LLM 的性能。
課程形式
- 互動講座和討論。
- 大量的練習和練習。
- 在現場實驗室環境中動手實施。
課程自定義選項
- 如需申請本課程的定製培訓,請聯繫我們進行安排。
課程簡介
Large Language Models (LLMs) 簡介
- LLM概述
- 定義和意義
- 當今人工智慧中的應用
變壓器架構
- 什麼是變壓器,它是如何工作的?
- 主要元件和特點
- 嵌入和位置編碼
- 多頭注意力
- 前饋神經網路
- 歸一化和殘差連接
變壓器型號
- 自注意力機制
- 編碼器-解碼器架構
- 位置嵌入
- BERT(來自 Transformer 的雙向編碼器表示)
- GPT(產生式預訓練轉換器)
性能優化和陷阱
- 上下文長度
- 曼巴和狀態空間模型
- 閃光注意力
- 稀疏變壓器
- 視覺變壓器
- 量化的重要性
改進變壓器
- 檢索增強文本生成
- 模型混合
- 思想之樹
微調
- 低秩適應理論
- 使用 QLora 進行微調
LLM 中的縮放定律和優化
- LLM擴展法的重要性
- 數據和模型大小縮放
- 計算擴展
- 參數效率縮放
優化
- 模型大小、數據大小、計算預算和推理需求之間的關係
- 優化 LLM 的性能和效率
- 用於訓練和微調 LLM 的最佳實踐和工具
訓練和微調 LLM
- 從頭開始培訓 LLM 的步驟和挑戰
- 數據採集與維護
- 大規模數據、CPU 和記憶體要求
- 優化挑戰
- 開源 LLM 的前景
Reinforcement Learning (RL) 的基礎知識
- Reinforcement Learning 簡介
- 通過積極強化學習
- 定義和核心概念
- 瑪律可夫決策過程 (MDP)
- 動態規劃
- 蒙特卡羅方法
- 時差學習
深 Reinforcement Learning
- 深度 Q 網路 (DQN)
- 近端策略優化 (PPO)
- Element秒,共 Reinforcement Learning
LLM 和 Reinforcement Learning 的集成
- 將 LLM 與 Reinforcement Learning 相結合
- RL在LLM中的使用方式
- Reinforcement Learning 人工反饋 (RLHF)
- RLHF的替代品
案例研究和應用
- 實際應用
- 成功案例和挑戰
高級主題
- 先進技術
- 高級優化方法
- 尖端研發
摘要和後續步驟
最低要求
- 基本瞭解 Machine Learning
觀眾
- 數據科學家
- 軟體工程師
Open Training Courses require 5+ participants.
大型語言模型(LLMs)與強化學習(RL)培訓 - Booking
大型語言模型(LLMs)與強化學習(RL)培訓 - Enquiry
大型語言模型(LLMs)與強化學習(RL) - 咨詢詢問
咨詢詢問
Upcoming Courses
相關課程
Advanced LangGraph: Optimization, Debugging, and Monitoring Complex Graphs
35 時間:LangGraph is a framework for building stateful, multi-actor LLM applications as composable graphs with persistent state and control over execution.
This instructor-led, live training (online or onsite) is aimed at advanced-level AI platform engineers, DevOps for AI, and ML architects who wish to optimize, debug, monitor, and operate production-grade LangGraph systems.
By the end of this training, participants will be able to:
- Design and optimize complex LangGraph topologies for speed, cost, and scalability.
- Engineer reliability with retries, timeouts, idempotency, and checkpoint-based recovery.
- Debug and trace graph executions, inspect state, and systematically reproduce production issues.
- Instrument graphs with logs, metrics, and traces, deploy to production, and monitor SLAs and costs.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
Advanced Ollama Model Debugging & Evaluation
35 時間:Advanced Ollama Model Debugging & Evaluation is an in-depth course focused on diagnosing, testing, and measuring model behavior when running local or private Ollama deployments.
This instructor-led, live training (online or onsite) is aimed at advanced-level AI engineers, ML Ops professionals, and QA practitioners who wish to ensure reliability, fidelity, and operational readiness of Ollama-based models in production.
By the end of this training, participants will be able to:
- Perform systematic debugging of Ollama-hosted models and reproduce failure modes reliably.
- Design and execute robust evaluation pipelines with quantitative and qualitative metrics.
- Implement observability (logs, traces, metrics) to monitor model health and drift.
- Automate testing, validation, and regression checks integrated into CI/CD pipelines.
Format of the Course
- Interactive lecture and discussion.
- Hands-on labs and debugging exercises using Ollama deployments.
- Case studies, group troubleshooting sessions, and automation workshops.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
使用Ollama構建私有AI工作流程
14 時間:這是一個由講師指導的線上或線下培訓,針對希望使用Ollama實現安全高效的AI驅動工作流程的高級專業人士。
通過本培訓,參與者將能夠:
- 部署和配置Ollama以進行私有AI處理。
- 將AI模型集成到安全的企業工作流程中。
- 在保持數據隱私的同時,優化AI性能。
- 利用本地AI功能自動化業務流程。
- 確保符合企業安全與治理政策。
Claude AI 面向開發者:構建AI驅動的應用程序
14 時間:這個面向希望將Claude AI集成到他們的應用程序中、構建AI驅動的聊天機器人以及通過AI驅動的自動化來增強軟件功能的中級軟件開發人員和AI工程師的香港(在線或現場) Instructor-led, live training (online or onsite)。
在這次培訓結束時,參與者將能夠:
- 使用Claude AI API將AI集成到應用程序中。
- 開發AI驅動的聊天機器人和虛擬助手。
- 利用AI驅動的自動化和NLP增強應用程序。
- 對不同的用例優化和微調Claude AI模型。
Claude AI 適用於 Workflow Automation 和 Productivity
14 時間:這門由講師主導的現場培訓在香港(在線或現場)旨在幫助有志於將Claude AI整合到日常工作流程中以提高效率和自動化的初級專業人員。
培訓結束時,參加者將能夠:
- 利用Claude AI來自動化重複性任務和精簡工作流程。
- 利用人工智慧驅動的自動化技術提高個人和團隊的生產力。
- 將Claude AI與現有的商業工具和平台整合。
- 優化AI驅動的決策制定和任務管理。
使用Ollama部署和優化LLM
14 時間:這是一場由講師指導的現場培訓,地點在香港(線上或現場),適合希望使用Ollama部署、優化和整合LLM的中級專業人士。
在培訓結束時,參與者將能夠:
- 使用Ollama設置和部署LLM。
- 優化AI模型以提升性能和效率。
- 利用GPU加速提升推理速度。
- 將Ollama整合到工作流程和應用程式中。
- 監控和維護AI模型的長期性能。
Fine-Tuning 與在 Ollama 上自訂 AI 模型
14 時間:這項由講師指導的香港(線上或線下)培訓,針對希望微調和自訂Ollama上的AI模型以提升性能和特定領域應用的高級專業人士。
在培訓結束時,參與者將能夠:
- 在Ollama上設置高效的AI模型微調環境。
- 準備用於監督式微調和強化學習的數據集。
- 優化AI模型的性能、準確性和效率。
- 在生產環境中部署自訂模型。
- 評估模型改進並確保其穩健性。
Claude AI 簡介:對話式 AI 與商業應用
14 時間:這項由講師主持的現場培訓(在線或現場)旨在幫助希望了解Claude AI的基本原理並利用它進行商業應用的初級商業專業人士、客戶支持團隊和技術愛好者。
培訓結束時,參與者將能夠:
- 了解Claude AI的能力和用例。
- 有效地設置和與Claude AI互動。
- 利用對話式AI自動化業務工作流程。
- 利用AI驅動的解決方案提升客戶互動和支持。
LangGraph Applications in Finance
35 時間:LangGraph is a framework for building stateful, multi-actor LLM applications as composable graphs with persistent state and control over execution.
This instructor-led, live training (online or onsite) is aimed at intermediate-level to advanced-level professionals who wish to design, implement, and operate LangGraph-based finance solutions with proper governance, observability, and compliance.
By the end of this training, participants will be able to:
- Design finance-specific LangGraph workflows aligned to regulatory and audit requirements.
- Integrate financial data standards and ontologies into graph state and tooling.
- Implement reliability, safety, and human-in-the-loop controls for critical processes.
- Deploy, monitor, and optimize LangGraph systems for performance, cost, and SLAs.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
LangGraph Foundations: Graph-Based LLM Prompting and Chaining
14 時間:LangGraph is a framework for building graph-structured LLM applications that support planning, branching, tool use, memory, and controllable execution.
This instructor-led, live training (online or onsite) is aimed at beginner-level developers, prompt engineers, and data practitioners who wish to design and build reliable, multi-step LLM workflows using LangGraph.
By the end of this training, participants will be able to:
- Explain core LangGraph concepts (nodes, edges, state) and when to use them.
- Build prompt chains that branch, call tools, and maintain memory.
- Integrate retrieval and external APIs into graph workflows.
- Test, debug, and evaluate LangGraph apps for reliability and safety.
Format of the Course
- Interactive lecture and facilitated discussion.
- Guided labs and code walkthroughs in a sandbox environment.
- Scenario-based exercises on design, testing, and evaluation.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
LangGraph in Healthcare: Workflow Orchestration for Regulated Environments
35 時間:LangGraph enables stateful, multi-actor workflows powered by LLMs with precise control over execution paths and state persistence. In healthcare, these capabilities are crucial for compliance, interoperability, and building decision-support systems that align with medical workflows.
This instructor-led, live training (online or onsite) is aimed at intermediate-level to advanced-level professionals who wish to design, implement, and manage LangGraph-based healthcare solutions while addressing regulatory, ethical, and operational challenges.
By the end of this training, participants will be able to:
- Design healthcare-specific LangGraph workflows with compliance and auditability in mind.
- Integrate LangGraph applications with medical ontologies and standards (FHIR, SNOMED CT, ICD).
- Apply best practices for reliability, traceability, and explainability in sensitive environments.
- Deploy, monitor, and validate LangGraph applications in healthcare production settings.
Format of the Course
- Interactive lecture and discussion.
- Hands-on exercises with real-world case studies.
- Implementation practice in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
LangGraph for Legal Applications
35 時間:LangGraph is a framework for building stateful, multi-actor LLM applications as composable graphs with persistent state and precise control over execution.
This instructor-led, live training (online or onsite) is aimed at intermediate-level to advanced-level professionals who wish to design, implement, and operate LangGraph-based legal solutions with the necessary compliance, traceability, and governance controls.
By the end of this training, participants will be able to:
- Design legal-specific LangGraph workflows that preserve auditability and compliance.
- Integrate legal ontologies and document standards into graph state and processing.
- Implement guardrails, human-in-the-loop approvals, and traceable decision paths.
- Deploy, monitor, and maintain LangGraph services in production with observability and cost controls.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
Building Dynamic Workflows with LangGraph and LLM Agents
14 時間:LangGraph is a framework for composing graph-structured LLM workflows that support branching, tool use, memory, and controllable execution.
This instructor-led, live training (online or onsite) is aimed at intermediate-level engineers and product teams who wish to combine LangGraph’s graph logic with LLM agent loops to build dynamic, context-aware applications such as customer support agents, decision trees, and information retrieval systems.
By the end of this training, participants will be able to:
- Design graph-based workflows that coordinate LLM agents, tools, and memory.
- Implement conditional routing, retries, and fallbacks for robust execution.
- Integrate retrieval, APIs, and structured outputs into agent loops.
- Evaluate, monitor, and harden agent behavior for reliability and safety.
Format of the Course
- Interactive lecture and facilitated discussion.
- Guided labs and code walkthroughs in a sandbox environment.
- Scenario-based design exercises and peer reviews.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
LangGraph for Marketing Automation
14 時間:LangGraph is a graph-based orchestration framework that enables conditional, multi-step LLM and tool workflows, ideal for automating and personalizing content pipelines.
This instructor-led, live training (online or onsite) is aimed at intermediate-level marketers, content strategists, and automation developers who wish to implement dynamic, branching email campaigns and content generation pipelines using LangGraph.
By the end of this training, participants will be able to:
- Design graph-structured content and email workflows with conditional logic.
- Integrate LLMs, APIs, and data sources for automated personalization.
- Manage state, memory, and context across multi-step campaigns.
- Evaluate, monitor, and optimize workflow performance and delivery outcomes.
Format of the Course
- Interactive lectures and group discussions.
- Hands-on labs implementing email workflows and content pipelines.
- Scenario-based exercises on personalization, segmentation, and branching logic.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
Ollama 入門指南:運行本地 AI 模型
7 時間:本課程為講師指導的香港(線上或線下)培訓,旨在幫助初級專業人士安裝、配置並使用Ollama在本地機器上運行AI模型。
培訓結束後,學員將能夠:
- 了解Ollama的基本原理及其功能。
- 設置Ollama以運行本地AI模型。
- 使用Ollama部署並與LLM進行互動。
- 優化AI工作負載的性能和資源使用。
- 探索本地AI部署在各行業中的應用案例。