Lordog/dive-into-llms
Summary
《动手学大模型》是由上海交通大学课程讲义拓展而来的开源编程实践教程,涵盖微调、提示学习、知识编辑、数学推理、模型水印、越狱攻击、隐写术等大模型相关主题,面向初学者完全免费开放。
View Cached Full Text
Cached at: 05/09/26, 12:31 PM
Lordog/dive-into-llms
Source: https://github.com/Lordog/dive-into-llms
《动手学大模型》系列编程实践教程
💡 Updates
2025/06/06 感谢各位朋友们的关注和积极反馈!我们从以下两个方面对本教程进行了更新:
- 上线国产化《大模型开发全流程》公益教程(含PPT、实验手册和视频),此处特别感谢华为昇腾社区的支持!
- 在原系列编程实践教程的基础上进行内容更新,并增加了新的主题(数学推理、GUI Agent、大模型对齐、隐写术等)!
🎯 项目动机
《动手学大模型》系列编程实践教程,由上海交通大学《自然语言处理前沿技术》(NIS8021)、《人工智能安全技术》课程(NIS3353)讲义拓展而来(教师:张倬胜),旨在提供大模型相关的入门编程参考。本教程属公益性质、完全免费。通过简单实践,帮助同学们快速入门大模型,更好地开展课程设计或学术研究。
📚 教程目录
| 教程内容 | 简介 | 地址 |
|---|---|---|
| 微调与部署 | 预训练模型微调与部署指南:想提升预训练模型在指定任务上的性能?让我们选择合适的预训练模型,在特定任务上进行微调,并将微调后的模型部署成方便使用的Demo! | [课件] [教程] [脚本] |
| 提示学习与思维链 | 大模型的API调用与推理指南:“AI在线求鼓励?大模型对一些问题的回答令人大跌眼镜,但它可能只是想要一句「鼓励」” | [课件] [教程] [脚本] |
| 知识编辑 | 语言模型的编辑方法和工具:想操控语言模型在对指定知识的记忆?让我们选择合适的编辑方法,对特定知识进行编辑,并将对编辑后的模型进行验证! | [课件] [教程] [脚本] |
| 数学推理 | 如何让大模型学会数学推理?让我们快速蒸馏一个迷你R1! | [课件] [教程] [脚本] |
| 模型水印 | 语言模型的文本水印:在语言模型生成的内容中嵌入人类不可见的水印 | [课件] [教程] [脚本] |
| 越狱攻击 | 想要得到更好的安全,要先从学会攻击开始。让我们了解越狱攻击如何撬开大模型的嘴! | [课件] [教程] [脚本] |
| 大模型隐写 | “看不见的墨水”!想让大模型在流畅回答的同时,悄悄携带只有“自己人”能识别的信息吗?大模型隐写告诉你! | [课件] [教程] [脚本] |
| 多模态模型 | 作为能够更充分模拟真实世界的多模态大语言模型,其如何实现更强大的多模态理解和生成能力?多模态大语言模型是否能够帮助实现AGI? | [课件] [教程] [脚本] |
| GUI智能体 | 想要饭来张口、解放双手?那么让我们一起来让AI Agent替你点外卖、回消息、购物比价吧! | [课件] [教程] [脚本] |
| 智能体安全 | 大模型智能体迈向了未来操作系统之旅。然而,大模型在开放智能体场景中能意识到风险威胁吗? | [课件] [教程] [脚本] |
| RLHF安全对齐 | 基于PPO的RLHF实验指南:本教程”十分危险“,阅读后请检查你的大模型是否在冷笑。 | [课件] [教程] [脚本] |
🔥 新上线:国产化《大模型开发全流程》
-
✨ 我们联合华为昇腾推出的《大模型开发全流程》公益教程正式上线!前沿技术+代码实践,手把手带你玩转AI大模型 ✨:
在《动手学大模型》原系列教程的基础上,我们联合华为开发了《大模型开发全流程》系列课程。本系列教程基于昇腾基础软硬件开发,覆盖PPT、实验手册、视频等教程形式。该教程分为初级、中级、高级系列,面向不同的大模型实践需求,旨在将前沿技术通过代码实践的方式,为相关研究者、开发者由浅入深地提供快速上手、应用昇腾已支持模型和全新模型迁移调优的全流程开发指南。
-
🚀 前往昇腾社区探索《大模型开发全流程》系列课程:
👉《大模型开发学习专区》@ 昇腾社区 👈
-
✨ 课程内容展示 ✨
🙏 免责声明
本教程所有内容仅仅来自于贡献者的个人经验、互联网数据、日常科研工作中的相关积累。所有技巧仅供参考,不保证百分百正确。若有任何问题,欢迎提交 Issue 或 PR。另本项目所用徽章来自互联网,如侵犯了您的图片版权请联系我们删除,谢谢。
🤝 欢迎贡献
本教程目前是一个正在进行中的项目,如有疏漏在所难免,欢迎任何的PR及issue讨论。
❤️ 贡献者列表
感谢以下老师和同学对本项目的支持与贡献:
《动手学大模型》系列教程开发团队:
《大模型开发全流程》系列教程开发团队:
🌟 Star History
Similar Articles
@wsl8297: Sharing an easy-to-read open-source book 'Foundations of Large Models'. From an introduction to large language models to architectural evolution, then to key technologies such as Prompt engineering, parameter-efficient fine-tuning, model editing, retrieval-augmented generation (RAG), all in one book. GitHub: https://github.com/ZJU-LLMs/…
The Zhejiang University team open-sourced an easy-to-understand textbook on large models 'Foundations of Large Models', covering from architectural evolution to key technologies like RAG, accompanied by the Agent-Kernel multi-agent framework.
@NFTCPS: Brothers, doing AI without large models is like doing nothing! Today I have to recommend an open-source masterpiece 'Foundations of LLMs' to you. Don't wait, just read it! This book doesn't beat around the bush—it goes deep from the start! From getting started with large language models to architectural evolution, and then it breaks down Prompt engineering, parameter-efficient fine-tuning, model editing, RAG (Retrieval-Augmented Generation) and other hardcore techniques in one go—a one-stop service.
This article promotes the open-source book 'Foundations of LLMs', which systematically explains knowledge about large language models, and introduces the multi-agent development framework Agent-Kernel.
@Khazix0918: https://x.com/Khazix0918/status/2046082879109959807
This article is a beginner's guide to installing and using Claude Code for Chinese users, covering installation methods for both Mac and Windows (including setups without a VPN), as well as how to integrate domestic models (such as GLM-5.1) as alternatives to Claude's native models.
@wsl8297: UC's Open Course on Reinforcement Learning for LLMs uses a 'theory + practice' approach to thoroughly explain key AI training techniques from the ground up, helping you systematically build a complete framework spanning from RL to LLM training. Comprehensive curriculum paired with complete resources: lecture slides, full videos, and practical exercises are all provided so you can start implementing right away…
Assistant Professor Ernest K. Ryu at UCLA offers the open course 'Reinforcement Learning for Large Language Models,' comprehensively analyzing key LLM training techniques like RLHF, PPO, and DPO alongside their supporting resources through a blend of theory and practice. The course provides developers and researchers with a systematic learning path from foundational algorithms to practical deployment.
We’re proud to open-source LIDARLearn [R] [D] [P]
LIDARLearn is an open-source PyTorch library for 3D point cloud deep learning that unifies 56 pre-configured models with built-in cross-validation and automatic publication-ready LaTeX report generation. The framework supports supervised, self-supervised, and parameter-efficient fine-tuning methods across datasets like ModelNet40, ShapeNet, and remote sensing benchmarks.