Checkout

Cart () Loading...

    • Quantity:
    • Delivery:
    • Dates:
    • Location:

    $

Contact Sales

Adding New Knowledge to LLMs

This course provides a comprehensive, hands-on guide to the essential techniques for augmenting and customizing LLMs.

This course takes you on a complete journey from raw data to a fine-tuned, optimized model. You will begin by learning how to curate high-quality datasets and generate synthetic data with NVIDIA NeMo Curator. Next, you will dive deep into the crucial process of model evaluation, using benchmarks, LLM-as-a-judge, and the NeMo Evaluator to rigorously assess model performance. With a solid foundation in evaluation, you will then explore a suite of powerful customization techniques, including Continued Pretraining to inject new knowledge, Supervised Fine-Tuning to teach new skills, and Direct Preference Optimization (DPO) to align model behavior with human preferences.

GK# 847000
Vendor Credits:
No matching courses available.
Start learning as soon as today! Click Add To Cart to continue shopping or Buy Now to check out immediately.
Access Period:
Scheduling a custom training event for your team is fast and easy! Click here to get started.

What You'll Learn

  • Curate high-quality datasets and generate synthetic data using NVIDIA NeMo Curator.
  • Rigorously evaluate LLM performance with benchmarks (MMLU), LLM-as-a-judge, and the NeMo Evaluator.
  • Inject new domain-specific knowledge into LLMs using Continued Pretraining (CPT).
  • Teach LLMs new skills and align them to specific tasks with Supervised Fine-Tuning (SFT).
  • Align model behavior to human preferences for style, tone, and safety using Direct Preference Optimization (DPO).
  • Compress and optimize LLMs for efficient deployment using Quantization, Pruning, and Knowledge Distillation with TensorRT-LLM and NeMo.
  • Apply end-to-end model customization workflows to solve real-world problems.
  • Topics Covered

Prerequisites

  • Familiarity with Python programming and Jupyter notebooks.
  • Basic understanding of Large Language Models and their applications.
  • Conceptual knowledge of deep learning and neural networks.
  • Tools, libraries, frameworks used: Python, NVIDIA NeMo, NVIDIA TensorRT-LLM, Docker, MLflow