AI Advances

Democratizing access to artificial intelligence

Follow publication

Member-only story

LLM Safeguard | NeMo Guardrails Tutorials | 01 Hello NeMo

01coder
AI Advances
Published in
5 min readFeb 17, 2024

NeMo Guardrails Tutorial — 01 Hello NeMo

Safeguarding LLM is an area we all know deserves significant attention.

Starting today, I will be launching a new course series — NeMo Guardrails Tutorial. Nemo Guardrails is an open-source toolkit released by Nvidia, providing safeguard solutions for LLM applications.

The course is open-sourced at GitHub. Feel free to follow and feedback.

Today marks the first installment of this course series, 01 Hello NeMo. We will learn what NeMo Guardrails is all about and take a quick look into its basic usage.

Organizations, regardless of size, that are eager to adopt generative AI face significant challenges in protecting their LLM applications. Addressing issues such as prompt injection, handling insecure outputs, and preventing the leakage of sensitive information are questions that every AI architect or engineer must answer. Without reliable solutions to these issues, enterprise-grade LLM applications cannot survive.

Create an account to read the full story.

The author made this story available to Medium members only.
If you’re new to Medium, create a new account to read this story on us.

Or, continue in mobile web

Already have an account? Sign in

No responses yet

Write a response