NVIDIA launches safety toolkit to keep AI chatbots more secure

NVIDIA NeMo Guardrails announced to make generative AI tools safer
Tecno Ad
Tecno Ad
Tecno Ad
Tecno Ad

NVIDIA just released an open-source toolkit called “NeMo Guardrails” to help generative AI programs like ChatGPT stay on track and answer queries securely. NVIDIA’s Senior Director for AI Software, Jonathan Cohen, says this is a move to make text-generating AI tools and chatbots safer and non-toxic.

NVIDIA NeMo Guardrails and AI:

According to the company, this “AI safety toolkit” has easy-to-implement features that anyone can use to guide their chatbots to act professionally. Since Large Language Models (LLMs) are prone to bias and “hallucinations”, they aren’t always ideal. Sometimes they might even become toxic or manipulative.

As a result, the need for a mechanism that can guide chat interactions as intended has been emerging in this space for quite some time. 

Companies like OpenAI use multiple layers of filters and human moderators to prevent unintended responses. But smaller organizations might not be able to do the same. It’s getting increasingly easier to create an LLM that is trained on your very own dataset and answers queries on that domain.

So the safety concern may not be the top priority—or rather out of technical reach—for small businesses. They might also choose to implement their own safety protocols that are either too broad and vague or too uptight and narrow for their use case. Hence, an open-source tool like NVIDIA NeMo Guardrails would make it accessible to everyone.

What is NeMo Guardrails?

NVIDIA announces NeMo to make AI like ChatGPT safer

Basically, NeMo Guardrails contains code constraint examples and documentation using which developers can easily guide their chatbots to behave in the intended manner. “These guardrails monitor, affect, and dictate a user’s interactions, like guardrails on a highway that define the width of a road and keep vehicles from veering off into unwanted territory”, says NVIDIA.

What can NeMo Guardrails do?

When implemented, it can have noticeable impacts on the user experience with text-based AI tools. As of now, NeMo Guardrails support these 3 categories of boundaries:

  • Topical guardrails, to prevent the bot from veering out of topic and going rogue.
  • Safety guardrails, to prevent misinformation, toxic responses, and biases.
  • Security guardrails, to prevent chatbots from executing malicious code and scripts.

NVIDIA NeMo Guardrails: Conclusion

NVIDIA says they have been working on NeMo Guardrails for years and acknowledge that its model isn’t perfect. That being said, this is definitely a step in the right direction.

A lot of businesses are looking to capitalize on the superhuman capabilities of text-generating chatbots and LLMs. This enables the need for a lot of these businesses to implement their bots safely. Hence, NVIDIA’s move toward AI safety is a huge stepping stone for the future.