ad
      ad

      NVIDIA launches safety toolkit to keep AI chatbots more secure

      GG

      By

      Published Apr 27, 2023

      ad
      NVIDIA launches safety toolkit to keep AI chatbots more secure
      ad

      No headings found

      NVIDIA just released an open-source toolkit called "NeMo Guardrails" to help generative AI programs like ChatGPT stay on track and answer queries securely. NVIDIA's Senior Director for AI Software, Jonathan Cohen, says this is a move to make text-generating AI tools and chatbots safer and non-toxic.

      NVIDIA NeMo Guardrails and AI:

      According to the company, this "AI safety toolkit" has easy-to-implement features that anyone can use to guide their chatbots to act professionally. Since Large Language Models (LLMs) are prone to bias and "hallucinations", they aren’t always ideal. Sometimes they might even become toxic or manipulative. As a result, the need for a mechanism that can guide chat interactions as intended has been emerging in this space for quite some time.  Companies like OpenAI use multiple layers of filters and human moderators to prevent unintended responses. But smaller organizations might not be able to do the same. It’s getting increasingly easier to create an LLM that is trained on your very own dataset and answers queries on that domain. So the safety concern may not be the top priority—or rather out of technical reach—for small businesses. They might also choose to implement their own safety protocols that are either too broad and vague or too uptight and narrow for their use case. Hence, an open-source tool like NVIDIA NeMo Guardrails would make it accessible to everyone. What is NeMo Guardrails?
      NVIDIA announces NeMo to make AI like ChatGPT safer
      Basically, NeMo Guardrails contains code constraint examples and documentation using which developers can easily guide their chatbots to behave in the intended manner. “These guardrails monitor, affect, and dictate a user’s interactions, like guardrails on a highway that define the width of a road and keep vehicles from veering off into unwanted territory”, says NVIDIA.

      What can NeMo Guardrails do?

      When implemented, it can have noticeable impacts on the user experience with text-based AI tools. As of now, NeMo Guardrails support these 3 categories of boundaries:
      • Topical guardrails, to prevent the bot from veering out of topic and going rogue.
      • Safety guardrails, to prevent misinformation, toxic responses, and biases.
      • Security guardrails, to prevent chatbots from executing malicious code and scripts.

      NVIDIA NeMo Guardrails: Conclusion

      NVIDIA says they have been working on NeMo Guardrails for years and acknowledge that its model isn’t perfect. That being said, this is definitely a step in the right direction. A lot of businesses are looking to capitalize on the superhuman capabilities of text-generating chatbots and LLMs. This enables the need for a lot of these businesses to implement their bots safely. Hence, NVIDIA's move toward AI safety is a huge stepping stone for the future.

      Article Last updated: April 27, 2023

      Best Tech Deals

      No Active Polls

      There are currently no polls available. Check back later for new polls to participate in!

      Polls will appear here when available

      How did we do with this article?

      Conversation

      We’d love to hear your thoughts! Let's keep it respectful and on-topic. Any inappropriate remarks may be removed. Happy commenting! Privacy Policy

      Be the first to share your thoughts-start the conversation!

      Latest Articles