Jailbreak & Prompt Injection
Prevent your LLM from getting jailbroken using prompt injection and other latest threat vectors.
Protect against jailbreaks, hallucinations, and malicious inputs, while monitoring LLM behavior with detailed traces, spans, and metrics.
Check user inputs and LLM outputs with our built-in evaluators, or create your custom guardrails. We support text, conversations, RAG, and even tools.
Prevent your LLM from getting jailbroken using prompt injection and other latest threat vectors.
Detect political, legal, medical, or even religious content from being discussed.
Filter out offensive and threatening language as well as hate speech from your LLM.
Catch personally identifiable information from your LLM outputs, especially when using RAG.
Block competitors from being mentioned in your LLM outputs and avoid embarrassement.
Identify and grade the tone and mood of your LLM outputs to ensure they are appropriate.
Ensure the LLM output matches the user input to avoid language confusion.
Compare input and output embeddings to ensure the LLM is on topic and relevant.
Detect the emotions in the content and ensure they are appropriate.
Check how helpful the replies have been using an LLM-as-judge.
const modelmetry = new ModelmetryClient()
const guardrails = modelmetry.guardrails()
const result = await guardrails.check("grd_jaohsfzgcbd523hbt1grwmvp", {
Input: {
Text: "What does the employee handbook say about vacation time during a busy period?",
},
})
if (result.failed) {
// handle the failure
return "Sorry user, I cannot help you with this query at the moment."
for (const entry of result.summarisedEntries) {
// You can have access to more data for debugging (scores, evaluation(s) that failed) in the Check
console.log(entry)
}
}
// carry on as normal
Modelmetry provides lightening fast advanced guardrails to keep your company safe from LLM risks.
Check user inputs and LLM outputs with our wide range of evaluators. We offer a wide range of guardrails to help you deploy your models with confidence.
Our pricing is simple and transparent.
Have any questions? Use our live chat or email us at [email protected]
Modelmetry is an advanced platform designed to enhance the safety, quality, and appropriateness of data and models in applications utilizing Large Language Models (LLMs) like chatbots. It offers a comprehensive suite of evaluators to assess critical aspects such as emotion analysis, PII leak detection, text moderation, relevancy, and security threat detection.
With customizable guardrails, early termination options, and detailed metrics and scores, Modelmetry ensures that your LLM applications meet high standards of performance and safety. This robust framework provides actionable insights, safeguarding the integrity and effectiveness of your AI-driven solutions.
Modelmetry is ideal for developers and software engineers aiming to ensure their AI-driven applications are safe, reliable, and compliant with regulations.
Modelmetry also benefits higher-level stakeholders, including product managers, compliance officers, and CEOs, by offering a robust framework to monitor and enhance application performance and security, ensuring high standards of safety and quality while mitigating risks.
Absolutely, our client SDKs are open source. Our backend is proprietary because, well, it's our secret sauce. We can expert all your data upon request.
Modelmetry is committed to protecting your data privacy and security. We do not access payloads on your behalf, ever. We are a security-focused company and have implemented robust measures to ensure the confidentiality and integrity of your data. We use encryption, secure connections, and other industry-standard security practices to safeguard your data.
We do not access payloads on your behalf, ever. And we do store inputs and outputs so you can review them alongside metrics and scores.