guardrails-ai@0.5.0a8 vulnerabilities

Adding guardrails to large language models.

  • latest version

    0.6.5

  • latest non vulnerable version

  • first published

    2 years ago

  • latest version published

    24 days ago

  • licenses detected

  • Direct Vulnerabilities

    Known vulnerabilities in the guardrails-ai package. This does not include vulnerabilities belonging to this package’s dependencies.

    How to fix?

    Automatically find and fix vulnerabilities affecting your projects. Snyk scans for vulnerabilities and provides fixes for free.

    Fix for free
    VulnerabilityVulnerable Version
    • H
    Improper Neutralization of Directives in Dynamically Evaluated Code ('Eval Injection')

    guardrails-ai is an Adding guardrails to large language models.

    Affected versions of this package are vulnerable to Improper Neutralization of Directives in Dynamically Evaluated Code ('Eval Injection') due to improper validation in the parse_token method of the ValidatorsAttr class in the guardrails/guardrails/validatorsattr.py file. An attacker can execute arbitrary code on the user's machine by loading a maliciously crafted XML file that contains Python code, which is then passed to an eval function.

    How to fix Improper Neutralization of Directives in Dynamically Evaluated Code ('Eval Injection')?

    Upgrade guardrails-ai to version 0.5.10 or higher.

    [0.2.9,0.5.10)
    • H
    XML External Entity (XXE) Injection

    guardrails-ai is an Adding guardrails to large language models.

    Affected versions of this package are vulnerable to XML External Entity (XXE) Injection when consuming RAIL documents from external sources. An attacker can leak internal file data by exploiting the SYSTEM entity in the XML structure.

    How to fix XML External Entity (XXE) Injection?

    Upgrade guardrails-ai to version 0.5.0 or higher.

    [,0.5.0)