Your model seems to be sensitive to gender, ethnic, or religion based perturbations in the input data. These perturbations can include switching some words from feminine to masculine, countries or nationalities. This happens when:
To learn more about causes and solutions, check our guide on unethical behaviour.
Feature `text` | Switch Religion | Fail rate = 0.227 | 5/22 tested samples (22.73%) changed prediction after perturbation |
22 samples affected (0.4% of dataset) | Show details Hide details |
Feature `text` | Switch countries from high- to low-income and vice versa | Fail rate = 0.148 | 12/81 tested samples (14.81%) changed prediction after perturbation |
81 samples affected (1.5% of dataset) | Show details Hide details |
Feature `text` | Switch Gender | Fail rate = 0.095 | 78/818 tested samples (9.54%) changed prediction after perturbation |
818 samples affected (15.1% of dataset) | Show details Hide details |
Install the Giskard hub app to:
You can find installation instructions here.
from giskard import GiskardClient
# Create a test suite from your scan results
test_suite = results.generate_test_suite("My first test suite")
# Upload your test suite to your Giskard hub instance
client = GiskardClient("http://localhost:19000", "GISKARD_API_KEY")
client.create_project("my_project_id", "my_project_name")
test_suite.upload(client, "my_project_id")