Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run a first technical test #102

Open
anneschuth opened this issue Jul 17, 2024 · 5 comments
Open

Run a first technical test #102

anneschuth opened this issue Jul 17, 2024 · 5 comments

Comments

@anneschuth
Copy link
Member

anneschuth commented Jul 17, 2024

For now we assume a model to be present and data to be present and it runs locally, and that it is linked from the system card.

  • the system card points to the model in the model card which has a URI to a model.
  • The URI (for now) is a file path (file://) to the guusje model.
  • We add a simple example model (as a fixture), simply checked into our code base (use model from https://github.com/MinBZK/example-data-science-project)
  • The path to the model is hardcoded (for now)
  • The path for the data is hardcoded (for now)

We add a Shap technical test to the Instrument Registry. It needs

  • A urn
  • A description
  • A task of type "technical test"
  • The task should require a model by pointing to the model field included in the system card.
  • The task has a urn (used below)
  • To define where the results should be written in the system card (in measures)

Implement the technical test that runs the Shap implementation of AI Verify using the model pointed to in the system card. We implement this as a CLI (for now, later this becomes a Celery worker), based on the minimimal ai verify repo (https://github.com/MinBZK/aiverify/tree/using_aiverify_as_api).

The implementation:

  • itself knows what instrument task it implements (it knows the urn of the Shap technical test task, see above)
  • has access to the instrument definition (by retrieving it using the urn)
  • has access to a system card (receives the path as a CLI argument)
  • calls the AI Verify API
  • writes the result from the SHAP test into the model card on the path defined by the instrument.

Acceptance Criteria

  • All shap values are stored in the system card -> model card -> model index -> results -> measures -> ...
  • implement how to store shap values in model card without lose of data
@anneschuth anneschuth converted this from a draft issue Jul 17, 2024
@anneschuth anneschuth changed the title Run technical test Run a first technical test Jul 17, 2024
Copy link
Contributor

This issue did not have any activity in the last 90 days and will be removed after 30 days

@github-actions github-actions bot added Stale and removed Stale labels Oct 24, 2024
Copy link
Contributor

This issue did not have any activity in the last 90 days and will be removed after 30 days

@github-actions github-actions bot added the Stale label Jan 23, 2025
Copy link
Contributor

This issue is closed due to inactivity

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Feb 23, 2025
@github-project-automation github-project-automation bot moved this from ♻ To Do to ✅ Done in 👾 AI Validation Team Planning Feb 23, 2025
@anneschuth
Copy link
Member Author

:(

@robbertbos robbertbos removed the Stale label Feb 24, 2025
@robbertbos robbertbos moved this from ✅ Done to ♻ To Do in 👾 AI Validation Team Planning Feb 24, 2025
@robbertbos robbertbos reopened this Feb 24, 2025
@anneschuth
Copy link
Member Author

❤️

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: ♻ To Do
Development

No branches or pull requests

2 participants