ai experimental version 0.1.0 hex unreleased package jido_eval elixir ~> 1.18

Jido Eval

Evaluation framework for LLM and Jido agent quality measurement

Experimental support. Open exploration with no stability guarantee.

View package metadata source →

RELATED PACKAGES

AT A GLANCE

Framework foundation for evaluating LLM/agent behavior
Structured project layout for datasets, scoring, and pipelines
Typed schemas and CSV utilities for experiment inputs/outputs
Quality automation aliases for repeatable evaluation workflows

DEEP DIVE

Overview

Jido Eval is an experimental framework focused on measuring and improving LLM and agent performance in Jido-based systems.

Purpose

Jido Eval provides an ecosystem home for evaluation datasets, scoring methods, and repeatable quality checks.

Major Components

Evaluation Core

Defines project primitives for assembling benchmark runs and collecting outputs.

Data Utilities

Uses typed structs and CSV tooling for controlled experiment datasets.

Quality Tooling

Ships project aliases oriented around formatting, static analysis, and docs-driven iteration.