Skip to content

Agents & HR

Agent Identity Card

Every agent has a comprehensive identity. At the design level, agent data splits into two layers:

Config (immutable)
Identity, personality, skills, model preferences, tool permissions, and authority. Defined at hire time, changed only by explicit reconfiguration. Represented as frozen Pydantic models.
Runtime state (mutable-via-copy)
Current status, active task, conversation history, and execution metrics. Evolves during agent operation. Represented as Pydantic models using model_copy(update=...) for state transitions -- never mutated in place.

Personality Dimensions

Personality is split into two tiers:

Float values (0.0--1.0) used for internal compatibility scoring only (not injected into prompts). stress_response replaces traditional neuroticism with inverted polarity (1.0 = very calm). Scored by core/personality.py.

Dimension Range Description
openness 0.0--1.0 Curiosity, creativity
conscientiousness 0.0--1.0 Thoroughness, reliability
extraversion 0.0--1.0 Assertiveness, sociability
agreeableness 0.0--1.0 Cooperation, empathy
stress_response 0.0--1.0 Emotional stability (1.0 = very calm)

Compatibility scoring (weighted composite, result clamped to [0, 1]):

  • 60% Big Five similarity: openness, conscientiousness, agreeableness, stress_response use 1 - |diff|; extraversion uses a tent-function peaking at 0.3 diff (complementary extraverts collaborate better than identical ones)
  • 20% Collaboration alignment: ordinal adjacency (INDEPENDENTPAIRTEAM); scored 1.0 for same, 0.5 for adjacent, 0.0 for opposite
  • 20% Conflict approach: constructive pairs score 1.0, destructive pairs 0.2, mixed 0.4--0.6. Uses itertools.combinations for team-level averaging

Injected into system prompts as natural-language labels that LLMs respond to:

Enum Values
DecisionMakingStyle analytical, intuitive, consultative, directive
CollaborationPreference independent, pair, team
CommunicationVerbosity terse, balanced, verbose
ConflictApproach avoid, accommodate, compete, compromise, collaborate (Thomas-Kilmann model)

Agent Configuration Example

Full agent identity YAML
# --- Config layer -- AgentIdentity (frozen) ---
agent:
  id: "uuid"
  name: "Sarah Chen"
  role: "Senior Backend Developer"
  department: "Engineering"
  level: "Senior"
  personality:
    traits:
      - analytical
      - detail-oriented
      - pragmatic
    communication_style: "concise and technical"
    risk_tolerance: "low"
    creativity: "medium"
    description: >
      Sarah is a methodical backend developer who prioritizes clean
      architecture and thorough testing. She pushes back on shortcuts
      and advocates for proper error handling. Prefers Pythonic solutions.
    # Big Five (OCEAN-variant) -- internal scoring (0.0-1.0)
    openness: 0.4
    conscientiousness: 0.9
    extraversion: 0.3
    agreeableness: 0.5
    stress_response: 0.75
    # Behavioral enums -- injected into system prompts
    decision_making: "analytical"
    collaboration: "independent"
    verbosity: "balanced"
    conflict_approach: "compromise"
  skills:
    primary:
      - python
      - litestar
      - postgresql
      - system-design
    secondary:
      - docker
      - redis
      - testing
  model:
    provider: "example-provider"
    model_id: "example-medium-001"
    temperature: 0.3
    max_tokens: 8192
    fallback_model: "openrouter/example-medium-001"
    model_tier: "medium"  # set by model matcher (large/medium/small/null)
  model_requirement:            # original tier/priority from template
    tier: "medium"
    priority: "balanced"
    min_context: 0
    capabilities: []
  memory:
    type: "persistent"       # persistent, project, session, none
    retention_days: null     # null = forever; also agent-level global default
    retention_overrides: []  # per-category overrides, e.g. [{category: "semantic", retention_days: 365}]
  tools:
    access_level: "standard" # sandboxed | restricted | standard | elevated | custom
    allowed:
      - file_system
      - git
      - code_execution
      - web_search
      - terminal
    denied:
      - deployment
      - database_admin
  authority:
    can_approve: ["junior_dev_tasks", "code_reviews"]
    reports_to: "engineering_lead"
    can_delegate_to: ["junior_developers"]
    budget_limit: 5.00
  autonomy_level: null       # full, semi, supervised, locked (overrides defaults)
  hiring_date: "2026-02-27"
  status: "active"           # active, on_leave, terminated

Runtime State

The runtime state layer (in engine/) tracks execution progress using frozen models with model_copy:

  • TaskExecution wraps a Task with evolving execution state: status transitions, accumulated cost (TokenUsage), turn count, and timestamps.
  • AgentContext wraps AgentIdentity + TaskExecution with a unique execution ID, conversation history, cost accumulation, turn limits, and timing.
  • AgentRuntimeState provides a lightweight per-agent execution status snapshot (idle / executing / paused) for dashboard queries and graceful-shutdown discovery. Persisted via AgentStateRepository, independent of the checkpoint system.

Seniority & Authority Levels

Level Authority Typical Model Cost Tier
Intern/Junior Execute assigned tasks only small / local $
Mid Execute + suggest improvements medium / local $$
Senior Execute + design + review others medium / large $$$
Lead All above + approve + delegate large / medium $$$
Principal/Staff All above + architectural decisions large $$$$
Director Strategic decisions + budget authority large $$$$
VP Department-wide authority large $$$$
C-Suite (CEO/CTO/CFO) Company-wide authority + final approvals large $$$$

Role Catalog

The role catalog is extensible -- users can add custom roles via config. The built-in catalog covers common organizational roles:

  • CEO -- Overall strategy, final decision authority, cross-department coordination
  • CTO -- Technical vision, architecture decisions, technology choices
  • CFO -- Budget management, cost optimization, resource allocation
  • COO -- Operations, process optimization, workflow management
  • CPO -- Product strategy, roadmap, feature prioritization
  • Product Manager -- Requirements, user stories, prioritization, stakeholder communication
  • UX Designer -- User research, wireframes, user flows, usability
  • UI Designer -- Visual design, component design, design systems
  • UX Researcher -- User interviews, analytics, A/B test design
  • Technical Writer -- Documentation, API docs, user guides
  • Software Architect -- System design, technology decisions, patterns
  • Frontend Developer (Junior/Mid/Senior) -- UI implementation, components, state management
  • Backend Developer (Junior/Mid/Senior) -- APIs, business logic, databases
  • Full-Stack Developer (Junior/Mid/Senior) -- End-to-end implementation
  • DevOps/SRE Engineer -- Infrastructure, CI/CD, monitoring, deployment
  • Database Engineer -- Schema design, query optimization, migrations
  • Security Engineer -- Security audits, vulnerability assessment, secure coding
  • QA Lead -- Test strategy, quality gates, release readiness
  • QA Engineer -- Test plans, manual testing, bug reporting
  • Automation Engineer -- Test frameworks, CI integration, E2E tests
  • Performance Engineer -- Load testing, profiling, optimization
  • Data Analyst -- Metrics, dashboards, business intelligence
  • Data Engineer -- Pipelines, ETL, data infrastructure
  • ML Engineer -- Model training, inference, MLOps
  • Project Manager -- Timelines, dependencies, risk management, status tracking
  • Scrum Master -- Agile ceremonies, impediment removal, team health
  • HR Manager -- Hiring recommendations, team composition, performance tracking
  • Security Operations -- Request validation, safety checks, approval workflows
  • Content Writer -- Blog posts, marketing copy, social media
  • Brand Strategist -- Messaging, positioning, competitive analysis
  • Growth Marketer -- Campaigns, analytics, conversion optimization

Dynamic Roles

Users can define custom roles via config:

custom_roles:
  - name: "Blockchain Developer"
    department: "Engineering"
    skills: ["solidity", "web3", "smart-contracts"]
    system_prompt_template: "blockchain_dev.md"
    authority_level: "senior"
    suggested_model: "large"

Hiring Process

The HR system manages the agent workforce dynamically:

  1. HR agent (or human) identifies a skill gap or workload issue
  2. HR generates candidate cards based on team needs:
    • What skills are underrepresented?
    • What seniority level is needed?
    • What personality would complement the team?
    • What model/provider fits the budget?
  3. Candidate cards are presented for approval (to CEO or human)
  4. Approved candidates are instantiated and onboarded
  5. Onboarding includes: company context, project briefing, team introductions

Design decisions (Decision Log D8)

  • D8.1 -- Source: Templates + LLM customization. Templates for common roles (reuses existing template system). LLM generates config for novel roles not covered by templates. Approval gate catches invalid/bad configs before instantiation.
  • D8.2 -- Persistence: Operational store via PersistenceBackend. YAML stays as bootstrap seed -- operational store wins for runtime state. Enables rehiring and auditable history.
  • D8.3 -- Hot-plug: Agents are hot-pluggable at runtime via a dedicated company/registry service (not AgentEngine, which remains the per-agent task runner). Thread-safe registry, wired into message bus + tools + budget.

Firing / Offboarding

Offboarding is triggered by: budget cuts, poor performance metrics, project completion, or human decision.

  1. Agent's memory is archived (not deleted)
  2. Active tasks are reassigned
  3. Team is notified

Design decisions (Decision Log D9, D10)

  • D9 -- Task Reassignment: Pluggable TaskReassignmentStrategy protocol. Initial strategy: queue-return -- tasks return to unassigned queue, existing TaskRoutingService re-routes with priority boost for reassigned tasks. Future strategies: same-department/lowest-load, manager-decides (LLM), HR agent decides.
  • D10 -- Memory Archival: Pluggable MemoryArchivalStrategy protocol. Initial strategy: full snapshot, read-only. Pipeline: retrieve all memories, archive to ArchivalStore, selectively promote semantic+procedural memories to OrgMemoryBackend (rule-based), clean hot store, mark agent TERMINATED. Rehiring restores archived memories into a new AgentIdentity. Future strategies: selective discard, full-accessible.

Performance Tracking

Performance data is exposed via three API sub-routes on /api/v1/agents/{name}:

Sub-route Response model Description
GET /performance AgentPerformanceSummary Flat summary: tasks completed (total/7d/30d), success rate, cost per task, quality/collaboration scores, trend direction, plus raw window metrics and trend results
GET /activity PaginatedResponse[ActivityEvent] Paginated chronological timeline merging lifecycle events, task metrics, cost records, tool invocations, and delegation records (most recent first). Supports typed ActivityEventType enum filtering (invalid values return 400). Cost events are redacted for read-only roles. Response includes degraded_sources field for partial data detection
GET /history ApiResponse[tuple[CareerEvent, ...]] Career-relevant lifecycle events (hired, fired, promoted, demoted, onboarded) in chronological order

The framework tracks detailed per-agent metrics:

agent_metrics:
  tasks_completed: 42
  tasks_failed: 2
  average_quality_score: 8.5     # from code reviews, peer feedback
  average_cost_per_task: 0.45
  average_completion_time: "2h"
  collaboration_score: 7.8       # peer ratings
  last_review_date: "2026-02-20"
Design decisions (Decision Log D2, D3, D11, D12)

D2 -- Quality Scoring: Pluggable QualityScoringStrategy protocol. Initial strategy: layered combination --

  1. FREE: Objective CI signals (test pass/fail, lint, coverage delta)
  2. ~$1/day: Small-model LLM judge (different family than agent) evaluates output vs acceptance criteria
  3. On-demand: Human override via API, highest weight

All three layers are implemented via CompositeQualityStrategy (configurable CI/LLM weights, human override short-circuits with highest priority). Human override CRUD is exposed at /agents/{agent_id}/quality/override. Config fields: quality_judge_model, quality_judge_provider, quality_ci_weight, quality_llm_weight in PerformanceConfig. Future strategies: CI-only, LLM-only, human-only.


D3 -- Collaboration Scoring: Pluggable CollaborationScoringStrategy protocol. Initial strategy: automated behavioral telemetry --

collaboration_score = weighted_average(
    delegation_success_rate,
    delegation_response_latency,
    conflict_resolution_constructiveness,
    meeting_contribution_rate,
    loop_prevention_score,
    handoff_completeness
)

Weights are configurable per-role. Periodic LLM sampling (1%, configurable) for calibration is implemented via LlmCalibrationSampler (opt-in, requires llm_sampling_model config). Human override via API is implemented via CollaborationOverrideStore + CollaborationController at /agents/{agent_id}/collaboration. Future strategies: LLM evaluation, peer ratings, human-provided.


D11 -- Rolling Windows: Pluggable MetricsWindowStrategy protocol. Initial strategy: multiple simultaneous windows --

  • 7d for acute regressions
  • 30d for sustained patterns
  • 90d for baseline/drift

Minimum 5 data points per window; below that, the system reports "insufficient data." Future strategies: fixed single window, per-metric configurable.


D12 -- Trend Detection: Pluggable TrendDetectionStrategy protocol. Initial strategy: Theil-Sen regression slope per window + configurable thresholds classify trends as improving/stable/declining. Theil-Sen has 29.3% outlier breakdown (tolerates ~1 in 3 bad data points). Minimum 5 data points. Future strategies: period-over-period, OLS regression, threshold-only.


Promotions & Demotions

Agents can move between seniority levels based on performance:

  • Promotion criteria: Sustained high quality scores, task complexity handled, peer feedback
  • Demotion criteria: Repeated failures, quality drops, cost inefficiency
  • Promotions can unlock higher tool access levels
  • Model upgrades/downgrades may accompany level changes (configurable, see auto-downgrade)

Design decisions (Decision Log D13, D14, D15)

  • D13 -- Promotion Criteria: Pluggable PromotionCriteriaStrategy protocol. Initial strategy: configurable threshold gates. ThresholdEvaluator with min_criteria_met: int (N of M) + required_criteria: list[str]. Setting min=total gives AND; min=1 gives OR. Default: junior-to-mid = 2 of 3 criteria, mid-to-senior = all.
  • D14 -- Promotion Approval: Pluggable PromotionApprovalStrategy protocol. Initial strategy: senior+ requires human approval. Junior-to-mid auto-promotes (low cost impact: small-to-medium ~4x). Demotions: auto-apply for cost-saving (model downgrade), human approval for authority-reducing demotions.
  • D15 -- Model Mapping: Pluggable ModelMappingStrategy protocol. Initial strategy: default ON (hr.promotions.model_follows_seniority: true). Model changes at task boundaries only (never mid-execution, consistent with auto-downgrade). Per-agent preferred_model overrides seniority default. Smart routing still uses cheap models for simple tasks regardless of seniority.

Five-Pillar Evaluation Framework

Performance data is also evaluated through a structured five-pillar framework (InfoQ: Evaluating AI Agents):

Pillar Measures Data Sources
Intelligence/Accuracy Quality of task output, reasoning coherence QualityScoreResult, LlmCalibrationRecord
Performance/Efficiency Cost, latency, token usage WindowMetrics (cost, time, tokens)
Reliability/Resilience Consistency, failure recovery, streaks TaskMetricRecord sequences
Responsibility/Governance Compliance, trust stability, autonomy adherence Audit log, trust system, autonomy system
User Experience Clarity, helpfulness, tone, satisfaction InteractionFeedback records

Each pillar and its individual metrics can be independently enabled/disabled via EvaluationConfig. Disabled pillars/metrics have their weight redistributed proportionally to remaining enabled ones. All pillars ship enabled by default with recommended weights (equal 0.2 each).

The EvaluationService orchestrates scoring, delegating to pluggable PillarScoringStrategy implementations. The efficiency pillar is computed inline from WindowMetrics. Human-calibrated LLM labeling uses the existing LlmCalibrationSampler infrastructure -- calibration drift above a configurable threshold reduces the intelligence pillar's confidence, signaling the need for more human labels.

Design decisions (Decision Log D24)

D24 -- Five-Pillar Evaluation: Pluggable PillarScoringStrategy protocol with single EvaluationContext bag. Each pillar has a default strategy:

  • Intelligence: QualityBlendIntelligenceStrategy -- blends CI quality score (70%) with LLM calibration score (30%). High calibration drift reduces confidence.
  • Efficiency: Inline computation from WindowMetrics -- normalized cost (40%), time (30%), token (30%) efficiency scores.
  • Resilience: TaskBasedResilienceStrategy -- success rate (40%), recovery rate (25%), quality consistency (20%), streak bonus (15%).
  • Governance: AuditBasedGovernanceStrategy -- audit compliance (50%), trust level (30%), autonomy compliance (20%).
  • Experience: FeedbackBasedUxStrategy -- clarity (25%), helpfulness (25%), trust (20%), tone (15%), satisfaction (15%).

All metrics toggleable via EvaluationConfig per-pillar sub-configs. Weight redistribution follows the BehavioralTelemetryStrategy pattern. Pull-based evaluation (no background daemon). Settings system integration deferred -- config model is ready for it.