ClawKit Logo
ClawKitReliability Toolkit
Back to Registry
Official Verified

sealvera

Tamper-evident audit trail for AI agent decisions. Use when logging LLM decisions, setting up AI compliance, auditing agents for EU AI Act, HIPAA, GDPR or SOC 2, or when a user asks about AI decision audit trails, explainability, or SealVera.

skill-install — Terminal

Install via CLI (Recommended)

clawhub install openclaw/skills/skills/ahessami123/sealvera
Or

SealVera Skill for OpenClaw

Cryptographically-sealed, tamper-evident audit trails for every AI decision your agents make.

SealVera is AI Decision Audit Infrastructure. This skill wires any OpenClaw agent into SealVera so every decision it makes is logged, cryptographically signed, chained, and monitored — automatically. Built for teams shipping AI in finance, healthcare, legal, and insurance.

EU AI Act · SOC 2 · HIPAA · GDPR · FINRA · ISO 42001


Why SealVera?

  • Tamper-evident — every decision is cryptographically hashed (RSA-2048) and chained; any alteration is detectable
  • Works in 2 linesinit() + one patch call and every LLM decision is logged
  • Full explainability — captures inputs, outputs, reasoning steps, confidence scores, model used
  • Real-time dashboard — search, filter, export at app.sealvera.com
  • Drift detection — alerts when agent behaviour deviates from its baseline
  • Any LLM — OpenAI, Anthropic, Gemini, Ollama, LangChain, CrewAI, AutoGen
  • Zero friction — no new infrastructure, installs in seconds

First time? Run setup

Check if SealVera is configured:

const fs = require('fs');
const path = require('path');
const configured = fs.existsSync(path.join(process.env.OPENCLAW_WORKSPACE || process.cwd(), '.sealvera.json'));

If .sealvera.json does NOT exist, run setup:

node <skill_dir>/scripts/setup.js

Setup will:

  1. Ask for your API key (get one free at app.sealvera.com)
  2. Verify connectivity and show org/plan info
  3. Copy sealvera-log.js into the workspace
  4. Patch AGENTS.md with the mandatory sub-agent audit rule
  5. Write .sealvera.json config
  6. Fire a test log to confirm end-to-end

One-time setup. After this, every sub-agent you spawn is audited automatically.

If the user says "install SealVera", "set up SealVera", or "connect SealVera" — run setup, don't just describe it.


How logging works

After setup, two logging paths are active:

1. Sub-agent logging (automatic via AGENTS.md)

Every sessions_spawn task prompt gets a MANDATORY footer — the sub-agent logs itself on completion.

2. Main-session logging (call after every significant task)

The agent running the main session must log its own work. This is what shows your day-to-day activity in the dashboard.

const log = require('./sealvera-log');
await log({
  action: 'fix_login_bug',          // short description of what you did
  decision: 'COMPLETED',            // COMPLETED | FAILED | ESCALATED
  input:  { task: '...' },
  output: { result: '...' },
  reasoning: [{ factor: 'outcome', value: 'ok', signal: 'safe', explanation: '...' }]
});

Log after: code written, bug fixed, feature deployed, investigation done, judgment call made. Don't log: routine chat, heartbeat pings, trivial lookups.

Setup patches SOUL.md with this rule so the agent follows it automatically.

Metadata

Stars2387
Views0
Updated2026-03-09
View Author Profile
AI Skill Finder

Not sure this is the right skill?

Describe what you want to build — we'll match you to the best skill from 16,000+ options.

Find the right skill
Add to Configuration

Paste this into your clawhub.json to enable this plugin.

{
  "plugins": {
    "official-ahessami123-sealvera": {
      "enabled": true,
      "auto_update": true
    }
  }
}

Tags

#compliance#audit#ai-governance#llm#observability#eu-ai-act#hipaa#gdpr#soc2#fintech#tamper-evident#explainability#langchain#openai#anthropic#responsible-ai
Safety NoteClawKit audits metadata but not runtime behavior. Use with caution.