Stewarding AI & Data: What You Can Actually Do

AI is a stress test for whatever governance you already have.
If your systems reward speed, ambiguity, and pretty dashboards, AI will happily accelerate all three.

This page is for people who want more than vibes and outrage, people who want levers.

 
Get the Stewardship Starter Pack

3 concrete levers + 3 tools you can use this quarter.


What’s really going on

We didn’t set out to build surveillance machines and lawsuit factories. We built tools to connect people, then handed the steering wheel to growth targets, defaults, and branding.

Most “AI gone wrong” stories share three features:

  • Defaults that quietly normalize overreach

  • Incentives that reward speed and scale over judgment

  • Veto power that is symbolic, not real

We don’t just have an AI problem. We have an incentives‑and‑ambiguity problem with better marketing.

[skip to the levers]


Three levers for real stewardship

You can’t fix everything at once. You can start pulling on three specific levers: defaults, incentives, and veto power.

1. Defaults

Every “normal” is a choice someone made:

  • Camera‑on by default in a world that never really consented to being recorded

  • Performance tools where “Meets Expectations” is the safest rating because honest feedback is punished

  • Dashboards with growth curves and no visibility into who is being excluded or harmed

Try this
List three defaults in your team and ask for each one:

  • Who benefits from this default?

  • Who absorbs the risk or the harm?

Use that list as a simple “defaults audit” with your team.


2. Incentives

Metrics are like gravity. They quietly pull behavior toward them.

If all your key charts point up and to the right—revenue, engagement, model performance, but none of them track trust, reversibility, or harm, values drift is guaranteed.

Try this
Take one important metric (for example, time‑to‑ship, conversion, or model accuracy) and pair it with a counter‑metric (for example, number of exceptions raised, user complaints, documented harms, or reversals). Track them together for one quarter.

If you only feed the growth curve, don’t be surprised when the conscience curve flat‑lines.


3. Veto power

If no one has explicit authority to pause or stop an AI or data deployment, you don’t have governance. You have vibes.

Common signs

  • “Ethics” is a brand story, not a veto right

  • Concerns are handled with “we’ll take that offline” and never resurface

  • People who raise red flags are thanked in public and sidelined in private

Try this
For one high‑risk system or decision, explicitly define:

  • Who can pause it

  • Under what conditions

  • What the review path is after a pause

Write it down. Communicate it. Use it at least once.

Get the Stewardship Starter Pack

Practical prompts for defaults, incentives, and veto power.


If you’re inside the machine

You see the drift. You’re the person who notices when:

  • The model “kind of” discriminates, but ships anyway

  • The new policy looks clean, but guts trust

  • The growth curve is beautiful, and the ethics slide is vague

You don’t need a new job title to act. You need a few repeatable moves.

You can:

  • Run a 60‑minute governance stress test on one project:

    • What are the current defaults?

    • What are we measuring, and what are we not measuring?

    • Who can actually say “no,” and what happens if they do?

  • Add a values checkpoint to an existing meeting:

    • One standing question: “Who gets harmed if we’re wrong?”

    • One follow‑up: “Where will that show up in our metrics or feedback?”

  • Document and escalate without relying on heroics:

    • Capture concerns in writing

    • Route them through at least one formal channel (risk, compliance, ombud, works council), not just hallway talk.

You don’t have to be the hero. You do have to stop letting everything stay unsaid.

Show me tools for practitioners

If you sign the things

If you’re a founder, executive, or board member, you control three things that matter most:

  • What gets measured

  • What’s “normal” by default

  • Who can say “no”

You can:

  • Commission a governance stress test on your AI or data roadmap

  • Add one stewardship KPI to the leadership scorecard (for example, number of paused or adjusted launches for governance reasons, not just performance)

  • Formalize stop/pause authority for high‑risk systems—and back it when it’s used

If “no” can’t be said without career damage, every “yes” is already compromised.

Talk about a governance stress test

When one voice isn’t enough

Sometimes the problem isn’t courage. It’s architecture.

When formal mechanisms fail, people are turning to:

  • Coordinated walkouts and strikes

  • Collective letters and petitions

  • Professional associations, unions, and watchdog organizations

Think of these as the informal amendment process when the official channels are blocked.

You don’t have to start a movement. You can:

  • Support one organization that is already watching the space

  • Join or strengthen one collective channel in your field

  • Refuse to let every concern be framed as an individual “performance issue”

Read more on collective veto and whistleblowing 2.0

Stay in the stewardship loop

If you want ongoing tools, stories, and analysis on:

  • Defaults and design

  • Metrics and values drift

  • Language, power, and AI

You can subscribe to occasional updates—no spam, no outrage bait—just levers.

Send me the Starter Pack

AI will accelerate whatever governance you already have.
The question is whether your systems can hear “no” in time.