I spent the next morning trying to quantify it.
Not the consequences. Those were already quantified. The system loved quantifying consequences—severity ratings, incident classifications, neat little rows in a ledger I'd only just discovered existed.
No, I was trying to quantify the guilt.
There had to be a pattern. A formula. Some way to measure how bad I should feel about Maya's knee, or the twelve other pending disasters, or the fact that I'd apparently been racking up collateral damage without even knowing it.
I pulled up the system interface during my first class.
Professor Went was lecturing about supply chains or market forces or something equally distant from my actual problems. I sat in the back and scrolled through tabs.
TRAITS
INTERACTIONS
EVOLUTION
CONSEQUENCES
HARM LEDGER
I checked each one.
Nowhere—not in the stats breakdown, not in the trait descriptions, not in the consequence severity scale—was there a metric for guilt.
No Remorse Index. No Moral Weight stat.
The system tracked everything else. It knew how many times I'd hesitated before making a decision. It logged proximity, intent classification, interaction duration down to the second.
But guilt?
That wasn't a variable.
I closed the interface and tried to pay attention to the lecture.
Lasted about three minutes.
Then I opened it again and started checking trait synergies, like maybe I'd missed something. Maybe guilt was bundled into one of the hidden modifiers. Maybe it showed up once you hit a certain threshold of collateral damage.
Nothing.
"You're doing it again."
I looked up.
Claire was sitting two seats over. I hadn't even noticed her come in.
"Doing what?" I asked.
"That thing where you stare at nothing and expect the universe to explain itself."
She'd moved closer now, sliding into the seat next to mine. Her notebook was open, but she clearly hadn't been taking notes.
"I'm not staring at nothing," I said.
"Right. You're staring at your weird system interface that no one else can see." She said it quietly, but there was an edge to it. "How's that going?"
"Fine."
"You're a terrible liar."
Professor Went's voice droned on in the background. Something about logistics bottlenecks. Someone near the front was asleep.
Claire leaned back in her chair. "Maya told me you texted her."
"Yeah."
"She said it was weird."
"It wasn't weird."
"She used the word weird three times." Claire pulled out her phone and scrolled through something. "Here. Quote: 'He was weird at breakfast, then weird at the health center, then sent me a weird text like he was apologizing for something I don't remember.'"
I didn't respond.
"So." Claire pocketed the phone. "What are you apologizing for?"
I could have lied. I could have deflected, changed the subject, made a joke about being generally awkward and left it at that.
Instead I said, "I think I'm the reason she collapsed."
Claire's expression didn't change. "Explain."
"The system. It logged a consequence. Indirect harm. She was the target."
"And you know this how?"
"It told me."
Claire was quiet for a moment. Then: "Did you do something to her?"
"No. I don't know. That's the problem—I don't know what I did. It could have been anything. Could have been weeks ago."
"So you're trying to figure out which of your past mistakes caused her to trip."
"She didn't trip. Her knee gave out."
"Because of you."
"According to the system, yeah."
Claire looked at me like she was deciding whether to be angry or just disappointed. She settled on something in between.
"You know how that sounds, right?"
"I know."
"You're treating people like a spreadsheet."
I started to argue. Stopped.
She was right.
I'd been scrolling through stats, looking for guilt like it was a build optimization problem. Searching for the perfect ratio of remorse to consequence severity, like if I just found the right formula I could balance it all out and feel okay again.
That's not how it worked.
"You can't optimize this," Claire said.
"I know."
"Do you?"
I didn't answer.
Professor Went wrapped up the lecture with some half-hearted reminder about a reading assignment. People started packing up. Claire didn't move.
"You want my advice?" she asked.
"Not really."
"Stop trying to solve it. You can't. Guilt isn't a puzzle. You just have to sit with it."
"That's a terrible strategy."
"It's the only one that works." She stood up, slinging her bag over her shoulder. "Also, maybe stop treating every interaction like it's going to explode. That's not helping anyone."
She left before I could respond.
I stayed in my seat for a while after the room cleared. My laptop was still open. The lecture slides were still on the screen, frozen on some chart about supply chain efficiency.
I closed it.
Then I pulled up the system interface one more time.
Still no guilt stat.
Still no metric for remorse.
Just the same cold architecture, the same clinical breakdowns, the same ledger tracking harm like it was inventory.
I was about to close it when something changed.
A new tab appeared.
CHAIN ANALYSIS
I tapped it.
The window that opened was different from the others. Less structured. More... relational.
It showed connections.
Me at the center. Lines radiating out to other nodes—people I'd interacted with, people who'd been affected. Some lines were solid. Some were dotted. A few pulsed faintly, like they were still active.
Maya's node was near the top. A dotted line connected her to me, then branched off to two other nodes I didn't recognize.
I zoomed in.
One of the secondary nodes had a timestamp from two weeks ago. A brief interaction. I didn't even remember it.
The other was older. A month back. Same story—something I'd done, something I'd forgotten, something that had quietly added Maya to the list of people close enough to catch fallout.
And now the system was showing me the whole chain.
Not to help.
To make sure I understood.
SYSTEM NOTICE
Chain visibility: Enabled.
Guilt is not a tracked variable.
Consequence is.
I closed the interface.
Sat there in the empty classroom for another minute.
Then I grabbed my bag and left.
Claire was right.
I couldn't optimize this.
I could only watch the chains grow.
