The Scale of Victims

(This column is posted at www.StevenSavage.com, Steve’s Tumblr, and Pillowfort.  Find out more at my newsletter, and all my social media at my linktr.ee)

It sure seems there’s a lot of IT security breaches lately. In fact, it’s to the point where I can’t remember which one inspired this column. It’s probably just as well, since you can map whatever horrific violation of privacy you heard of this week onto this column. There, I’ve sort of written something relatively timeless because people are dumb.

One of the things I wonder about is why more CTOs, CIOs, and so forth aren’t being taken to court, followed by reporters, and in general held freaking responsible for their companies having lousy security. Yes there’s all sorts of shielding from accountability, but you think we’d see some effort, but I think one thing protecting them is that the company is seen mostly as a victim.

I’d argue that’s technically right, the companies were attacked by some external force. But treating companies as equivalent of people ignores their responsibilities. People, individual moral agents, can be victims, but corporations are not people and not moral agents, and treating them as victims like people lets them out of responsibilities. Sorry, Mitt Romney.

Think about a person who is a victim of a crime. Though people often try to blame victims, those blamers are usually both wrong and assholes (and sometimes justifying their own crimes). A person who is victim of a crime is a victim in that someone else chose to behave criminally.. Even if said victim enhanced their own danger it doesn’t remove the culpability of the criminal, who violated social and legal norms that people are expected to follow.

When I watch people shrug as corporation after corporation has customer records placed on the dark web, I see comments about how crappy their security is, but it doesn’t seem particularly judgmental. This impresses me as an echo of the don’t-blame-the-victim mentality.

But corporations are groups of people – organizations. That organization makes certain agreements and promises in order to exist. Security of data is, obviously, part of them. If one’s data is breached, despite the criminals actions, you also take responsibility as you are responsible. If you’re leadership, you should be on the line because you made a promise that this probably won’t happen.

Organizations are about promises and responsibility. Screw that up, and no matter why, someone has to pay as your failure hurt the organization and the people involved. You don’t have to restrain yourself on going after the people who did the actual crime, but corporations have made promises. If you can’t keep them, you’ve got a problem.

In fact, I’d say a corporation that suffers a data breach or similar failure must be investigated to see if it violated social norms. If the corporation made guarantees it could not and did not keep, if good faith effort was not made, the corporation was responsible. There is a failure of the company that echoes the action of the criminal, it too violated norms.

Of course we all know that if we at all ask this we’ll find a lot of corporations have done terrible at security. It’s all cost cutting, half-assed integration, and big bonuses. A lot of companies, if they were really investigated for security problems, would be locked down and sold off for being terrible.

(And yes, I work in Healthcare, which has insanely strict rules, but everyone should for everything, and we remember that these rules protect people.)

We don’t need to act like corporations are victims like people. If they can’t keep their promises, if security violations reveal they’ve done a poor job of protecting people, they’re part of the problem. Some of them should pay. Some shouldn’t exist.

Steven Savage

Living In The Future We Were Sold

(This column is posted at www.StevenSavage.com, Steve’s Tumblr, and Pillowfort.  Find out more at my newsletter, and all my social media at my linktr.ee)

We’re living in the future, and it’s lousy.

So-called AI is just Ultra-Clippy being shoehorned into everything that will temporarily goose stock prices. We’ve got computerized cars that allow us to bluescreen while driving, and universal automated cars are many dreams and lawsuits away. Phones gave us something like Star Trek gadgets, but we’re using them to become depressed by doomscrolling. I could make a comment on the Cybertruck, but honestly, that seems pointless.

We’ve got a lot of things that we think are futuristic, and a lot of them are lame, terrible, pointless, or have side effects. Plus you know, we’ve got climate change, Nazis, and pandemics as well.

The future isn’t what it used to be? No, the problem is we’re living in the future we were sold.

A lot of our futuristic ideas derive from popular culture, but that popular culture has nothing to do with what we can, should, or even may want to do. A lot of or popular culture is what people could sell us or what worked in media of the time. It has nothing to do with the possible or the necessary.

AI? It’s easier to just have Hunky Space Captain talk to the computer, because no one wants to watch someone scroll on a monitor. Besides, it sounds cool. Also if you’re bored eventually the computer can try to murder people as part of the plot, a real horror film twist. But do we need it?

Automated cars are a dream, especially if you’ve ever driven . . . well, anywhere. It’s a dream that’s cool and convenient and doesn’t have messy people, and looks awesome in films. It doesn’t deal with the reality that driving needs a moral actor to make decisions, even if you’re paying them by the mile. Also it doesn’t deal with outages, software updates, and crashes.

Then there’s our phones, our pocket computers. This is a totally understandable dream of course, going back to hand-held sci-fi gizmos and communicators. It’s just we never asked how we’d misuse them, as if people won’t find some weird use for technology five seconds after inventing it.

All of these are things we’ve seen in pop culture media since the 60’s (and I’d argue a lot of what we’re living in is very 80’s). But it’s not stuff from speculative fiction or deep analysis or asking hard questions of what we want and need in the future. It’s stuff that was fun to put into movies, tv, and comics.

That’s it. For many of us, the future we envision is something that was marketable.

So of course all the backfire we’re experiencing is a surprise. We weren’t buying a warning, we were buying a cool experience.

“A good science fiction story should be able to predict not the automobile but the traffic jam,” said Fredrick Pohl. Indeed it should. It’s just sometimes the warnings don’t sell – and other times people think the warning is cool (see many a stylish dystopia with lots of leather for no reason).

So much of the future that people want – or are trying to sell us at least – seems to just be whatever was laying around in pop culture for a while. It doesn’t have anything to do with speculation, or possibility, or what we need. It’s what many of us assume the future is supposed to be because we bought it.

But what is the future we really want and need? The struggle is to find that, and perhaps in this time where the future we bought is failing us, we have a chance to find it.

Steven Savage

The Future Was Never What It Was

(This column is posted at www.StevenSavage.com, Steve’s Tumblr, and Pillowfort.  Find out more at my newsletter, and all my social media at my linktr.ee)

“The Future Isn’t What It Used To Be” has been a saying for a while. In a time of resource-sucking hallucinating AIs, climate change, and ad-saturated social media, the saying seems more relevant. We’re not getting the future we expected, want, or needed.

As I muse over this, I think the problem is that we had expectations as opposed to asking what we really needed. We wanted a future that was past and present.

Let’s take the Cybertruck, which is one man’s vision of a futuristic vehicle. The Cybertruck – for whatever valid critiques may be made of it – a deliberate creation, from its tech to the low-poly appearance. It’s something out of past science fiction, shoehorned poorly into current technology The thing is it turns out what we want isn’t, well, that vision or its janky implementation.

Or Microsoft’s Recall feature, which records what you’re doing for some kind of recovery purpose, all while basically being a security nightmare. A cybersecurity writer noted that maybe this is just what you get when an aging group of leaders keeps forgetting things. Is it evil opportunism, or just people thinking of a future that solves only something they might think of?

I could of course go on, from wasteful AI today to cuecat in the past and so on. A whole lot of people are inventing, selling, and sometimes just lying about how they’re making the future we want or expect. Which really means what too many people wan tis a future based on old videogames and movies and current ill-thought-out-needs.

We’re not humanity wants or needs because it really seems we’re not trained to think about that.

We look at what we want, and assume it’s for everyone. We look at our childhood media fixation and figure it’s how it should be. Even when people are lying their butts off trying to make “number go up” they’re justifying it with such explanations. I’m pretty sure enough supposed “leaders” of the tech world have been justifying things so long they actually believe it.

I’d feel kind of better knowing I’ve been lied to more, but am really starting to feel a little too much kool-aid has been drunk. A lot of that kool-aid came from 80’s direct-to-video.

And right now people’s egos and money are on the line in these various bad tech decisions, so they’re not going to reverse without some pretty hard bumps. Delusion, short-sightedness, and personal income and reputation are pretty compelling. Besides The Market doesn’t reward you for insight and the news doesn’t fawn over you for saying what a dumbass you were.

I’m starting to think being able to make the future (and make it better) is sort of its own skillset. Clearly a business degree doesn’t help you. But neither does a writing degree as you might just create a new mental straightjacket. Designing a future that works doesn’t necessarily come from pushing around numbers and making pretty words.

But it’s a skill we desperately need right now, and maybe recognizing it is a start.

Steven Savage