When you cross your fingers, you’re either blindly hoping for the best, or you’re lying. Or both.

As a director for a media literacy non-profit who is in fairly regular contact with a couple big-tech/media leaders on the front lines in the AI field, I’ve been deep-diving into this for months, now, and here’s where I’m at:

I believe the proper analogy is: AI is crypto, social media algorithms, and nuclear tech all rolled into one.


Like crypto, the positives are over-hyped (in the case of AI, that’s due to the hallucination problem, which will require human oversight for anything important). But, the similarity is striking: in its current iteration, there turns out to be very few good use-cases, really…except to facilitate crimes. I’ve been told the answer to this (y’know, for self-driving cars or medical diagnoses) is to have, non-AI algorithms to supervise and quality control all AI decisions.

Uh, okay. But where do humans fit into all of this again? Tell me again how much of a hit in the job market this will cause and why it’s worth it?

Like Social Media algorithms, a really robust regulation regime might possibly leverage it into some real positives; but, instead, untested and unregulated prior to rollout, it will accelerate the deterioration of society (genocides and political lurching toward authoritarianism, genocides in places where institutions are already weakened, all Art will continue to regress toward the mean, and negative mental health outcomes will continue to spike…all while we continue to largely ignore other existential problems like Global Warming and Pandemic readiness).

And like Nuclear Tech—due to its sheer power, emergent ‘theory of mind’ properties and exponential growth—the downside is existential and probably no upside would really be worth it, anyway.

But, the genie is out of the bottle, so that’s where we’re at.

The (I think, obvious, and only truly ethical) answer is to use the “Precautionary Principle” and stop it as much as possible until its tested and a robust international regulatory regime is in place.

Meanwhile, AI tech leaders are asking for regulation even while they race to roll it out: Fingers Crossed, indeed.

Leave a Reply

Your email address will not be published. Required fields are marked *