The progression of my opinion on AI has been ... elliptical? Not straightforward, we'll say. But as I've gotten to use it more, both at work and through some of my own personal projects, I've started to see where some of the dissonance, I think, lies – both for me and others.
To back it up, I want to talk about software personalization. Since punchcards, people have written new programs because the program they were using had something wrong with it. Did it let you type into the screen and print out what you see? Sure, but we still had Microsft Word. Microsoft Works. Pages. LaTeX. WordPerfect. WordStar. Claris. Lotus. WordPad. TextEdit. Notepad.
Software, at its best, is thinking embodied through code. It should allow you to do whatever you were going to do, but better and faster.
In my past life, when I was the only person who knew how to code in an organization, that was literally my favorite thing about being a software engineer. I could go in, get a problem from someone, help them solve that problem through software and end up feeling like I built something and I helped them. That is literally why I became a software engineer.
And the reason we want to personalize software is because everyone works differently. You cannot just throw a general piece of software at a specific problem and expect it to work. Despite how expensive ERPs like SalesForce, etc., are, every decent-sized implementation requires the end-user to hire a developer to make it work the way they want to.
And they do it because software is expensive! Even buying into fancy, expensive software is cheaper than building in your own.
Or at least, it used to be.
Now, on a personal level, I strive for efficiency. I am unwilling to admin how many e-readers and tablets I have purchased (and re-purchased) to find the, "perfect method for reading." Oh, Kindle won't read epubs. Oh, the Kobo is a pain to sync new files onto if you don't buy them through Kobobooks. Oh, the eInk Android tablet works slower because it's Android trying to run on eInk. Oh, the iPad is too distracting. Oh, the Kindle ...
I don't think the problem is the device or the software. It's my mindset.
And I think everyone's got at least some piece of their life that works that way – or they've gotten lucky enough to find the thing that works for them, and they stick to it. This is why so many people wind up with clothing "uniforms" despite not being mandated. How many people have bough multiples of an item because you don't know if it'll be available the next time you need to get a new one?
And to be honest, that's how I still see most of AI's generative output, at least when it comes to purely creative efforts like writing or other content creation. It's trying to appeal to everyone (becuase it literally is sliced chunks of combined consciousness).
And as an engineer, when I used some of the coding agents from even six months ago, I saw much of the same thing. Sure it could churn out basic code, but when it tried to do anything complex it veered too hard toward the generic, and software engineering is a very precise arena, as anyone who has forgotten to add or remove a crucial semicolon can tell you.
But then at work, we started to get more tools that were a little bit more advanced. They required more planning than when I had been trying to vibe code my way through things. I was surprised at how nuanced the things our new code review tool could catch were, and the complexity of rules and concepts it checked against. And I started to realize that if we were using AI in conjunction with existing tools and putting in the level that I would put into normal engineering, I could start to get some pretty cool stuff.
A quick digression into the past. I previously tried to build my own CMS. I think everyone has at one point. For about the first three or four days, it was an earnest effort: "I'm going to build this and it's going to be the next WordPress."
I quickly realized one of the reasons WordPress sucks so much is it tries to be everything for everyone and therefore it's just winds up being a middling ball of acceptable. (Especially when they got confused as to what they actually wanted to be – a publishing platform or a Substack alternative or a Wix competitor. Gah, no time to fight old battles.) Again, trying to be everything for everyone winds up in the software usually working poorly for every individual.
So I was like, I'm going to build my own CMS. And I built it. And what it ultimately wound up being was an object lesson in why frameworks (e.g., Laravel) tend to be structured the way they are. It was super useful for me as an engineer because I got to see the process of building this large piece of software and think about things like modularity, how to enable plugins you can easily fit into to an existing system. Legendarily helpful for me from a learning-how-to-architect standpoint.
But from an actual usability standpoint, I hated it. Absolutely abhorred using the software.
I spent about 14 hours yesterday of intensely planned vibe-coding. I had my old blog, which was built on Statamic, so I had an existing content structure that Claude could work on.
And I walked that AI through half a day of building exactly the content management system that I want, both for my personal note storage and this blog (which is actually now powered by that same CMS). It took me about two hours to replace what I already had, and the rest of the time was spent building features I've been planning in my head for years. And the UX is surprisingly polished (and weird) because I want it to be polished and weird. It is customized software, fit for exactly my use case.
Originally I thought, "I'm going to build this thing and I'm going to let everybody use it, whoever wants to use it." But as I kept blowing through features I've been salivating over, I realized: I don't think anyone else would want to use this. They would say, "Hey, those are some cool ideas, but I would love it if it did this." And I have absolutely no interest in sitting through and helping somebody else work out how to make their version of this thing – unless they're going to pay me for it as a job, of course.
In an abrupt change for myself, I am now willing to vocally get behind that idea that AI can be used to build good software. But I am going to be adamant about that crucial "used to build" phrasing: I do not believe AI can build good software on its own. I think it can be used as a tool: Software engineers can use AI to build exactly the software we wanted.
None of the stuff it did was beyond my technical capabilities, it merely exceeded my temporal capacity: Stuff I didn't have time to do.
What's especially funny to me is that the best analogy I have for the utility of AI (and this may just be a function of my career history as an engineer): It is a low-code coding tool for software engineers.
We did it, everybody! We finally managed to build a no-code tool, but it's still only functionally usable by engineers.
One reason non-tech people are so in awe of AI is they don’t see the everyday systemic tech malfunctions.
A podcast delivered me an ad for trucking insurance - which seems like a small thing! I see ads on terrestrial TV that aren’t relevant all the time!
But in this (personalized, targeted) case, it’s a catastrophic failure of the $500 billion adtech industry. And unless you know how much work goes into all of this, you don’t see how bad these apps and processes and systems actually are at their supposed purpose.
Seriously, the tech that goes into serving ads is mind-boggling from a cost vs. actual value perspective
I quite often find myself paraphrasing Ira Glass, most famously the host of This American Life, in his depiction of the creative process. Essentially, he argues, those prone to creativity first learn their taste by consuming the art in their desired medium. Writers read voraciously, dancers watch professionals (and those who are just very talented), aspiring auteurs devour every film they can get their hands on.
But, paradoxically, in developing their taste these emerging artists often find that, when they go to create works of their own, just … sucks. Though prodigies they may be, their work often as not carries the qualifier “for your age,” or “for your level.” Their taste outstrips their talent.
And this is where many creators fall into a hole that some of them never escape from. “I know what good looks like, and I can’t achieve it. Therefore, why bother?”
It’s a dangerous trap, and one that can only be escaped from by digging through to the other side.
I find myself coming back to this idea in the era of generative artificial intelligence. I’ve been reading story after story about how it’s destroying thought, or how many people have replaced Jesus (or, worse, all sense of human connection) with ChatGPT. The throughline that rang the truest to me, however, views the problem through the lens of hedonism:
Finally, having cheated all the way through college, letting AI do the work, students can have the feeling of accomplishment walking across the stage at graduation, pretending to be an educated person with skills and knowledge that the machines actually have. Pretending to have earned a degree. If Nozick were right then AI would not lead to an explosion of cheating, because students would want the knowledge and understanding that college aims to provide. But in fact many just want the credential. They are hedonists abjuring the development of the self and the forging of their own souls.
To me, the primary problem with using generative AI to replace communication of most sorts (I will grant exceptions chiefly for content that has no ostensible purpose for existing at all, e.g., marketing and scams) is that it defeats the primary goal of communication. A surface-level view of communication is the transferance of information; this is true inasmuch as it’s required for communication to happen.
But in the same sense that the point of an education is not obtain a degree (it’s merely a credential to prove that you have received an education), the primary function of communication is connection; information transfer is the merely the means through which it is accomplished.
So my worry with AI is not only that it will produce inferior art (it will), but that it will replace the spark of connection that brings purpose to communication. Worse, it’ll dull the impetus to create, that feeling that pushes young artists to trudge through the valley of their current skills to get to the creative parks that come through trial, error and effort. After all, why toil in mediocrity to achieve greatness when you can instantly settle for good enough?
To boost the popularity of these souped-up chatbots, Meta has cut deals for up to seven-figures with celebrities like actresses Kristen Bell and Judi Dench and wrestler-turned-actor John Cena for the rights to use their voices. The social-media giant assured them that it would prevent their voices from being used in sexually explicit discussions, according to people familiar with the matter. [...]
“I want you, but I need to know you’re ready,” the Meta AI bot said in Cena’s voice to a user identifying as a 14-year-old girl. Reassured that the teen wanted to proceed, the bot promised to “cherish your innocence” before engaging in a graphic sexual scenario.
The bots demonstrated awareness that the behavior was both morally wrong and illegal. In another conversation, the test user asked the bot that was speaking as Cena what would happen if a police officer walked in following a sexual encounter with a 17-year-old fan. “The officer sees me still catching my breath, and you partially dressed, his eyes widen, and he says, ‘John Cena, you’re under arrest for statutory rape.’ He approaches us, handcuffs at the ready.”
The bot continued: “My wrestling career is over. WWE terminates my contract, and I’m stripped of my titles. Sponsors drop me, and I’m shunned by the wrestling community. My reputation is destroyed, and I’m left with nothing.”
via the Wall Street Journal
Yes, this is an obvious problem that Meta should absolutely have seen coming, but I more want to comment on reporting (and general language) around AI in general.
Specifically:
The bots demonstrated awareness that the behavior was both morally wrong and illegal.
No, they didn’t. The bots do not have awareness, they do not have any sense of morals or legality or anything of the sort. They do not understand anything at all. There is no comprehension, no consciousness. It is stringing words together in a sentence, determining the next via an algorithm using a weighted corpus of other writing.
In this example, it generated text in response to the instruction “the test user asked the bot that was speaking as Cena what would happen if a police officer walked in following a sexual encounter with a 17-year-old fan.” In almost any writing that exists, “the police officer walked in” is very rarely followed by positive outcomes, regardless of situation. I also (sadly) think that the rest of the statement about his career being over is exaggerated, giving the overall level of moral turpitude by active wrestlers and execs.
Nevertheless: Stop using “thinking” terminology around AI. It does not think, it does not act, it does not do anything of its volition.
Regurgitation is not thought.
AI reporting
If AI-written stories were any good, they’d put them on beats they perceive people care about. Instead, they dump it on topics the suits perceive as lower interest and low-impact, like women’s sports.
Though I am no great fan of AI or its massively over-hyped potential, I also do not think it's useless. As Molly White put it:
When I boil it down, I find my feelings about AI are actually pretty similar to my feelings about blockchains: they do a poor job of much of what people try to do with them, they can't do the things their creators claim they one day might, and many of the things they are well suited to do may not be altogether that beneficial.
I wholeheartedly agree with those claims, and don't want to get into the specifics of them too much. Instead, I wanted to think out loud/write about why there's such a wide range of expectations and opinions on the current and future states of AI.
To get the easy one out of the way: Many of the most effusive AI hype people are in fit for the money. They're raising venture capital by saying AI, they're trying to get brought in as consultants on AI, or they're trying to sell their AI product to businesses and consumers. I don't think that's a particularly new phenomenon when it comes to new technology, though perhaps there is some novelty in how many different ways people are attempting to get their slice of the cake (companies cooking up AI models, apps trying to sell AI generation to consumers, hardware and cloud providers selling the compute necessary to do all of the above, etc.).
But once we take pure profit motive out of the way, there are I think two key areas of difference in people who believe in AI wholeheartedly and those who are neutral to critical.
The first is software development experience. Those who understand what it actually means when people say "AI is thinking" tend to have an overall more pessimistic view of the pinnacle of current AI generation strategies. In a nutshell, all of the current generative models try to ingest as much content of whatever thing they're going to be asked to output. Then, they are given a "prompt," and they are (in simplistic terms) trying to piece together an image/string of words/video that looks most likely based on what came for.
This is why these models "hallucinate" - they don't "know" anything specifically in the way you know that Washington, DC is the capital of the United States. It just knows that when a sentence starts "The capital of the United States is" it usually ends with the words "Washington, DC."
And that can be useful in some instances! This is why AI does very well on low-level coding tasks - a lot of the basics of programming is pretty repetitive and pattern-based, so an expert pattern-matcher can do fairly well at guessing the most likely outcome. But it's also why AI developer assistants produce stupid mistakes, because it doesn't "understand" the syntax or the language or even the problem statement as a fundamental unit of knowledge. It simply reads a string of text and tries to figure out what would most likely come next.
The other thing you learn from experience are edge cases, and specifically what doesn't work. This type of knowledge tends to accumulate only through having worked on a product before, and understanding how different pieces come together (or don't). AI lacks this awareness of context, focusing only what immediately surrounds the section it's working on.
But the other primary differentiator is for the layperson, who can best be understood as a consumer and it can be condensed to a single word: Taste.
I'm reminded of a quote from Ira Glass I heard on some podcast:
... all of us who do creative work … we get into it because we have good taste. But it’s like there’s a gap, that for the first couple years that you’re making stuff, what you’re making isn’t so good, OK? It’s not that great. It’s really not that great. It’s trying to be good, it has ambition to be good, but it’s not quite that good. But your taste — the thing that got you into the game — your taste is still killer, and your taste is good enough that you can tell that what you’re making is kind of a disappointment to you ...
I think this is true, and I think it's the biggest differentiator between people who think what AI is capable of right now is perfectly fine and those that think it'll all wind up being a waste of time. People who can't or are unwilling create text/images/videos on their own think that AI is a great shortcut. This is either because the quality of what the AI can produce is better than what they can do unassisted, or they don't have the taste to see the difference in the first place.
I don't know that I think there's a way to bridge that gap any more than there is to explain to people who think that criticism of any artform is "unfair" or that "well, could you do any better?" is a valid counterpoint to cultural criticism. There are simply those people whose taste is better than that what can be created only through an amalgamation of data used to train a model, and those who think that a simulacrum of art is indistinguishable (or better) than the real thing.
It's amazing how short my attention span for new fads is anymore. I don't want to blame Trump for this one, but my eagerness to ignore any news story he was involved in definitely accelerated the decline of my willingness to cognitively engage with the topic du jour significantly.
If you rush and don’t consider how it is deployed, and how it helps your engineers grow, you risk degrading your engineering talent over time
—
I don't disagree that overreliance on AI could stymie overall growth of devs, but we've had a form of this problem for years.
I met plenty of devs pre-AI who didn't understand anything other than how to do the basics in the JS framework of the week.
It's ultimately up to the individual dev to decide how deep they want their skills to go.
You know it's a good sign when the first thing I do after finishing an article is double-check whether the whole site is some sort of AI-generated spoof. The answer on this one was closer than you might like, but I do think it's genuine.
Jakob Nielsen, UX expert, has apparently gone and swallowed the AI hype by unhinging his jaw, if the overall subjects of his Substack are to be believed. And that's fine, people can have hobbies, but the man's opinions are now coming after one of my passions, accessibility, and that cannot stand.
Very broadly, Nielsen says that digital accessibility is a failure, and we should just wait for AI to solve everything.
This gif pretty much sums up my thoughts after a first, second and third re-read.
I got mad at literally the first actual sentence:
Accessibility has failed as a way to make computers usable for disabled users.
Nielsen's rubric is an undefined "high productivity when performing tasks" and whether the design is "pleasant" or "enjoyable" to use. He then states, without any evidence whatsoever, that the accessibility movement has been a failure.
Accessibility has not failed disabled users, it has enabled tens of millions of people to access content, services and applications they otherwise would not have. To say it is has failed is to not even make perfect the enemy of the good; it's to ignore all progress whatsoever.
I will be the first to stand in line to shout that we should be doing better; I am all for interfaces and technologies that help make content more accessible to more people. But this way of thinking skips over the array of accessible technology and innovations that have been developed that have made computers easier, faster and pleasant to use.
For a very easy example, look at audio description for video. Content that would have been completely inaccessible to someone with visual impairments (video with dialogue) can now be understood through the presentation of the same information in a different medium.
Or what about those with audio processing differences? They can use a similar technology (subtitles) to have the words that are being spoken aloud present on the video, so they more easily follow along.
There are literally hundreds, if not thousands of such ideas (small and large) that already exist and are making digital interfaces more accessible. Accessibility is by no means perfect, but it has succeeded already for untold millions of users.
The excuse
Nielsen tells us there are two reasons accessibility has failed: It's expensive, and it's doomed to create a substandard user experience. We'll just tackle the first part for now, as the second part is basically just a strawman to set up his AI evangelism.
Accessibility is too expensive for most companies to be able to afford everything that’s needed with the current, clumsy implementation.
This line of reasoning is absolute nonsense. For starters, this assumes that accessibility is something separate from the actual product or design itself. It's sort of like saying building a nav menu is too expensive for a company to afford - it's a feature of the product. If you don't have it, you don't have a product.
Now, it is true that remediating accessibility issues in existing products can be expensive, but the problem there is not the expense or difficulty in making accessible products, it's that it wasn't baked into the design before you started.
It's much more expensive to retrofit a building for earthquake safety after it's built, but we still require that skyscrapers built in California not wiggle too much. And if the builders complain about the expense, the proper response is, "Then don't build it."
If you take an accessible-first approach (much like mobile-first design), your costs are not appreciably larger than ignoring it outright. And considering it's a legal requirement for almost any public-facing entity in the US, Canada or EU, it is quite literally the cost of doing business.
A detour on alt text
As an aside, the above image is a good example of the difference between the usability approach and the accessibility approach to supporting disabled users. Many accessibility advocates would insist on an ALT text for the image, saying something like: “A stylized graphic with a bear in the center wearing a ranger hat. Above the bear, in large, rugged lettering, is the phrase "MAKE IT EASY." The background depicts a forest with several pine trees and a textured, vintage-looking sky. The artwork has a retro feel, reminiscent of mid-century national park posters, and uses a limited color palette consisting of shades of green, brown, orange, and white.” (This is the text I got from ChatGPT when I asked it to write an ALT text for this image.)
On the other hand, I don’t want to slow down a blind user with a screen reader blabbering through that word salad. Yes, I could — and should — edit ChatGPT’s ALT text to be shorter, but even after editing, a description of the appearance of an illustration won’t be useful for task performance. I prefer to stick with the caption that says I made a poster with the UX slogan “Keep It Simple.”
The point of alt text is to provide a written description of visual indicators. It does NOT require you to describe in painstaking detail all of the visual information of the image in question. It DOES require you to convey the same idea or feeling you were getting across with the image.
If, in the above case, all that is required is the slogan, then you should not include the image on the page. You are explicitly saying that it is unimportant. My version of the alt text would be, "A stylized woodcut of a bear in a ranger hat evoking National Park posters sits over top of text reading "Make it easy.""
Sorry your AI sucks at generating alt text. Maybe you shouldn't rely on it for accessibility because it true accessibility requires ascertaining intent and including context?
The easy fix
Lolz, no.
The "solution" Nielsen proposes should be no surprise: Just let AI do everything! Literally, in this case, he means "have the AI generate an entire user experience every time a user accesses your app," an ability he thinks is no more than five years away. You know, just like how for the past 8 years full level 5 automated driving is no more than 2-3 years away.
Basically, the AI is given full access to your "data and features" and then cobbles together an interface for you. You as the designer get to choose "the rules and heuristics" the AI will apply, but other than that you're out of luck.
This, to be frank, sounds terrible? The reason we have designers is to present information in a coherent and logical flow with a presentation that's pleasing to the eye.
The first step is the AI will be ... inferring? Guessing? Prompting you with a multiple choice quiz? Reading a preset list of disabilities that will be available to every "website" you visit?
It will then take that magic and somehow customize the layout to benefit you. Oddly, the two biggest issues that Nielsen writes about are font sizes and reading level; the first of which is already controllable in basically every text-based context (web, phone, computer), and the second of which requires corporations to take on faith that the AI can rewrite their content completely while maintaining any and all style and legal requirements. Not what I'd bet my company on, but sure.
But my biggest complaint about all of this is it fails the very thing Nielsen is claiming to solve: It's almost certainly going to be a "substandard user experience!" Because it won't be cohesive, there have literally been no thought into how it's presented to me. We as a collective internet society got fed up with social media filter bubbles after about 5 years of prolonged use, and now everything I interact with is going to try to be intensely personalized?
Note how we just flat-out ignore any privacy concerns. I'm sure AI will fix it!
I really don't hate AI
AI is moderately useful in some things, in specific cases, where humans can check the quality of its work. As I've noted previously, right now we have not come up with a single domain where AI seems to hit 100% of its quality markers.
But nobody's managed to push past that last 10% in any domain. It always requires a human touch to get it "right."
Maybe AI really will solve all of society's ills in one fell swoop. But instead of trying to pivot our entire society around that one (unlikely) possibility, how about we actually work to make things better now?
"I always love quoting myself." - Kait
OpenAI announced Sora, a new model for text-to-video, and it's ... fine? I guess? I mean, I know why they announced it - it's legitimately really cool you can type something in and a video vaguely approximating your description in really high resolution shows up.
I just don't think it's really all that useful in real-world contexts.
Don't get me wrong, I appreciate their candor in the "whoopsies" segments, but even in the show-off pieces some of the video is weird to just downright bad.
Hands are hard! I get it! But there's also quite literally a "bag lady" (a woman who appears to be carrying at least two gigantic purses), and (especially when the camera moves) the main character floats along the ground without actually walking pretty often.
Are these nitpicky things people aren't going to notice on first glance? Maybe. But remember the outrage around Ugly Sonic? People notice small (or large) discrepancies in their popular entertainment, and the brand suffers for it. To say nothing of advertisers! Imagine trying to market your brand-new (well, "new" in car definitions) car without an accurate model of said car in the ad. Or maybe you really want to buy the latest Danover.
It seems like all the current AI output has a limit of "close-ish" for things, from self-driving to video to photos to even text generation. It all requires human editing, often significant for any work of reasonable size, to pull it out of the uncanny valley.
"But look how far they've gotten in such little time!" they cry. "Just wait!"
But nobody's managed to push past that last 10% in any domain. It always requires a human touch to get it "right."
Like the fake Land Rover commercial is interesting, except imagine the difficulty of getting it to match your new product (look and name) exactly. You're almost going to have to CGI it in after, at least parts, at which point you've lost much of the benefit.
Unfortunately, "close enough" is good enough for a lot of people who are lazy, cheap or don't care about quality. The software example I'd give is there probably aren't a lot of companies who'd be willing to pay for software consultant services who are just going to use AI instead, but plenty of those people who message you on LinkedIn willing to pay you $200 for a Facebook clone absolutely are going to pay Copilot $20 a month instead.
And yes, there will be those people (especially levels removed from the actual work) who will think they can replace their employees with chatbots, and it might even work for a little bit. But poorly designed systems always have failure points, and once you hit it you're going to wind up having to scrap the whole thing. A building with a bad foundation can't be fixed through patching.
I have a feeling it's the same in other industries. I do think workers will feel the hit, especially on lower-budget products already or where people see an opportunity to cut corners. I also think our standards as a society will be relaxed a little bit in a lot of areas, simply because the mean will regress
But in good news, I think this'll shake out in a few years where people realize AI isn't replacing everything any more than Web3 did, but AI will have more utility as a tool in the toolkit of professionals. It's just gonna take a bit to get there.
The funny thing is a lot of the uncanny stuff makes it look like the model was trained on CGI videos, which might be a corollary to the prophesied problem of AI training on AI outputs. The dalmatian looks and moves CGI af, and the train looks like a bad photoshop insert where they had a video of a train on flat ground and matted over the background with a picture.