Back in the days when AI development meant working towards something like an artificial human, I had three rules to bear in mind when reading most popular articles on the subject :
- AI does not yet have the the same kind of understanding as human intelligence.
- There is no guarantee rule 1 will always hold true.
- It is not necessary to violate rule 1 for AI to have a massive impact, positive or otherwise, intentional or otherwise.
These were from the earliest days of LLMs when there were many other types of AI floating around in the popular press. Most AI stories were about whether AI was or could be conscious or not, a position most serious people have moved on from completely (though not quite all). The "rules" made sense as a way to keep perspective, to independently remind myself that the author might well have gotten carried away or missed their own point.
Given the massive developments in the last few years, I think it's time for an update. There's also a nice piece on Clearer Thinking which I think has a sensible take, especially if you want more practical advice.
Personally, I still have more than a little sense of wonder about the whole thing. Regardless of what AI is used for or who controls it, I just think it's an astonishing feat to essentially teach a rock to think. I find it somewhat dismaying, much as with skeptics of the space program, that the realisation of one of mankind's oldest dreams is being treated so often with cynicism and fear more than wonder and enthusiasm. It's as though everyone is focusing on the corrupt capitalism in Jurassic Park more than they are the reincarnation of freakin' dinosaurs. Still, for this very reason, if I might vainly hope to rekindle some sense of fascination in the more cynical reader, I should also try to temper my own admiration.
Here, then, are my offerings. I'll try to keep them neutral-ish, but you've been duly forewarned as to my own bias.
0) LLMs are not human
This should be a default presumption. When I say LLMs are thinking, reasoning, or understanding, I am not saying they do so in an entirely human-like way. While I think it's legitimate to say they do all of these things, the sense in which this is meant must be very carefully defined or else presumed to be linguistic shorthand. But to be direct, LLMs are not conscious, have no will, no desires of their own, no inner awareness, no coherent long-term memory, no personality, function differently according to their current context window, etc. etc. etc.
In some narrow but important ways, they probably are doing something closely analogous to human thinking. In the right conditions, those similarities are fascinating, and we probably shouldn't dismiss LLMs as a dead-end in intelligence more broadly. But in more general ways, LLMs are absolutely nothing like humans. I get very frustrated when people dismiss LLMs out of hand because of their differences to human cognition when the similarities actually are interesting, but nevertheless, I completely agree with the basic premise that a net of linguistic probabilities doesn't count in any way as "alive".
1) The imperfect nature of AI does not render it useless
And the useful nature of AI does not render it perfect !
Much the most common flawed argument is rotten cherry-picking : focusing entirely on the mistakes that AI makes, especially the silly ones, and thereby extrapolating that it can't do anything at all – or at the very least that it's completely untrustworthy. Less common among my feeds is the opposite view, that because AI is able to do some incredibly complex analysis very well, it can be completely relied upon in all things, or at least that it's silly mistakes in simple problems are just not worth worrying about at all.
Both of these are wrong-headed. A better way to look at it might be for pessimists to say, "just because an AI isn't useful for me, it doesn't follow that it's of no use for anyone else". Conversely, the optimist's take would be, "just because I find AI useful, that doesn't mean that everyone else will necessarily do so as well".
The "jaggedness" of LLM-intelligence seems to cause people no end of strife. Sure, it can't understand some common sense things. So what ? All that should tell you is Rule 0 : that it isn't reasoning like a human. It does not tell you that its answers on more complex topics are therefore wrong. At most, it should act as a reminder to what's best practise in all situations : when something is important, you need to check any proposed solution from any source, rather than assuming blindly that the proffered answer is correct and immediately implementing it into your workflow.
A hilarious example : this case of ChatGPT showing blatant sycophancy in analysing a fart track as a serious musical composition. True, absolutely, it shouldn't do this. But to conclude that "your product sucks" is... I mean, I honestly don't understand this mentality at all*.
* Although I do understand it as a joke, of course, and I laughed along with the ending. Here I'm criticising people who actually do think like this, of which their numbers appear to be legion.
A much better, more nuanced take comes from this article on the use of AI in mathematics. Time was when LLMs couldn't even use a calculator, but that time is no more. Used correctly, they can be hella productive. It's worth reading that one in full – it covers the downsides quite nicely as well as the upsides – but the most interesting bit to me was the following :
The LLMs he spoke with inevitably made lots of mistakes, leading some mathematicians to dismiss them outright. Many researchers, he said, decide that if “everything it says is kind of wrong, I will just not talk to it.” But others — he puts himself in this camp — have a higher tolerance for “the pain of talking to this bullshitting model. They say, I can still get something out of this conversation; even if not every idea is good, I can ignore the bad ones and take the good ones.” And the mistakes, Schmitt noted, are weird ones: There is virtually no way that a person with any training in mathematics would make such a plethora of basic errors while also succeeding in coming up with subtle, original, and correct ideas.
Maybe LLMs annoy certain people because they're still thinking that they must be human-like to be useful, or are simply not prepared to accept anything except the smallest error rate : either they have to be fully human, or as perfect as a calculator, and anything in between constitutes an unacceptable uncanny valley.
I personally have always preferred to use the AI output as inspirational more than authoritative, and with that sort of mindset, even GPT-3.5 could be quite useful. If you're looking for a Truth Engine, go home, but then... why did you ever believe there was any such authority anyway ? Why would you assume that human experts have an error rate of zero ?
I think there's a lot of double standards being applied here. Apparently, people can accept that other people might sometimes be wrong without dismissing them entirely, but such errors in LLMs seem to render them as useless junk for some reason. I find it weird. I also find the opposite techbro mentality weird, mind you : just because some mistakes seem trivial doesn't mean they don't matter at all, and just because they're very good in some situtations, it doesn't follow they should be shoved into absolutely everything.
2) AI is used by real humans in the real world, including very stupid and very clever people
Following on from that, I think AI-skeptics should approach any AI article from the stance that a vast number of people do find using AI beneficial, and that they're not all deluding themselves. Conversely, those of us who are more optimistic should acknowledge that not every negative study is necessarily flawed, and that some concerns are motivated out of entirely sensible considerations based on human psychology rather than cynical views of the techbro ilk. The scale of AI adoption is vast, and it makes no sense to say that all these hundreds of millions of users aren't seeing any benefit at all, nor to dismiss the possibility of downsides from such a rapid, enormous uptake.
Two contrasting pieces : this one in The Conversation (a usually skeptical website) finds that most students aren't just using AI to do all their homework for them, but actively engage with its output and revise it according to their own needs. This is much my own approach : I almost always reword AI text (on the rare occasions I use it for text, which I dislike doing) to suit my own style, even if the AI version might sound better. Conversely, this piece in Ars Technica gives a detailed description of the problems AI has caused for teachers, with the temptation to simply go to an AI for the answer – even if the student then rewords the thing – being sometimes irresistible.
Quite honestly I don't know what to do about that. In school, we probably want to keep AI and maybe even computer use down to a minimum, with single-use devices like books, pencil and paper etc. being innately better at creating focus. It's always seemed to me that this trajectory is obvious : begin with training on the basics so you have a full, deep understanding of what's going on and can do it on your own, then gradually transition to using more and more learning aids like reference books and calculators and so on. In this way you move slowly into the real world, with a solid grounding in the fundamentals so you can make better use of all the productivity boosters everyone uses when they have to actually get stuff done for real. Keep exams device-free when necessary and that's all there is to it.
The difficulty with this is coursework. In principle, this is the best guide to how to use knowledge and skills in the real world. In the pre-LLM world it was relatively easy to set a task that couldn't be automated, and I personally always preferred this to examinations. Exams carry a weird kind of stress that isn't replicated in real life, whereas coursework can be done more at one's own pace. I preferred it and would have encouraged it to replace examinations as much as possible. But with LLMs in the picture, I honestly don't know what to do.
The only thing I can offer is to acknowledge that coursework is still a chore. I believe quite strongly in a work-life balance, and the need to continue working outside of working hours is something I always found depressing : even when it's something I enjoy doing, I dislike being compelled to do it during what seems like it should be my own personal time, even when I can largely set my own schedule. So I wouldn't want to knee-jerk to "students are cheating" here* : they're doing exactly what the rest of us are doing, a perfectly natural reaction to avoiding things they'd rather not do.
* Indeed, some of my students would probably benefit from using LLMs a lot more to polish their language.
Maybe the only solution here is, as with multi-functional devices, to take them away. Give students a good working environment where they can go at any time for coursework in which LLM-use is absolutely restricted... I don't know.
Using AI definitely isn't something we should allow to be complete free-range in all situations for all people, but at the same time, it definitely isn't something we should strangle at birth either. While I think the suicide/murder stories are not something worth taking very seriously – there are hundreds of millions of users, and if AI was a causal factor in this, then violence would already have skyrocketed – there are definite concerns about over-use and the degradation of critical thinking and suchlike. I just think that while it might have negative effects on some, this does not automatically offset the positive benefits for others. Cherry-picking on either side isn't helpful.
3) This is the worst it will ever be
Finally, even quite recently I would have said that AI could never do a whole bunch of things it's now reasonably competent at. How far this is going to progress is a matter of debate, but while I find this video to be largely hyperbole, it has one outstanding point : if we're not good at intuiting exponential progress, we're even worse at understanding S-curve exponentials. That is, development follows an exponential trajectory but only by averages. Sometimes there can be protracted periods – months or more – when development plateaus or increases only slowly, but these are followed by short periods of enormous breakthroughs.
Thus far the pattern has held every time the nay-sayers have insisted that AI development is hitting a wall. The CEO of Microsoft (for whatever that's worth, which is not much but not nothing) says that there's no sign of this happening in the foreseeable future, while if you follow AI news, you'll know that there are plenty of other avenues under investigation for advancement beyond raw computing power.
The takeaway from this one is simple : to say that "AI can't do this and therefore it's useless to me" is a largely vacuous statement. There is absolutely no guarantee that it won't be able to what you need in the (very) near future. Some things will likely take longer than others, but realistically, nothing is off the cards. Full automation even for the most complex of jobs looks like a real possibility, and some of the seeming hype is worth taking seriously. To pretend that we're still in the era of GPT-3.5 is not at all sensible, and to "hope" that development will somehow just stop here is scarcely any better.
Whether you think that this will be a good thing or not is another matter. If you think that LLMs thus far have been generally positive, presumably you think that further advances will be more of the same. Conversely, if you think they're been detrimental, you probably don't want to see them continue. Neither is correct : if AI use thus far has been generally positive, it does not follow that further developments cannot be problematic; equally, and conversely, if AI has been harmful thus far, it does not follow that future developments must inevitably be more of the same.
My point is that it's so easy to pick and choose whatever set of stories you want to support your position, whereas reality is likely more complex than either. If we can predict how LLMs will advance for at least the next few years, predicting what humans will do with them is another matter entirely : here I would tend to side with the cynics much more than the techbros, even I think they're hardly going in to usher some kind of apocalypse. Reaching for the heuristics of "I dis/like what's happened so far, so we can expect more of the same in the future", is, however, simply not good enough, especially in this most non-linear of development trajectories.
That's my take then. That LLMs aren't human doesn't mean they're useless, nor are they perfect or their mistakes inconsequential. Neither the benefits nor the downsides can be taken to completely offset the other and cherry-picking from either perspective is a trap which is perilously easy to fall into. Accounting for how people, both those who do and don't understand their operation, actually use them, is already a mixed bag, and predicting what comes next is only going to get harder even though we can actually make quite a reasonable extrapolation as to future LLM performance abilities. Critiquing LLMs for current shortcomings is valid, but it's worth getting some perspective and realising they have already made truly astonishing gains, and insisting that current problems are unsolvable just lacks any common sense.
The future is coming whether we want it to or not, and to try and force everyone who benefits from it to put it back in the bottle even for those who are genuinely badly affected by it... is just not a realistic expectation of humanity. LLMs are not truth engines, but they are certainly powerful thinking engines of a sort, which can no more be stopped than the rise of steam power. Whether they will have the same degree of impact I don't know, I still tend towards thinking probably not, or at least not yet, but one thing I am confident on is that to dismiss them entirely is just not a sensible thing at all.