Exploring and Evaluating Generative Artificial Intelligence Number Three

I decided to make a Header Image (above) for this little series of posts and have retrofitted it to the two previous posts here and here. So I asked the Midjourney app on Discord, to depict a silver-skinned Android, firstly, standing at an easel painting, and then at a computer typing. I am fairly sure that the AI known as Midjourney had no sense of the irony of asking it to anthropomorphise an Android doing these activities, because current forms of AI are so far from having the sentience required to appreciate concepts as subtle as irony. Spoiler alert, I approached this evaluative exploration with certain preconceptions about the likely conclusion although I didn’t know for sure, how those conclusions might be reached because I didn’t know how AI’s work, in detail. What I am going to show you today is what I have learned, but I am also going to link you to a very erudite analysis of why we should not be worried about AI taking over the world – in a piece calledWhy the AI Doomers Are Wrong“, Nafeez Ahmed explains why the direction of travel of AI development, simply can’t lead to a human-brain-like sentience. I will quote from his article later.

First of all, look at the left-hand side of the header picture, in particular, the easel. On close inspection, you can see that the easel is the wrong way round and that the painter/android, is standing behind the easel. Midjourney produces four images by default, in the remarkable time of about 60 seconds which is almost like magic – indeed, in 1962, Arthur C. Clarke, a science fiction writer, stated in his book “Profiles of the Future: An Inquiry into the Limits of the Possible” that Any sufficiently advanced technology is indistinguishable from magic. So despite the apparent magic of creating these images so quickly, the AI has made a fundamental mistake that reveals that it doesn’t really understand what an easel is or how it should be used. Nafeez Ahmed is mostly talking about text generative interactions with AI – ChatGPT and the like, but what he says below, is equally applicable to images generated by AI…

The stunning capabilities of the new chatbots and their uncanny ability to engage in human-like conversations has sparked speculation about sentience. Yet there is zero tangible evidence of this beyond a sense of amazement.
What many miss is that the amazing nature of these conversations is not a function of an internal subjective ‘understanding’ by an identifiable entity, but rather the operation of algorithms made-up of hundreds of billions of parameters trained on massive datasets of human text to predict words that should follow given strings of text. {…} This is not a breakthrough in intelligence, although it is, certainly, a breakthrough in being able to synthesise human responses to similar questions and thereby mimic patterns in human interaction. This model of AI, therefore, cannot, in itself, generate fundamentally new knowledge or understanding – let alone sentience.

Nafeez Ahmed

Nafeez goes into great detail about how the research is headed in the wrong direction and indeed, how it is unlikely it is that it will ever succeed in equating to human sentience, so if you want to put your mind at rest about a Terminator-style future in which humans are subjugated by machines – nip on over and read the full article. Meanwhile I am going to show you some more examples of how Midjourney gets things “wrong” and how to get the “right” results and what that says about how useful such a programme can be.

You interact with the Midjourney app by sending it a message (just as if it was really an intelligent entity) containing a prompt, and once you receive your four images, you can choose one to enlarge, if you are satisfied with it, or run variations on one or all of them. Here is the prompt that produced the above set of images. “Silver android painting at an easel by Roy Lichtenstenstein” – the AI places most importance on the object at the beginning of the prompt, then on the activity described and lastly, it attempts to produce the images, in this case, in the style of the Pop Artist Roy Lichtenstein – famous for painting s in the style of close-ups of comic book pictures. These close-ups show the dot screens that were used to shade the illustrations of the comic book plus the hard black outlining and Midjourney has picked up well on these style features, particularly the top right and bottom left pictures. The top-left shows a sculpture vaguely resembling an up-cycled easel made of silver and the bottom right shows a silver-skinned figure with dot-screen effect, holding a brush and painting a picture but with no easel. In the [op-right picture, the top of the easel is just showing in the bottom corner and the android “artist” is holding a small blank canvas in her hand and drawing on it. Having seen the header image at top, and these pictures were as near as I could get to what I wanted, from multiple attempts, you can see that what I wanted was an all-over silver-skinned android and in the images above, top-right has a human face although “her” body is robotic – perhaps cyborg is a better description, whilst the other pictures show a sculpture, a woman and a totally abstract figure. So I decided to change the prompts to “Robot” rather than “Android” which produced better results. The reason I had started with “Andriod” was because robots range from automatic hoovers that move around your rooms looking like little miniature flying saucers sucking up dirt to more anthropomorphic devices – which is what I wanted.

“standing silver robot painting at an easel by Roy Lichtenstein” produced(among others) the above image in which the robot, possibly standing, is grasping what looks like the top of an easel but the “painting” does not appear to be on the easel. So I tried “Robot standing painting at an easel” and got this rather cute robot who looks like he is sitting on an invisible chair – “Hey Midjourney” just because you don’t show the chair, doesn’t make it standing!” Notice that with the style reference to Roy Lictensten gone, this image is very different. I would like to show you more of the iterations but Midjourney froze and when I reinstalled it, it had lost the entire session of work – you just can’t get the staff…

Another thing that I have discovered in my experiments, is that both Midjourney and ChatGPT, like to add unspecified embellishments – remember in my first report, how ChatGPT found the correct explanation for the phrase “Cold enough to freeze the balls off a brass monkey” but then added a totally made up explanation? Well Midjourney does the same thing too. Here is a picture of the railway viaduct at Knaresborough in West Yorkshire , an hours drive from where I live.

I wanted to see if Midjourney could produce a collage image using fragments of maps which it tried but didn’t really understand the concept – although I am not saying that it can’t, but at the very least, my prompt wasn’t sufficient (one of the oldest sayings amongst computer programmers is “Garbage in – garbage out!”) Here is Midjourneys best effort…

There are some map elements and the whole scene has been chopped up and rearranged but not in a way that makes sense – this one is closer to the real view…

But my first attempt, before I added the collage style, was simply to see how Midjourney would find and represent the viaduct and it generated the four images below. The top left image, Midjourney has added another railway running beneath the viaduct, likewise, lower-left it has added a canal and in the images on the right, Midjourney has transported us into a past sans Knaresborough and a post apocalyptic future where vegetation is growing over the viaduct.

Enough with all the pretty pictures – what does all this reveal about the way that the AI Midjourney works! Referring to the work – Erik J. Larson in his book, The Myth of Artificial Intelligence, Nafeez Ahmed cites a summary of the work by Ben Chugg, lead research analyst at Stanford University (Iknow – quotes within quotes) as follows:-

“Larson points out that current machine learning models are built on the principle of induction: inferring patterns from specific observations or, more generally, acquiring knowledge from experience. This partially explains the current focus on ‘big-data’ — the more observations, the better the model. We feed an algorithm thousands of labelled pictures of cats, or have it play millions of games of chess, and it correlates which relationships among the input result in the best prediction accuracy. Some models are faster than others, or more sophisticated in their pattern recognition, but at bottom they’re all doing the same thing: statistical generalization from observations.
This inductive approach is useful for building tools for specific tasks on well-defined inputs; analyzing satellite imagery, recommending movies, and detecting cancerous cells, for example. But induction is incapable of the general-purpose knowledge creation exemplified by the human mind.”

https://towardsdatascience.com/the-false-philosophy-plaguing-ai-bdcfd4872c45

Nafeez goes on:-

Current AI has become proficient at both deductive and inductive inference, with the latter becoming a primary focus.
Larson points out that human intelligence is based on a far more creative approach to generating knowledge called ‘abduction’. Abductive inference allows us to creatively select and test hypotheses, quickly eliminate the ones which are proven wrong, and create new ones as we go along before reaching a reliable conclusion. “We guess, out of a background of effectively infinite possibilities, which hypotheses seem likely or plausible,” writes Larson in The Myth of Artificial Intelligence. {…}
And here is Larson’s killer diagnosis: We don’t have a good theory of how abductive inference works in the human mind, and we have no idea how to recreate abductive inference for AI: “We are unlikely to get innovation if we choose to ignore a core mystery rather than face it up,” he writes with reference to the mystery of human intelligence.
Before we can generate genuine artificial intelligence that approaches human capabilities, we need a philosophical and scientific revolution that explains abductive inference. “As long as we keep relying on induction, AI programs will be forever prediction machines hopelessly limited by what data they are fed”, explains Chugg

https://www.bylinesupplement.com/p/why-the-ai-doomers-are-wrong?utm_source=substack&utm_medium=email

To relate this back to my experiments with Midjourney, the AI could identify what an easel looked like and include it in an image but it didn’t really know what it was or how it was used. Easels are probably present in thousands of pictures of artist’s studios as well as adverts but I bet there isn’t a Painters 101 that “First you will need an easel and this is how you use it” because when a human artist goes into a studio and sees canvasses fixed into easels, even if he has never seen one before, he is there to paint canvasses and it is obvious what they are and what they are for. It might be obvious to a human being with our ability to use inference, deduction and abductive capabilities, but an AI might identify an easel, but without finding a clear description of the usage, it cannot fully fathom how to use it…

As for the the tendency to add extraneous details, well the algorithms that govern Generative AI’s, are designed to mimic human conversational style, so when it has found a relevant answer to the requested information or task, it looks to extend the conversation in what it has learned might possibly follow – it doesn’t know whether it is true or not, because that is way above it’s paygrade ( a metaphor which it probably wouldn’t understand either). This phenomena of AI’s making things up is called hallucination – a very anthropocentric term…

I will make one more report on my attempts to get exactly what I wanted from Midjourney and how I found a compromise to be able to work with the results…

R – Robot – robota – Czech for forced labour – “Foreign” words appropriated – Rule of Thumb…

The origin of the word Robot, is the Czech word robota, meaning “forced labour”, from a Slavic root – rab, meaning “slave”. Herein lies much of our fear and angst when we consider the future of robots because Slave implies a Master and so slaves are capable of revolt – of turning against their masters…

The pursuit of developing robots, is that they might assist humans in doing jobs which are too difficult, dangerous or just too plain boring for humans – things which in the past, and even today – (think wage-slaves, modern slavery) – have been done by human slaves and so these fears have a foundation in fact – there have been many slave revolts!

A quote from the play that coined the word “robot”. The author said – “The product of the human brain has escaped the control of human hands. This is the comedy of science.”

The problem of creating robots is twofold – body, and mind. On the body side, we have long had Automatons – they range from say a music box which can play a tune, to the most sophisticated machines that are now being tested for their ability to play football. On the mind side, we have the quest for AI – Artificial Intelligence which is the subject of hot debate at present for reasons varying from “Will AI take our jobs away?” to “Will AI outgrow and destroy human beings?” which brings us back to the man who first coined the word robot in his play “R.U.R.” (which stands for “Rossum’s Universal Robots”) premiered in Prague in 1921, Karel Čapek. Like many of his generation, just out of the horrors of the First World War, Čapek was sceptical of the utopian benefits of science and technology – or rather the uses which human beings put those things. You can read a more detailed account of his play here. But what “R.U.R.” illustrates is that science fiction is the way we explore the possibilities and problems of what may be achieved in the development of robots and our relationship to them.

Starting with the body problem, long before the amazingly intricate creations of the 18th century with watchmaker ingenuity inside, the ancient Egyptians imbued statues with souls and the Greeks envisaged artificial men such as Talos But those later amazing mechanical figures who might play a tune on an inbuilt musical box, are only built to perform one task, albeit a potentially complex task and in this respect, modern technology has created many robots which assist us today without posing any threat except to the workers they superseded. Car plants use many robot arms to manufacture cars with greater strength, dexterity (programmed), speed and accuracy than the human beings who used to do the jobs. Still, without reprogramming, these robot arms do one thing only and their “intelligence” is limited to a programmable computer.

Automaton in the Swiss Museum CIMA. Rama, CC BY-SA 2.0 FR https://creativecommons.org/licenses/by-sa/2.0/fr/deed.en, via Wikimedia Commons

Turning to the issue of robotic minds – Artificial Intelligence is progressing in leaps and bounds, to use an anthropocentric metaphor – several people doing this year’s A to Z Challenge have experimented with AI including Misky who has used the AI graphic app Midjourney to create amazing illustrations for the poems she has posted – check out her site! She has talked about how many times she has to try and prompt the app in order to get these illustrations the way she wants them, tweaking the style and content descriptors and this shows that although the Midjourney app is incredibly powerful, it is still soft AI. The limitations of soft AI are best illustrated by Alexa the Amazon speaker app – instruct it thus “Alexa – tell me a joke!” and she will indeed tell you a joke, but then say “Alexa – tell me another…” and Alexa doesn’t know what you mean unless you specify “another joke”. Hard AI would be the kind which was truly capable of independently sentient thought, and we are some way off from that if Alexa is anything to go by. Alan Turing, imagining (stupendously) the future possibilities of AI devised the Turing Test in which an evaluator would hold two remote conversations, with a human being and with an AI and if they could not distinguish which was the AI, then it might be said to have truly intelligent behaviour. We may be approaching this watershed moment but I like to think that, writing this blog, for example, an AI would not be drawing out the ideas that I have – at least not without close supervision – anyway, I am going to do my own evaluative exploration of AI right here once the A to Z is finished, so watch this space…

The problem for designing robots which are indistinguishable from human beings, is how to cram an AI sizes computer into the body of a robot – Chat GPT can generate text which only another special app can discern to have been written by an AI, but the computers necessary to run ChatGPT are enormous and the miracle which is the human brain is most notable for its compact size – given it’s power. Nevertheless, there are many who are afraid that AI alone, without human-looking robots, can outmatch the human race and destroy it. Of course, we have many science fiction thought experiments to thank for that particular trope from Čapek’s “R.U.R.”, Fritz Lang’s “Metropolis”, through Asimov’s “I Robot”, My own favourite, “Bladerunner” and looming large in these debates about the dangers of rampant AI – the “Terminator” series of films.

(From left) Alfred Abel, Brigitte Helm, and Rudolf Klein-Rogge in Metropolis, directed by Fritz Lang, 1927. From a private collection.

In all these fictional considerations of the relationship between men and machines, different solutions are proposed to keeping the “robota -slaves” in check. Asimov came up with The Three Laws of Robotics

  • First Law: A robot may not injure a human being, or, through inaction, allow a human being to come to harm.
  • Second Law: A robot must obey orders given it by human beings, except where such orders would conflict with the First Law.
  • Third Law: A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

These laws would be encoded into the AI controlling all robots at their inception and would be sacrosanct. This idea escaped the creators of the Terminator series robots whilst in Bladerunner, the manufacturers of the “replicants” had to build in auto-destruction of their products after a small number of years lest their self-learning robots get too big for their boots and turn on their creators – sensibly too, replicants were not allowed on Earth but only sent to do the typical jobs of slaves, out in space as miners, soldiers, builders – dirty, dangerous jobs. Ultimately, the Replicants in Philip K. Dick’s story, show us something about what it is to be human ( as all good science fiction does, if only because it is written by humans), at the end of the story, (spoiler alert) the human Bladerunner, charged with tracking down and destroying renegade Replicants who have made it to Earth in pursuit of getting their lifespans extended, is being dangled over the edge of a roof by the leader of the group of Replicants. Rutger Hauer, the actor who plays The Replicant Roy Batty, whilst dangling the Bladerunner, makes a speech which has become known as the ”Tears in the Rain monologue” as follows:

“I’ve seen things… seen things you little people wouldn’t believe. Attack ships on fire off the shoulder of Orion bright as magnesium… I rode on the back decks of a blinker and watched C-beams glitter in the dark near the Tannhäuser Gate. All those moments… they’ll be gone.”

Moments later he lifts the Bladerunner to safety just before his brain self-destructs – talk about saved by the bell! But of course, the Bladerunner -played by Harrison Ford, was saved by the robot/ replicant, perhaps out of mercy (a very human trait which the replicant might have developed) but perhaps in the hope that the Bladerunner might respond by trying to change things for Replicant To me this gets to the crux of being human – just when we are getting the hang of life, out time is up…

Incidentally, Rutger Hauer, not liking the lines that had been written for him, and with only a few minutes of shooting time left, famously improvised the “Tears in the Rain” monologue – kudos!

In our fears over the development of robots and AI, are we perhaps projecting our fears about our own human traits onto them, we aspired to create them to be our helpers, such as AI interpreting MRI scans more effectively than human beings, so why would we think that AI would want to destroy or enslave human beings (The Matrix)? Perhaps it’s because we human beings, given the chance, have all too often enslaved other human beings, abused, exploited and been ready to lynch them at the first sign of independence, let alone revolt. Do we need to take heed of the results of science fiction’s thought experiment warnings – of course we do! We should no more allow the unregulated development of AI any more than we would allow our children to play with loaded guns, but those regulations would be to keep in check the humans who would use AI, or rather misuse it. We should watch out for governments who want to create AI-powered weapons or control their people with ever more efficient propaganda – even before AI we are struggling to know what is true in the news. We have plenty to worry about in the humans – let alone AI and Robots. However, between regulation, commonsense, and perhaps most of all. the fact that AI might have learned lessons from the human mistakes that are messing up the planet, might we be pleasantly surprised to find that AIs assist us as they were intended to do, solving problems of the environment, working out how to operate an economy not based on permanent growth and war – might we be headed for Iain Banks’ “Culture” series rather than “Terminator”…

So one final sci-fi thought experiment from the great Marge Piercy – “He, She and It”, which goes to the question of what would happen if we could achieve hard AI and make it small enough to fit into an android-style (human shaped) robot. In an environmentally post-apocalyptic world, a woman is given the task of socialising just such an android, because Marge Piercy imagines that true sentience, as opposed to a glorified search engine, would require teaching and guiding as to what sense to make of the world – rather like a child. If she is right, then the great leap forward to hard AI, true sentience, would not be a runaway Terminator scenario, but a chance to imbueAI with the best qualities of human beings rather than the worst. Running through the book is the story of The Golem – a man made from mud and brought to”life” by Cabbalistic magic in order to protect the ghetto and in this story is that familiar warning of the dangers inherent in creating powerful androids which has echoed through speculative fiction ever since Mary Shelley’s “Frankenstein”… “He, She and It” is such an amazing book that I am surprised that it has not been made into a movie…

So what do you believe will be the future of AI and Robots -slave, nemesis of the human race or willing and able helper?

And so lastly, to – as pernicious a law as humans could devise, The Rule of Thumb – another phrase that has disputed meanings. Some people imagine the Rule of Thumb to describe a readily available way of measuring things – an Inch is approximately equivalent to the breadth of the top joint of an adult thumb, but a ruling by Judge Francis Buller in 1782, allows that a man could legally beat his wife, as long as he used a stick that was no thicker than his thumb – see here. I suspect that those who favour the measuring theory and disparage the beating one, are men…

This Gillray cartoon of Judge Buller from 1782 shows ‘Judge Thumb’ selling sticks for wife-beating Bridgeman

The Cant languages from the Wikipedia article for the letter “R” are: