Leary Philosophers
Can a language model outperform old Edward Lear in describing philosophers in Limerick form? “There once was a Scotsman named Hume…”
Can a language model outperform old Edward Lear in describing philosophers in Limerick form? “There once was a Scotsman named Hume…”
Will large language models acknowledge authorship of their own generated texts? Will a language model claim authorship or ownership of texts which it did not create? A mistaken comprehension of ChatGPT’s abilities throws up a distinctive problem of intellectual property rights.
What is the relationship between perplexity, creativity and novelty? Following on from ‘Perplexing Perplexity’, I set out to demonstrate that high perplexity texts are not always creative, and to showcase ChatGPT’s ability to work with and even generate novel words – culminating in the Tale of Zonkamoozle.
Detectors such as GPTZero use the property of ‘perplexity’ to try to detect AI authorship of texts. But I show that by writing in a specifically dull style, or engineering the prompt given to a language model, we can easily and systematically fool such detectors to label AI text as human and vice versa.
A recent study by Gao et al. (2022) validates the warning of ‘Machine Evidence’ (Blunt, 2019) that language models would soon become capable of beating detection attempts by human peer reviewers. This piece looks at the near-term steps that journal editors and conference organisers can take to prevent AI-generated abstracts bypassing their screening processes, along with a warning for the long-term viability of those strategies.
Can a new iteration of GPT-3 write pop songs, raps and limericks…about Immanuel Kant’s Categorical Imperative? It is a moral duty to find out.
DALLE 2 offers a far more powerful image generation AI than the popular open access ‘Craiyon’/’DALLE Mini’ model. How does DALLE 2 compare to DALLE Mini’s visions of hierarchies and pyramids of evidence?
How does a machine learning algorithm picture hierarchies of evidence and evidence-based medicine – and what do these visions of evidence remind us of the way we understand, order and assemble the information we use to guide clinical practice?
AI21 Labs have just released a public demo of their giant language model, Jurassic-1. At 178bn parameters, it rivals GPT-3. Feeding it my own work, it generated some interesting and potentially novel views on evidence hierarchies… and then attributed them to CAM researcher Marc Micozzi! Is Jurassic Micozzi’s critique of evidential pluralism in medicine sound?
In their 2020 paper, Floridi and Chiriatti subject giant language model GPT-3 to three tests: mathematical, semantic and ethical. I show that these tests are misconfigured to prove the points Floridi and Chiriatti are trying to make. We should attend to how such giant language models function to understand both their responses to questions and the ethical and societal impacts.