hallucination problem

back to index

11 results

The Singularity Is Nearer: When We Merge with AI
by Ray Kurzweil
Published 25 Jun 2024

BACK TO NOTE REFERENCE 160 For more information on the problems LLMs have with hallucination, see Tom Simonite, “AI Has a Hallucination Problem That’s Proving Tough to Fix,” Wired, March 9, 2018, https://www.wired.com/story/ai-has-a-hallucination-problem-thats-proving-tough-to-fix; Craig S. Smith, “Hallucinations Could Blunt ChatGPT’s Success,” IEEE Spectrum, March 13, 2023, https://spectrum.ieee.org/ai-hallucination; Cade Metz, “What Makes A.I. Chatbots Go Wrong?,” New York Times, March 29, 2023 (updated April 4, 2023), https://www.nytimes.com/2023/03/29/technology/ai-chatbots-hallucinations.html; Ziwei Ji et al., “Survey of Hallucination in Natural Language Generation,” ACM Computing Surveys 55, no. 12, article 248 (March 3, 2023): 1–38, https://doi.org/10.1145/3571730.

pages: 336 words: 91,806

Code Dependent: Living in the Shadow of AI
by Madhumita Murgia
Published 20 Mar 2024

When the opposing counsel had challenged the cases cited, Schwartz went back to ChatGPT, but it doubled down and ‘lied’ to him, his lawyer said. Schwartz, his voice breaking, told the judge that he was ‘embarrassed, humiliated and extremely remorseful.’ ChatGPT and all other conversational AI chatbots have a disclaimer that warns users about the hallucination problem, pointing out that large language models sometimes make up facts. ChatGPT, for instance, has a warning on its webpage: ‘ChatGPT may produce inaccurate information about people, places, or facts.’ Judge Castel: Do you have something new to say? Schwartz’s lawyer: Yes.

Through years of writing about the technology, the pattern that has emerged for me is the extent of the impact of AI on society’s marginalized and excluded groups: refugees and migrants, precarious workers, socioeconomic and racial minorities, and women. These same groups are disproportionately affected by generative AI’s technical limitations too: hallucinations and negative stereotypes perpetuated in the software’s text and image outputs.7 And it’s because they rarely have a voice in the echo chambers in which AI is being built. It was why I had chosen to narrate the perspectives of people outside of Silicon Valley – those whose views are so often ignored in the design or implementation of new technologies like AI.

.: ‘The Machine Stops’ ref1 Fortnite ref1 Foxglove ref1 Framestore ref1 Francis, Pope ref1, ref2 fraudulent activity benefits ref1 gig workers and ref1, ref2, ref3 free will ref1, ref2 Freedom of Information requests ref1, ref2, ref3 ‘Fuck the algorithm’ ref1 Fussey, Pete ref1 Galeano, Eduardo ref1 gang rape ref1, ref2 gang violence ref1, ref2, ref3, ref4 Gebru, Timnit ref1, ref2, ref3 Generative Adversarial Networks (GANs) ref1 generative AI ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9, ref10 AI alignment and ref1, ref2, ref3 ChatGPT see ChatGPT creativity and ref1, ref2, ref3, ref4 deepfakes and ref1, ref2, ref3 GPT (Generative Pre-trained Transformer) ref1, ref2, ref3, ref4 job losses and ref1 ‘The Machine Stops’ and ref1 Georgetown University ref1 gig work ref1, ref2, ref3, ref4, ref5 Amsterdam court Uber ruling ref1 autonomy and ref1 collective bargaining and ref1 colonialism and ref1, ref2, ref3 #DeclineNow’ hashtag ref1 driver profiles ref1 facial recognition technologies ref1, ref2, ref3, ref4 fraudulent activity and ref1, ref2, ref3, ref4 ‘going Karura’ ref1 ‘hiddenness’ of algorithmic management and ref1 job allocation algorithm ref1, ref2, ref3, ref4, ref5, ref6 location-checking ref1 migrants and ref1 ‘no-fly’ zones ref1 race and ref1 resistance movement ref1 ‘slaveroo’ ref1 ‘therapy services’ ref1 UberCheats ref1, ref2, ref3 UberEats ref1, ref2 UK Supreme Court ruling ref1 unions and ref1, ref2, ref3 vocabulary to describe AI-driven work ref1 wages ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9, ref10, ref11 work systems built to keep drivers apart or turn workers’ lives into games ref1, ref2 Gil, Dario ref1 GitHub ref1 ‘give work, not aid’ ref1 Glastonbury Festival ref1 Glovo ref1 Gojek ref1 ‘going Karura’ ref1 Goldberg, Carrie ref1 golem (inanimate humanoid) ref1 Gonzalez, Wendy ref1 Google ref1 advertising and ref1 AI alignment and ref1 AI diagnostics and ref1, ref2, ref3 Chrome ref1 deepfakes and ref1, ref2, ref3, ref4 DeepMind ref1, ref2, ref3, ref4 driverless cars and ref1 Imagen AI models ref1 Maps ref1, ref2, ref3 Reverse Image ref1 Sama ref1 Search ref1, ref2, ref3, ref4, ref5 Transformer model and ref1 Translate ref1, ref2, ref3, ref4 Gordon’s Wine Bar London ref1 GPT (Generative Pre-trained Transformer) ref1, ref2, ref3, ref4 GPT-4 ref1 Graeber, David ref1 Granary Square, London ref1, ref2 ‘graveyard of pilots’ ref1 Greater Manchester Coalition of Disabled People ref1 Groenendaal, Eline ref1 Guantanamo Bay, political prisoners in ref1 Guardian ref1 Gucci ref1 guiding questions checklist ref1 Gulu ref1 Gumnishka, Iva ref1, ref2, ref3, ref4 Gutiarraz, Norma ref1, ref2, ref3, ref4, ref5 hallucination problem ref1, ref2, ref3 Halsema, Femke ref1, ref2 Hanks, Tom ref1, ref2 Hart, Anna ref1 Hassabis, Demis ref1 Harvey, Adam ref1 Have I Been Trained ref1 healthcare/diagnostics Accredited Social Health Activists (ASHAs) ref1, ref2, ref3 bias in ref1 Covid-19 and ref1, ref2 digital colonialism and ref1 ‘graveyard of pilots’ ref1 heart attacks and ref1, ref2 India and ref1 malaria and ref1 Optum ref1 pain, African Americans and ref1 qTrack ref1, ref2, ref3 Qure.ai ref1, ref2, ref3, ref4 qXR ref1 radiologists ref1, ref2, ref3, ref4, ref5, ref6 Tezpur ref1 tuberculosis ref1, ref2, ref3 without trained doctors ref1 X-ray screening and ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9, ref10 heart attacks ref1, ref2 Herndon, Holly ref1 Het Parool ref1, ref2 ‘hiddenness’ of algorithmic management ref1 Hikvision ref1, ref2 Hinton, Geoffrey ref1 Hive Micro ref1 Home Office ref1, ref2, ref3 Hong Kong ref1, ref2, ref3, ref4, ref5 Horizon Worlds ref1 Hornig, Jess ref1 Horus Foundation ref1 Huawei ref1, ref2, ref3 Hui Muslims ref1 Human Rights Watch ref1, ref2, ref3, ref4 ‘humanist’ AI ethics ref1 Humans in the Loop ref1, ref2, ref3, ref4 Hyderabad, India ref1 IBM ref1, ref2, ref3, ref4 Iftimie, Alexandru ref1, ref2, ref3, ref4, ref5 IJburg, Amsterdam ref1 Imagen AI models ref1 iMerit ref1 India ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9 facial recognition in ref1, ref2, ref3 healthcare in ref1, ref2, ref3 Industrial Light and Magic ref1 Information Commissioner’s Office ref1 Instacart ref1, ref2 Instagram ref1, ref2 Clearview AI and ref1 content moderators ref1, ref2, ref3, ref4 deepfakes and ref1, ref2, ref3 Integrated Joint Operations Platform (IJOP) ref1, ref2 iPhone ref1 IRA ref1 Iradi, Carina ref1 Iranian coup (1953) ref1 Islam ref1, ref2, ref3, ref4, ref5 Israel ref1, ref2, ref3 Italian government ref1 Jaber, Faisal bin Ali ref1 Jainabai ref1 Janah, Leila ref1, ref2, ref3 Jay Gould, Stephen ref1 Jewish faith ref1, ref2, ref3, ref4 Jiang, Mr ref1 Jim Crow era ref1 jobs application ref1, ref2, ref3 ‘bullshit jobs’ ref1 data annotation and data-labelling ref1 gig work allocation ref1, ref2, ref3, ref4, ref5, ref6 losses ref1, ref2, ref3 Johannesburg ref1, ref2 Johnny Depp–Amber Heard trial (2022) ref1 Jones, Llion ref1 Joske, Alex ref1 Julian-Borchak Williams, Robert ref1 Juncosa, Maripi ref1 Kafka, Franz ref1, ref2, ref3, ref4 Kaiser, Lukasz ref1 Kampala, Uganda ref1, ref2, ref3 Kellgren & Lawrence classification system. ref1 Kelly, John ref1 Kibera, Nairobi ref1 Kinzer, Stephen: All the Shah’s Men ref1 Knights League ref1 Koli, Ian ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9, ref10 Kolkata, India ref1 Koning, Anouk de ref1 Laan, Eberhard van der ref1 labour unions ref1, ref2, ref3, ref4, ref5, ref6 La Fors, Karolina ref1 LAION-5B ref1 Lanata, Jorge ref1 Lapetus Solutions ref1 large language model (LLM) ref1, ref2, ref3 Lawrence, John ref1 Leigh, Manchester ref1 Lensa ref1 Leon ref1 life expectancy ref1 Limited Liability Corporations ref1 LinkedIn ref1 liver transplant ref1 Loew, Rabbi ref1 London delivery apps in ref1, ref2 facial recognition in ref1, ref2, ref3, ref4 riots (2011) ref1 Underground terrorist attacks (2001) and (2005) ref1 Louis Vuitton ref1 Lyft ref1, ref2 McGlynn, Clare ref1, ref2 machine learning advertising and ref1 data annotation and ref1 data colonialism and ref1 gig workers and ref1, ref2, ref3 healthcare and ref1, ref2, ref3 predictive policing and. ref1, ref2, ref3, ref4 rise of ref1 teenage pregnancy and ref1, ref2, ref3 Mahmoud, Ala Shaker ref1 Majeed, Amara ref1, ref2 malaria ref1 Manchester Metropolitan University ref1 marginalized people ref1, ref2, ref3, ref4, ref5, ref6, ref7, ref8, ref9 Martin, Noelle ref1, ref2, ref3, ref4, ref5, ref6, ref7 Masood, S.

pages: 444 words: 117,770

The Coming Wave: Technology, Power, and the Twenty-First Century's Greatest Dilemma
by Mustafa Suleyman
Published 4 Sep 2023

In AI, technical safety also means sandboxes and secure simulations to create provably secure air gaps so that advanced AIs can be rigorously tested before they are given access to the real world. It means much more work on uncertainty, a major focus right now—that is, how does an AI communicate when it might be wrong? One of the issues with LLMs is that they still suffer from the hallucination problem, whereby they often confidently claim wildly wrong information as accurate. This is doubly dangerous given they often are right, to an expert level. As a user, it’s all too easy to be lulled into a false sense of security and assume anything coming out of the system is true.

Brian, 56 artificial capable intelligence (ACI), vii, 77–78, 115, 164, 210 artificial general intelligence (AGI) catastrophe scenarios and, 209, 210 chatbots and, 114 DeepMind founding and, 8 defined, vii, 51 gorilla problem and, 115–16 gradual nature of, 75 superintelligence and, 75, 77, 78, 115 yet to come, 73–74 artificial intelligence (AI) aspirations for, 7–8 autonomy and, 114, 115 as basis of coming wave, 55 benefits of, 10–11 catastrophe scenarios and, 208, 209–11 chatbots, 64, 68, 70, 113–14 Chinese development of, 120–21 choke points in, 251 climate change and, 139 consciousness and, 74, 75 contradictions and, 202 costs of, 64, 68 current applications, 61–62 current capabilities of, 8–9 cyberattacks and, 162–63, 166–67 defined, vii early experiments in, 51–54 efficiency of, 68–69 ego and, 140 ethics and, 254 explanation and, 243 future of, 78 future ubiquity of, 284–85 global reach of, 9–10 hallucination problem and, 243 human brain as fixed target, 67–68 hyper-evolution and, 109 invisibility of, 73 limitations of, 73 medical applications, 110 military applications, 104, 165 Modern Turing Test, 76–77, 78, 115, 190, 210 narrow nature of, 73–74 near-term capabilities, 77 omni-use technology and, 111, 130 openness imperative and, 128–29 potential of, 56, 70, 135 as priority, 60 profit motive and, 134, 135, 136 proliferation of, 68–69 protein structure and, 88–89 red teaming and, 246 regulation attempts, 229, 260–61 research unpredictability and, 130 robotics and, 95, 96, 98 safety and, 241, 243–44 scaling hypothesis, 67–68, 74 self-critical culture and, 270 sentience claims, 72, 75 skepticism about, 72, 179 surveillance and, 193–94, 195, 196 synthetic biology and, 89–90, 109 technological unemployment and, 177–81 Turing test, 75 See also coming wave; deep learning; machine learning arXiv, 129 Asilomar principles, 269–70, 272–73 ASML, 251 asymmetrical impact, 105–7, 234 Atlantis, 5 Atmanirbhar Bharat program (India), 125–26 attention, 63 attention maps, 63 audits, 245–48, 267 Aum Shinrikyo, 212–13, 214 authoritarianism, 153, 158–59, 191–96, 216–17 autocomplete, 63 automated drug discovery, 110 automation, 177–81 autonomy, 105, 113–15, 166, 234 Autor, David, 179 al-Awlaki, Anwar, 171 B backpropagation, 59 bad actor empowerment, 165–66, 208, 266 See also terrorism B corps, 258 Bell, Alexander Graham, 31 Benz, Carl, 24, 285 Berg, Paul, 269–70 BGI Group, 122 bias, 69–70, 239–40 Bioforge, 86 Biological Weapons Convention, 241, 263 biotech.

See climate change Go, 53–54, 113, 117–19, 120 Google corporate power of, 187 DeepMind purchase, 60, 255–57 efficiency and, 68 LaMDA and, 71, 72 large language models and, 66 quantum computing and, 97–98, 122 robotics and, 95 on transformers, 64 Google Scholar, 128 Gopher, 68 gorilla problem, 115–16 governments containment and, 258–63 organizational limitations of, 148–50 See also nation-states GPS (Global Positioning System), 110 GPT-2, 64, 70 GPT-3, 64, 68 GPT-4, 64, 113–14 GPUs, 130, 251 grand bargain, defined, viii Great Britain corporations and, 186, 189 surveillance, 193, 195–96 great power competition. See geopolitics Gutenberg, Johannes, 30, 35 H H1N1 flu, 173–74 hallucination problem, 243 Harvard Wyss Institute, 95 Hassabis, Demis, 8 health care. See medical applications Henrich, Joseph, 28 Heritage Foundation, 257 Hershberg, Elliot, 87 Hezbollah, 196–97 Hidalgo, César, 108 hierarchical planning, 76–77 Hinton, Geoffrey, 59, 60, 130 Hiroshima/Nagasaki bombings, 41–42 Hobbes, Thomas, 216 Homo technologicus, 6 Hugging Face, 199 Human Genome Project, 80–81 Huskisson, William, 131 Hutchins, Marcus, 161 hyper-evolution, 105, 107–9 chips and, 32–33, 57, 81, 108 containment and, 250 large language models and, 66, 68 I India, 125–26, 169–70 Industrial Revolution containment attempts, 39, 40, 281–83 openness imperative and, 127 profit motive and, 133, 134 technology waves and, 28–29 inertial confinement, 100 Inflection AI, 66, 68, 243, 244 information dematerialization and, 55–56 DNA as, 79, 87–88 Institute of Electrical and Electronics Engineers, 241 integrated circuit, 32 intelligence action and, 75–76 corporations and, 186–87 economic value of, 136 gorilla problem, 115–16 prediction and, 62 See also artificial intelligence interconnectedness, 28 Intergovernmental Panel on Climate Change, 138–39 internal combustion engine, 24–25, 26, 35–36 International Atomic Energy Agency, 241 international cooperation, 263–67 internet, 33, 107–8, 202 iPhone, 187 Iran, 165 Israel, 165 J James, Kay Coles, 257 Japan, containment attempts, 39, 40 jobs, technology impact on, 177–81, 261, 262 Joint European Torus, 100 K Kasparov, Garry, 53 Kay, John, 39 Ke Jie, 118–19, 121 Kennan, George F., 37 Keynes, John Maynard, 178 Khan, A.

pages: 321 words: 113,564

AI in Museums: Reflections, Perspectives and Applications
by Sonja Thiel and Johannes C. Bernhardt
Published 31 Dec 2023

The constant work of removing racist and crucial or sensitive content from foundation models is also pursued under often neo-colonial work conditions, which can be analysed through the ‘data-production dispositif’ (Miceli/Posada 2022). Any museum using a foundation model for data-related work needs to be aware of the conditions of its production, as well as the options for adjustments and integration into a specific product and the range of end-user scenarios, particularly against the backdrop of the so-called hallucination problem. One solution in which museums might contribute within their field of expertise and enhance their data with AI is the field of language sensitivity, as explored in the development of Sabio, a tool designed to detect biases in the metadata of museum collections.4 Another interesting option would be to build alliances within the cultural heritage community in order to build our own models trained on heritage data.

As stated there and also demanded by museum users, AI-generated content should be labelled as such; the training sources and finetuning of them should be made transparent; copyrighted material should be specially marked and excluded from training processes or foundation models; and the rights of artists and photographers should be protected. The hallucination problem, that is, the generation of information based not on facts but instead on the output of a statistical language model, can be highlighted as an existing problem, but nonetheless be made use of experimentally or creatively until better solutions are provided by research and development. Many people have already incorporated language models into their daily lives for improving texts, structuring presentations, writing speeches, or generating code.

pages: 848 words: 227,015

On the Edge: The Art of Risking Everything
by Nate Silver
Published 12 Aug 2024

pages: 688 words: 147,571

Robot Rules: Regulating Artificial Intelligence
by Jacob Turner
Published 29 Oct 2018

Hart, Punishment and Responsibility: Essays in the Philosophy of Law (Oxford: Clarendon Press, 1978). 130Carlsmith and Darley, “Psychological Aspects of Retributive Justice”, in Advances in Experimental Social Psychology, edited by Mark Zanna (San Diego, CA: Elsevier, 2008). 131In evidence to the Royal Commission on Capital Punishment, Cmd. 8932, para. 53 (1953). 132Exodus 21:24, King James Bible. 133John Danaher, “Robots, Law and the Retribution Gap”, Ethics and Information Technology, Vol. 18, No. 4 (December 2016), 299–309. 134Recent experiments conducted by Zachary Mainen involving the use of the hormone serotonin on biological systems may provide one avenue for future AI to experience emotions in a similar manner to humans. See Matthew Hutson, “Could Artificial Intelligence Get Depressed and Have Hallucinations?”, Science Magazine, 9 April 2018, http://​www.​sciencemag.​org/​news/​2018/​04/​could-artificial-intelligence-get-depressed-and-have-hallucinations, accessed 1 June 2018. 135In a gruesome example of public retribution being exacted against insensate “perpetrators”, in 1661 following the restoration of the English monarchy after the English Civil War and the rebublican Protectorate, three of the already deceased regicides who had participated in the execution of Charles I were disinterred from their graves and tried for treason.

Visual Thinking: The Hidden Gifts of People Who Think in Pictures, Patterns, and Abstractions
by Temple Grandin, Ph.d.
Published 11 Oct 2022

The odds of something offensive coming out is 100 percent.” AI applications are being developed for simulations and analytics, and in industry, transportation, cybersecurity, and the military. What are the failsafes? Would you want an AI program running a nuclear reactor? What if the AI operator started hallucinating because a hacker inserted a feedback loop that forced it to perceive the high pressures and temperatures of a meltdown that did not exist? Maybe it would create an actual meltdown. Some computer scientists will admit that they are not completely sure how AI works. In an article by Arthur I.

pages: 584 words: 170,388

Hyperion
by Dan Simmons
Published 15 Sep 1990

pages: 169 words: 41,887

Literary Theory for Robots: How Computers Learned to Write
by Dennis Yi Tenen
Published 6 Feb 2024

pages: 574 words: 164,509

Superintelligence: Paths, Dangers, Strategies
by Nick Bostrom
Published 3 Jun 2014

It could inspect each one, over and over, to reduce the risk that any of the paperclips fail to meet the design specifications. It could build an unlimited amount of computronium in an effort to clarify its thinking, in the hope of reducing the risk that it has overlooked some obscure way in which it might have somehow failed to achieve its goal. Since the AI may always assign a nonzero probability to having merely hallucinated making the million paperclips, or to having false memories, it would quite possibly always assign a higher expected utility to continued action—and continued infrastructure production—than to halting. The claim here is not that there is no possible way to avoid this failure mode.

pages: 1,028 words: 267,392

Wanderers: A Novel
by Chuck Wendig
Published 1 Jul 2019