ΑΙhub 04月22日 20:14
Why AI can’t take over creative writing
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了大型语言模型(LLM)如ChatGPT在创意写作中的应用。文章回顾了信息论创始人Shannon提出的概率语言模型,并讨论了ChatGPT的运作机制,它通过预测下一个词的概率来生成文本。作者区分了LLM的“鹦鹉式”输出与人类的创造力,并指出LLM在缺乏特定作者写作资料的情况下,难以产生真正具有创造力的作品。尽管如此,文章也提到了LLM在辅助写作方面的积极作用,如生成标准文本或作为大型项目的辅助工具,并强调了“提示工程”的重要性。文章最后呼吁读者在信息时代保持批判性思维,理解生成式AI的本质。

🗣️ **LLM的工作原理:** ChatGPT等LLM基于互联网上的大量文本数据进行训练,通过预测给定上下文(提示和已生成的词)中下一个词的概率来生成文本。这类似于从一个“词汇帽子”中随机抽取词语,高概率的词被选中的可能性更大。

🦜 **LLM的局限性:** 作者将LLM比作“鹦鹉”,它们重复训练数据中的内容,而非真正具有创造力。LLM难以捕捉特定作者的写作风格和意图,生成的文本可能只是随机组合,缺乏深度和独特性。

✍️ **LLM在写作中的应用:** LLM可以用于生成标准文本、辅助写作或作为大型项目的组成部分。然而,要充分利用LLM,需要生成多个输出并进行编辑,并掌握“提示工程”技巧,以获得更符合预期的结果。

💡 **人类的创造力:** 真正的创造力源于人类的思想和意图。作者认为,创意写作需要作者的思考和表达,而不仅仅是依赖LLM生成文本。重要的是,要区分LLM的输出与人类的创造力。

🧐 **批判性思考:** 在信息时代,保持批判性思维至关重要。对于生成式AI,我们需要理解其局限性,避免被过度宣传所误导。生成式AI没有魔法,只有大量数据,用于预测人们可能写的内容。

By David Poole, University of British Columbia

In 1948, the founder of information theory, Claude Shannon, proposed modelling language in terms of the probability of the next word in a sentence given the previous words. These types of probabilistic language models were largely derided, most famously by linguist Noam Chomsky: “The notion of ‘probability of a sentence’ is an entirely useless one.”

In 2022, 74 years after Shannon’s proposal, ChatGPT appeared, which caught the attention of the public, with some even suggesting it was a gateway to super-human intelligence. Going from Shannon’s proposal to ChatGPT took so long because the amount of data and computing time used was unimaginable even a few years before.

ChatGPT is a large language model (LLM) learned from a huge corpus of text from the internet. It predicts the probability of the next word given the context: a prompt and the previously generated words.

ChatGPT uses this model to generate language by choosing the next word according to the probabilistic prediction. Think about drawing words from a hat, where the words predicted to have a higher probability have more copies in the hat. ChatGPT produces text that seems intelligent.

There is a lot of controversy about how these tools can help or hinder learning and practising creative writing. As a professor of computer science who has authored hundreds of works on artificial intelligence (AI), including AI textbooks that cover the social impact of large language models, I think understanding how the models work can help writers and educators consider the limitations and potential uses of AI for what might be called “creative” writing.

LLMs as parrots or plagiarists

It’s important to distinguish between “creativity” by the LLM and creativity by a human. For people who had low expectations of what a computer could generate, it’s been easy to assign creativity to the computer. Others were more skeptical. Cognitive scientist Douglas Hofstadter saw “a mind-boggling hollowness hidden just beneath its flashy surface.”

Linguist Emily Bender and colleagues described the language models as stochastic parrots, meaning they repeat what is in the data they were trained on with randomness. To understand this, consider why a particular word was generated. It’s because it has a relatively high probability, and it has a high probability because a lot of text in the training corpus used that word in similar contexts.

Selecting a word according to the probability distribution is like selecting text with a similar context and using its next word. Generating text from LLMs can be seen as plagiarism, one word at a time.

The creativity of a human

Consider the creativity of a human who has ideas they want to convey. With generative AI, they put their ideas into a prompt and the AI will produce text (or images or sounds). If someone doesn’t care what is generated, it doesn’t really matter what they use as a prompt. But what if they do care about what is generated?

An LLM tries to generate what a random person who had written the previous text would produce. Most creative writers do not want what a random person would write. They want to use their creativity, and may want a tool to produce what they would write if they had the time to produce it.

LLMs don’t typically have a large corpus of what a particular author has written to learn from. The author will undoubtedly want to produce something different. If the output is expected to be more detailed than the input, the LLM has to make up details. These may or may not be what the writer intended.

Some positive uses of LLMs for creative writing

Writing is like software development: Given an idea of what is wanted, software developers produce code (text in a computer language) analogously to how writers produce text in a natural language. LLMs treat writing code and writing natural language text the same way; the corpus each LLM is trained on contains both natural language and code. What’s produced depends on the context.

Writers can learn from the experience of software developers. LLMs are good for small projects that have been done previously by many other people, such as database queries or writing standard letters. They are also useful for parts of larger projects, such as a pop-up box in a graphical user interface.

If programmers want to use them for bigger projects, they need to be prepared to generate multiple outputs and edit the one that is closest to what is intended. The problem in software development has always been specifying exactly what is wanted; coding is the easy part.

Generating good prompts

How to generate good prompts has been advocated as an art form called “prompt engineering.” Proponents of prompt engineering have suggested multiple techniques that improve the output of current LLMs, such as asking for an outline and then asking for the text based on the original prompt augmented with the outline.

Another is to ask the LLM to show its reasoning steps, as in so-called chain of thought. The LLM outputs don’t just the answer a question, but explains the steps that could be taken to answer it. The LLM uses those steps as part of its prompt to get its final answer.

Such advice is bound to be ephemeral. If some prompt-engineering technique works, it will be incorporated into a future release of the LLM, so that the effect happens without the need for the explicit use of the technique. Recent models that claim to reason have incorporated such step-by-step prompts.

People want to believe

Computer scientist Joseph Weizenbaum, describing his ELIZA program written in 1964–66, said: “I was startled to see how quickly and how very deeply people conversing with (the program) became emotionally involved with the computer and how unequivocally they anthropomorphized it.” The tools have changed, but people still want to believe.

In this age of misinformation, it is important for everyone to have a way to judge the often self-serving hype.

There is no magic in generative AI, but there is lots of data from which to predict what someone could write. I hope that creativity is more than regurgitating what others have written.

David Poole, Professor Emeritus of Computer Science, University of British Columbia

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

ChatGPT 大型语言模型 创意写作 人工智能 提示工程
相关文章