TechCrunch News 03月02日
The TechCrunch AI glossary
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

人工智能领域术语繁多,为了更好地理解AI技术的发展,本文整理了一份AI术语表,定期更新,涵盖了AI agent、Chain of thought、Deep learning、Fine tuning、LLM、Neural network和Weights等关键概念。AI agent是能代表用户执行一系列任务的AI工具,而Chain of thought是一种将问题分解为中间步骤的推理方法,提高了解答的准确性。Deep learning是机器学习的一个分支,通过多层神经网络进行复杂的数据关联。Fine tuning则是通过专业数据进一步训练AI模型,优化其在特定任务上的表现。LLM是驱动AI助手的大型语言模型,Neural network是深度学习的基础结构。Weights决定了训练数据中不同特征的重要性,影响AI模型的输出。

🤖 AI Agent:指利用AI技术代表用户执行一系列任务的工具,例如预定餐厅、购买电影票等,它是一个自主系统,可以调用多个AI系统来完成多步骤任务。

🧠 Chain of Thought:这是一种大型语言模型的推理方法,将复杂问题分解为更小的中间步骤,从而提高最终结果的质量和准确性,尤其在逻辑或编码环境中效果显著。

💡 Deep Learning:深度学习是一种机器学习的子集,其AI算法采用多层人工神经网络结构,能够识别数据中的重要特征,并通过重复和调整来改进输出,但需要大量数据进行训练。

🎯 Fine Tuning:对AI模型进行进一步训练,通过输入新的、专业的任务导向型数据,优化其在特定任务或领域的性能,许多AI初创公司都采用这种方法来增强LLM的实用性。

🕸️ Neural Network:神经网络是深度学习的基础,也是生成式AI工具蓬勃发展的关键。它模拟人脑的神经元连接方式,通过图形处理硬件进行训练,从而在语音识别、自动导航和药物发现等领域实现卓越性能。

Artificial intelligence is a deep and convoluted world. The scientists who work in this field often rely on jargon and lingo to explain what they’re working on. As a result, we frequently have to use those technical terms in our coverage of the artificial intelligence industry. That’s why we thought it would be helpful to put together a glossary with definitions of some of the most important words and phrases that we use in our articles.

We will regularly update this glossary to add new entries as researchers continually uncover novel methods to push the frontier of artificial intelligence while identifying emerging safety risks.

An AI agent refers to a tool that makes use of AI technologies to perform a series of tasks on your behalf — beyond what a more basic AI chatbot could do — such as filing expenses, booking tickets or a table at a restaurant, or even writing and maintaining code. However, as we’ve explained before, there are lots of moving pieces in this emergent space, so different people can mean different things when they refer to an AI agent. Infrastructure is also still being built out to deliver on envisaged capabilities. But the basic concept implies an autonomous system that may draw on multiple AI systems to carry out multi-step tasks.

Given a simple question, a human brain can answer without even thinking too much about it — things like “which animal is taller between a giraffe and a cat?” But in many cases, you often need a pen and paper to come up with the right answer because there are intermediary steps. For instance, if a farmer has chickens and cows, and together they have 40 heads and 120 legs, you might need to write down a simple equation to come up with the answer (20 chickens and 20 cows).

In an AI context, chain-of-thought reasoning for large language models means breaking down a problem into smaller, intermediate steps to improve the quality of the end result. It usually takes longer to get an answer, but the answer is more likely to be right, especially in a logic or coding context. So-called reasoning models are developed from traditional large language models and optimized for chain-of-thought thinking thanks to reinforcement learning.

(See: Large language model)

A subset of self-improving machine learning in which AI algorithms are designed with a multi-layered, artificial neural network (ANN) structure. This allows them to make more complex correlations compared to simpler machine learning-based systems, such as linear models or decision trees. The structure of deep learning algorithms draws inspiration from the interconnected pathways of neurons in the human brain.

Deep learning AIs are able to identify important characteristics in data themselves, rather than requiring human engineers to define these features. The structure also supports algorithms that can learn from errors and, through a process of repetition and adjustment, improve their own outputs. However, deep learning systems require a lot of data points to yield good results (millions or more). It also typically takes longer to train deep learning vs. simpler machine learning algorithms — so development costs tend to be higher.

(See: Neural network)

This means further training of an AI model that’s intended to optimize performance for a more specific task or area than was previously a focal point of its training — typically by feeding in new, specialized (i.e. task-oriented) data. 

Many AI startups are taking large language models as a starting point to build a commercial product but vying to amp up utility for a target sector or task by supplementing earlier training cycles with fine-tuning based on their own domain-specific knowledge and expertise.

(See: Large language model (LLM))

Large language models, or LLMs, are the AI models used by popular AI assistants, such as ChatGPT, Claude, Google’s Gemini, Meta’s AI Llama, Microsoft Copilot, or Mistral’s Le Chat. When you chat with an AI assistant, you interact with a large language model that processes your request directly or with the help of different available tools, such as web browsing or code interpreters.

AI assistants and LLMs can have different names. For instance, GPT is OpenAI’s large language model and ChatGPT is the AI assistant product.

LLMs are deep neural networks made of billions of numerical parameters (or weights, see below) that learn the relationships between words and phrases and create a representation of language, a sort of multidimensional map of words.

Those are created from encoding the patterns they find in billions of books, articles, and transcripts. When you prompt an LLM, the model generates the most likely pattern that fits the prompt. It then evaluates the most probable next word after the last one based on what was said before. Repeat, repeat, and repeat.

(See: Neural network)

Neural network refers to the multi-layered algorithmic structure that underpins deep learning — and, more broadly, the whole boom in generative AI tools following the emergence of large language models. 

Although the idea to take inspiration from the densely interconnected pathways of the human brain as a design structure for data processing algorithms dates all the way back to the 1940s, it was the much more recent rise of graphical processing hardware (GPUs) — via the video game industry — that really unlocked the power of theory. These chips proved well suited to training algorithms with many more layers than was possible in earlier epochs — enabling neural network-based AI systems to achieve far better performance across many domains, whether for voice recognition, autonomous navigation, or drug discovery.

(See: Large language model (LLM))

Weights are core to AI training as they determine how much importance (or weight) is given to different features (or input variables) in the data used for training the system — thereby shaping the AI model’s output. 

Put another way, weights are numerical parameters that define what’s most salient in a data set for the given training task. They achieve their function by applying multiplication to inputs. Model training typically begins with weights that are randomly assigned, but as the process unfolds, the weights adjust as the model seeks to arrive at an output that more closely matches the target.

For example, an AI model for predicting house prices that’s trained on historical real estate data for a target location could include weights for features such as the number of bedrooms and bathrooms, whether a property is detached, semi-detached, if it has or doesn’t have parking, a garage, and so on. 

Ultimately, the weights the model attaches to each of these inputs is a reflection of how much they influence the value of a property, based on the given data set.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 AI术语 深度学习 大型语言模型 神经网络
相关文章