OpenAI GPT-3 API: What is the difference between davinci and text-davinci-003?

news/2024/9/24 5:22:25/
aidu_pl">

题意:OpenAI GPT-3 API:davinci 和 text-davinci-003 有什么区别

问题背景:

I'm testing the different models for OpenAI, and I noticed that not all of them are developed or trained enough to give a reliable response.

我正在测试 OpenAI 的不同模型,我发现并不是所有模型都足够完善或训练充分,无法提供可靠的回应

The models I tested are the following:

我测试的模型如下:

model_engine = "text-davinci-003"
model_engine = "davinci" 
model_engine = "curie" 
model_engine = "babbage" 
model_engine = "ada" 

I need to understand what the difference is between davinci and text-davinci-003, and how to improve the responses to match that response when you use ChatGPT.

我需要了解 davincitext-davinci-003 之间的区别,以及如何改进响应,使其与使用 ChatGPT 时的响应相匹配

问题解决:

TL;DR   摘要

  • text-davinci-003 is the newer and more capable model than davinci

text-davinci-003 是比 davinci 更新且更强大的模型

  • text-davinci-003 supports a longer context window than davinci

text-davinci-003 支持比 davinci 更长的上下文窗口

  • text-davinci-003 was trained on a more recent dataset than davinci

text-davinci-003 使用比 davinci 更新的数据集进行训练

  • text-davinci-003 is cheaper than davinci

text-davinci-003davinci 更便宜

  • text-davinci-003 is not available for fine-tuning, while davinci is

text-davinci-003 不支持微调,而 davinci 支持

Capabilities   能力

As stated in the official OpenAI article:   正如 OpenAI 官方文章中所述

While both davinci and text-davinci-003 are powerful models, they differ in a few key ways.

text-davinci-003 is the newer and more capable model, designed specifically for instruction-following tasks. This enables it to respond concisely and more accurately - even in zero-shot scenarios, i.e. without the need for any examples given in the prompt. davinci, on the other hand, can be fine-tuned on a specific task, which can make it very effective if you have access to at least a few hundred training examples.

Additionally, text-davinci-003 supports a longer context window (max prompt+completion length) than davinci - 4097 tokens compared to davinci's 2049.

Finally, text-davinci-003 was trained on a more recent dataset, containing data up to June 2021. These updates, along with its support for Inserting text, make text-davinci-003 a particularly versatile and powerful model we recommend for most use-cases.

Use text-davinci-003 because the other models you mentioned in your question are less capable.

使用 text-davinci-003,因为您在问题中提到的其他模型能力较弱

If you buy a ChatGPT Plus subscription, you can also use gpt-3.5-turbo or gpt-4. So, to get similar responses as you get from ChatGPT, it depends on whether you are subscribed or not. For sure, gpt-3.5-turbo and gpt-4 are even more capable than text-davinci-003.

如果您购买 ChatGPT Plus 订阅,您还可以使用 gpt-3.5-turbogpt-4。因此,获得与 ChatGPT 相似的响应取决于您是否订阅。可以肯定的是,gpt-3.5-turbogpt-4 的能力甚至比 text-davinci-003 更强

Costs   成本

text-davinci-003 is cheaper than davinci, as stated on the official OpenAI website:

正如 OpenAI 官方网站所述,text-davinci-003davinci 更便宜

MODELUSAGE
davinci$0.1200 / 1K tokens
text-davinci-003$0.0200 / 1K tokens

Fine-tuning availability   微调可用性

text-davinci-003 is not available for fine-tuning, while davinci is, as stated in the official OpenAI documentation:

正如 OpenAI 官方文档所述,text-davinci-003 不支持微调,而 davinci 支持

Fine-tuning is currently only available for the following base models: davincicuriebabbage, and ada. These are the original models that do not have any instruction following training (like text-davinci-003 does for example).

MODELFINE-TUNING AVAILABILITYTRAINING
davinciYes$0.0300 / 1K tokens
text-davinci-003No


http://www.ppmy.cn/news/1529662.html

相关文章

解决Vue 3中Element Plus el-color-picker 组件消失的问题

概述 在使用Vue 3和Element Plus框架构建应用程序时,你可能会遇到el-color-picker组件无法正常显示的情况。这可能是由于多种原因造成的,包括但不限于CSS加载问题、JavaScript错误、版本兼容性等。本文将指导你如何排查并解决这个问题。 在 el-color-p…

STM32——输入捕获

输入捕获模式可以用来测量脉冲宽度或者测量频率。STM32的定时器,除了TIM6、TIM7,其他的定时器都有输入捕获的功能。应用场景是编码器。 测量频率 当捕获通道TIx 上出现上升沿时,发生第一次捕获,计数器CNT 的值会被锁存到捕获寄存…

JIT(Just-In-Time Compilation)编译器及其原理实现

什么是JIT编译器? JIT编译器是一种特殊的编译器,它在程序运行期间动态地将字节码(Bytecode)转换为机器码。这种转换发生在程序执行时,而不是在程序启动之前。JIT编译器的主要目的是提高程序的执行速度和效率&#xff…

AOT源码解析4.1-model主体解析

1 输入数据 VOS的数据集处理操作可见数据集操作,这里是进行数据集提取完毕后的操作。 图2:如图所示,使用datasets提取出数据之后,在模型训练阶段对数据做图中操作。即:将batch_size大小的ref_imgs、prev_imgs&#x…

洛谷-P3916 图的遍历

题目描述 给出 N 个点,M 条边的有向图,对于每个点 v,求A(v) 表示从点 v 出发,能到达编号最大的点。 思路 既然是要找到最大的点,那么我从最大的点开始DFS是否可以? 于是可以反向建图,然后从最…

OpenHarmony(鸿蒙南向开发)——小型系统内核(LiteOS-A)【扩展组件】上

往期知识点记录: 鸿蒙(HarmonyOS)应用层开发(北向)知识点汇总 鸿蒙(OpenHarmony)南向开发保姆级知识点汇总~ 子系统开发内核 轻量系统内核(LiteOS-M) 轻量系统内核&#…

[vulnhub] LAMPSecurity: CTF4

https://www.vulnhub.com/entry/lampsecurity-ctf4,83/ 端口扫描主机发现 探测存活主机,138是靶机 nmap -sP 192.168.75.0/24 // Starting Nmap 7.93 ( https://nmap.org ) at 2024-09-23 14:13 CST Nmap scan report for 192…

Actions Speak Louder than Words Meta史诗级的端到端推荐大模型落地

发现好久之前整理的推荐系统被遗忘在了草稿箱,让它出来见见世面。。。后续空了持续更新 1.Background 大模型生成用于推荐场景有如下几个难点: 特征缺乏显式结构。存在sparse和dense特征,其中sparse特征指的是一些离散特征,这部…