You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
'description': 'The OpenAI model used to generate the response.',
53
53
# https://beta.openai.com/docs/models
54
54
'enum': [
55
+
'gpt-4',
56
+
'gpt-4-32k',
57
+
'gpt-3.5-turbo',
55
58
'text-davinci-003',
59
+
'text-davinci-002',
56
60
'text-curie-001',
57
61
'text-babbage-001',
58
62
'text-babbage-002',
59
63
'code-davinci-002',
60
64
'code-cushman-001'
61
65
],
62
66
'enumDescriptions': [
63
-
'Most capable GPT-3 model. Can do any task the other models can do, often with higher quality, longer output and better instruction-following. Also supports inserting completions within text.',
67
+
'More capable than any GPT-3.5 model, able to do more complex tasks, and optimized for chat. Will be updated with our latest model iteration.',
68
+
'Same capabilities as the base gpt-4 mode but with 4x the context length. Will be updated with our latest model iteration.',
69
+
'Most capable GPT-3.5 model and optimized for chat at 1/10th the cost of text-davinci-003. Will be updated with our latest model iteration.',
70
+
'Can do any language task with better quality, longer output, and consistent instruction-following than the curie, babbage, or ada models.',
71
+
'Similar capabilities to text-davinci-003 but trained with supervised fine-tuning instead of reinforcement learning'
64
72
'Very capable, but faster and lower cost than Davinci.',
65
73
'Capable of straightforward tasks, very fast, and lower cost.',
66
74
'Capable of very simple tasks, usually the fastest model in the GPT-3 series, and lowest cost.',
67
75
'Most capable Codex model. Particularly good at translating natural language to code. In addition to completing code, also supports inserting completions within code.',
68
76
'Almost as capable as Davinci Codex, but slightly faster. This speed advantage may make it preferable for real-time applications.'
'description': 'The maximum number of tokens used for completion. Tokens do not necessarily align with word or instruction count. Typically, each token is four characters. If your function is very large, you may need to decrease this value, as the number of tokens in your prompt counts against the total number of tokens supported by the model. Not all models support the same number of maximum tokens; most support 2,048 tokens. For larger functions, check out text-davinci-003 and code-davinci-002 which support 4,000 and 8,000 respectively.',
92
+
'description': 'The maximum number of tokens used for completion. Tokens do not necessarily align with word or instruction count. Typically, each token is four characters. If your function is very large, you may need to decrease this value, as the number of tokens in your prompt counts against the total number of tokens supported by the model. Not all models support the same number of maximum tokens; most support 2,048 tokens. For larger functions, check out text-davinci-003 and code-davinci-002 which support 4,000 and 8,000 respectively. For the maximum amount of tokens, check out gpt-4-32k which supports up to 32,768 tokens. This may be costly, however.',
0 commit comments