🚀 Gate.io #Launchpad# for Puffverse (PFVS) is Live!
💎 Start with Just 1 $USDT — the More You Commit, The More #PFVS# You Receive!
Commit Now 👉 https://www.gate.io/launchpad/2300
⏰ Commitment Time: 03:00 AM, May 13th - 12:00 PM, May 16th (UTC)
💰 Total Allocation: 10,000,000 #PFVS#
⏳ Limited-Time Offer — Don’t Miss Out!
Learn More: https://www.gate.io/article/44878
#GateioLaunchpad# #GameeFi#
OpenAI has been updated heavily, and it has become stronger again! The API has a terrifying function call capability, can handle longer contexts, and the price has dropped by 75%!
On June 13, OpenAI's official website suddenly released the latest capability update of ChatGPT. The key points are as follows:
OpenAI also emphasized that all of these models continue to maintain the data privacy and security guarantees introduced on March 1-clients own all the outputs they request to generate, and their API data will not be used for training. With these updates, we're inviting even more people on the waitlist to try GPT-4, and we look forward to seeing what projects you build with GPT-4! We encourage developer feedback to help us ensure a smooth transition of model updates.
The following is the full text of the announcement:
We released the gpt-3.5-turbo and gpt-4 models earlier this year, and in just a few months we've seen developers build incredible applications on top of these models. Today, we're following up with some exciting updates:
New function call functionality in the Chat Completions API
· Updated gpt-4 and gpt-3.5-turbo versions with added controllability
· New 16k context version of gpt-3.5-turbo (vs. standard 4k version)
75% reduction in the price of our state-of-the-art embedding model
25% reduction in the price of gpt-3.5-turbo input tokens
Announced deprecation timeline for gpt-3.5-turbo-0301 and gpt-4-0314 models
All of these models come with the same data privacy and security guarantees we launched on March 1st - customers own all output generated by their requests, and their API data is not used for training. function call
Developers can now describe functions to gpt-4-0613 and gpt-3.5-turbo-0613, and have the model intelligently choose to output a JSON object containing the arguments for calling those functions. This is a new way to more reliably connect GPT's capabilities with external tools and APIs. These models have been fine-tuned to both detect when a function needs to be called (depending on user input) and respond with JSON that conforms to the function signature. Function calls allow developers to more reliably retrieve structured data from models. For example, developers can:
Create chatbots that answer questions by calling external tools such as the ChatGPT plugin;
Turning a question like "Ask Anya if she wants coffee next Friday" into a function call like send_email(to: string, body: string), or "What's the weather in Boston right now?" into get_current _weather(location: string, unit: 'celsius' | 'fahrenheit');
· Convert natural language into API calls or database queries;
Convert "Who are my top ten customers this month?" into an internal API call such as get_customers_by_revenue(start_date: string, end_date: string, limit: int), or "on How many orders did Acme, Inc. place in a month?" Use sql_query(query: string) to convert to SQL query;
Extract structured data from text;
· Define a function called extract_people_data(people: [{name: string, birthday: string, location: string}]) to extract all mentions of people from a Wikipedia article.
These use cases are all enabled by new API parameters in our /v1/chat/completions endpoint, functions and function_call, which allow developers to describe functions to the model via JSON Schema, and optionally ask it to call specific functions. Please get started with our developer docs, and if you see situations where function calls could be improved, please add an evaluation.
function call example
What's the weather like in Boston right now?
OpenAI API
Invoke the model with the function and the user's input
step 2
Third-party APIs
Call your API with the model's response
step 3
OpenAI API
Send the response back to the model for summarization The weather in Boston right now is sunny with a temperature of 22 degrees Celsius. Since the alpha release of the ChatGPT plugin, we've learned a lot about how to make tools and language models work together safely. However, there are still some open research questions. For example, a proof-of-concept vulnerability illustrates how obtaining untrusted data from a tool's output can cause the model to perform unexpected actions. We are working hard to mitigate these and other risks. Developers can protect their apps by only using information from trusted tools and including user confirmation steps before taking actions that have real-world impact, such as sending an email, posting online, or making a purchase.
new model
GPT-4
gpt-4-0613 includes updated and improved models, and function call functionality.
gpt-4-32k-0613 includes the same improvements as gpt-4-0613, plus increased context length for better understanding of larger texts.
With these updates, we will be inviting more people on the waitlist to try out GPT-4 in the coming weeks, with the goal of removing the waitlist for this model entirely. Thanks to everyone who has been patient, we look forward to seeing what you build with GPT-4!
GPT-3.5 Turbo
gpt-3.5-turbo-0613 includes the same function call capabilities as GPT-4, as well as more reliable control through system messages, both of which allow developers to more efficiently guide model responses.
gpt-3.5-turbo-16k provides four times the context length of gpt-3.5-turbo, but at double the price: $0.003 per 1K input tokens and $0.004 per 1K output tokens. 16k context means models can now be used in About 20 pages of text are supported in one request.
model deprecation
Today we will begin the process of upgrading and deprecating for the initial versions of gpt-4 and gpt-3.5-turbo that we announced in March. Apps using the stable model names (gpt-3.5-turbo, gpt-4, and gpt-4-32k) will automatically upgrade to the new models listed above on June 27th. To compare performance between model versions, our s library supports public and private evaluations to demonstrate how model changes will affect your use case.
Developers who need more time to transition can continue to use the old model by specifying gpt-3.5-turbo-0301, gpt-4-0314, or gpt-4-32k-0314 in the 'model' parameter of their API requests . These old models will still be accessible after September 13th, after which requests specifying these model names will fail. You can stay up-to-date on model deprecations on our model deprecations page. This is the first update to these models; as such, we welcome developer feedback to help us ensure a smooth transition.
Lower the price
Effective today, we continue to improve the efficiency of our systems and pass on these cost savings to developers.
Embeddings
text-embedding-ada-002 is our most popular embedding model. Today we reduced its cost by 75% to $0.0001 per 1K tokens.
GPT-3.5 Turbo
gpt-3.5-turbo is our most popular chat model, providing ChatGPT service for millions of users. Today we reduced the input token cost of gpt-3.5-turbo by 25%. Developers can now use this model for $0.0015 per 1K input tokens and $0.002 per 1K output tokens, which equates to roughly 700 pages per dollar.
The price of gpt-3.5-turbo-16k will be $0.003 per 1K input tokens and $0.004 per 1K output tokens.