# Send messages to ChatGPT action

This action sends a message to Chat GPT and gathers a response using model GPT 3.5 Turbo from Azure OpenAI. Use this action simply as a single question and answer action or to provide a chat experience as a part of your recipes.

Send messages to ChatGPT action Send messages to ChatGPT action

# Input

Field Description Deployment ID Enter the deployment ID of the model you plan to use. You can find the deployment ID in Azure AI Studio > Deployment. Single Message Content The message to be sent over to the model with the role as user. System role message An optional message that provides specific instructions to the model before starting a conversation. Chat Transcript Role Select the role of the user corresponding to the specific message. Content The message to be sent over to the model for each corresponding role. System role message Optional message to provide specific instructions to the model before starting a conversation. Name Name of the author of a specific message. Often utilized for tracking chat transcripts easily. This can contain uppercase and lowercase characters, numbers, and underscores with a maximum length of 64 characters (no blank spaces). Model Select the OpenAI model to which you plan to send the message. Temperature Enter a value between 0 and 2 for controlling the randomness of completions. Higher values will make the output more random, while lower values will make it more focused and deterministic. We recommend using this or top p but not both. Learn more here. Number of chat completions The number of completions to generate as the message response. Stop phrase A specific stop phrase that will end generation. For example, if you set the stop phrase to a period (.) the model will generate text until it reaches a period, and then it will stop. Use this to control the amount of text generated. Maximum tokens The maximum number of tokens to generate in the completion. The token count of your prompt plus the value here cannot exceed the model's context length. For longer prompts, we recommend setting a low value and if the prompt is likely to vary in length, we recommend leaving it blank. Presence penalty A number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. User A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse.

# Output

Field Description Created The datetime stamp of when the response was generated. ID A unique identifier denoting the specific request and response that was sent over. Model The model used to generate the text completion. Choices Message The response of the model for the specified input. The role will always be Assistant. Finish reason The reason why the model stopped generating more text. Can be one of "stop", "length", "content_filter" and "null". More information can be found here. Response Contains the response which OpenAI probabilistically considers to be the ideal selection. Usage Prompt tokens The number of tokens used by the prompt. Completions tokens The number of tokens used for the completions of text. Total tokens The total number of tokens used by the prompt and response.

Last updated: 11/23/2023, 12:32:46 AM

On this page