A streaming interface allows you to display partial answers as they are generated by the underlying GPT model, the same as what users see with ChatGPT. Due to the metadata involved, this requires a sequence of API calls to execute.
-
POST =>
/chat/message/createCreate record of the message, receive back messageId and conversationId, if one was not passed in the request.
-
POST =>
/chat/message/streamInitiate answer streaming by passing messageId and options.
-
GET =>
/chat/message/{messageId}Get answerMessageId and references associated with the given message. References are ready as soon as there is an answer streaming.
-
POST =>
/chat/message/followup-suggestionsGenerate suggested followup questions.
-
POST =>
/chat/message/feedbackRecord feedback given by the user for the generated answer.
