Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

added ask_stream endpoint #726

Open
wants to merge 4 commits into
base: main
Choose a base branch
from

Conversation

carlodek
Copy link

@carlodek carlodek commented Aug 1, 2024

you can now call azure openAI and OpenAI with streaming

Motivation and Context (Why the change? What's the scenario?)

I love repo but I'd love to call aureOpenAI and openAI in streaming way

High level description (Approach, Design)

I've added a new endpoint called /ask_stream, there is the possibility to customize the eosToken for chunks. Service will give back data:{MemoryAnswer object}.

you can now call azure openAI and OpenAI with streaming
@carlodek carlodek requested a review from dluc as a code owner August 1, 2024 14:06
@carlodek
Copy link
Author

carlodek commented Aug 1, 2024 via email

@dluc
Copy link
Collaborator

dluc commented Oct 16, 2024

Update: for this feature to be merged, there's a couple of things to do:

  • Check this similar PR Implement new streaming ask endpoint (WIP) #400 and decide which approach to take
  • Support content moderation. The stream of tokens needs to be validated while streamed, on a configurable frequence. If at any point the text moderation fails, the stream needs to be reset, e.g. sending a special token or similar.

@dluc dluc added the waiting for author Waiting for author to reply or address comments label Oct 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
waiting for author Waiting for author to reply or address comments
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants