Hello Streaming Markdown
View App View Code on GitHubshiny create --template basic-markdown-stream --mode express --github posit-dev/py-shiny-templates/gen-aishiny create --template basic-markdown-stream --mode core --github posit-dev/py-shiny-templates/gen-aiA basic example of collecting user input, using it to fill a LLM prompt template, then sending the result to the LLM for response generation. The response is then streamed back to the user in real-time via MarkdownStream().
To learn more, see the article on Gen AI streaming.
Other model providers
This particular template uses chatlas.ChatAnthropic() to do response generation via Anthropic. With chatlas, it’s easy to switch to other providers. Just change ChatAnthropic() to another provider (e.g., ChatOpenAI()).