📄 ai-sdk/elements/components/conversation

File: conversation.md | Updated: 11/15/2025

Source: https://ai-sdk.dev/elements/components/conversation

Vercel

Slash Forward

Sparkles

AI SDK

Docs Cookbook Providers Playground AI ElementsAI ElementsLeft sparkleRight sparkle AI GatewayGateway

Search...⌘KFeedback GitHub Vercel LogoSign in with Vercel

Chatbot

Conversation

Wraps messages and automatically scrolls to the bottom. Also includes a scroll button that appears when not at the bottom.

The Conversation component wraps messages and automatically scrolls to the bottom. Also includes a scroll button that appears when not at the bottom.

Preview

Code

Start a conversation

Messages will appear here as the conversation progresses.

Installation


AI Elements

shadcn CLI

Manual

npx ai-elements@latest add conversation

Usage with AI SDK


Build a simple conversational UI with Conversation and PromptInput :

Add the following component to your frontend:

app/page.tsx

'use client';

import {
  Conversation,
  ConversationContent,
  ConversationEmptyState,
  ConversationScrollButton,
} from '@/components/ai-elements/conversation';
import { Message, MessageContent } from '@/components/ai-elements/message';
import {
  Input,
  PromptInputTextarea,
  PromptInputSubmit,
} from '@/components/ai-elements/prompt-input';
import { MessageSquare } from 'lucide-react';
import { useState } from 'react';
import { useChat } from '@ai-sdk/react';
import { MessageResponse } from '@/components/ai-elements/message';

const ConversationDemo = () => {
  const [input, setInput] = useState('');
  const { messages, sendMessage, status } = useChat();

  const handleSubmit = (e: React.FormEvent) => {
    e.preventDefault();
    if (input.trim()) {
      sendMessage({ text: input });
      setInput('');
    }
  };

  return (
    <div className="max-w-4xl mx-auto p-6 relative size-full rounded-lg border h-[600px]">
      <div className="flex flex-col h-full">
        <Conversation>
          <ConversationContent>
            {messages.length === 0 ? (
              <ConversationEmptyState
                icon={<MessageSquare className="size-12" />}
                title="Start a conversation"
                description="Type a message below to begin chatting"
              />
            ) : (
              messages.map((message) => (
                <Message from={message.role} key={message.id}>
                  <MessageContent>
                    {message.parts.map((part, i) => {
                      switch (part.type) {
                        case 'text': // we don't use any reasoning or tool calls in this example
                          return (
                            <MessageResponse key={`${message.id}-${i}`}>
                              {part.text}
                            </MessageResponse>
                          );
                        default:
                          return null;
                      }
                    })}
                  </MessageContent>
                </Message>
              ))
            )}
          </ConversationContent>
          <ConversationScrollButton />
        </Conversation>

        <Input
          onSubmit={handleSubmit}
          className="mt-4 w-full max-w-2xl mx-auto relative"
        >
          <PromptInputTextarea
            value={input}
            placeholder="Say something..."
            onChange={(e) => setInput(e.currentTarget.value)}
            className="pr-12"
          />
          <PromptInputSubmit
            status={status === 'streaming' ? 'streaming' : 'ready'}
            disabled={!input.trim()}
            className="absolute bottom-1 right-1"
          />
        </Input>
      </div>
    </div>
  );
};

export default ConversationDemo;

Add the following route to your backend:

api/chat/route.ts

import { streamText, UIMessage, convertToModelMessages } from 'ai';

// Allow streaming responses up to 30 seconds
export const maxDuration = 30;

export async function POST(req: Request) {
  const { messages }: { messages: UIMessage[] } = await req.json();

  const result = streamText({
    model: 'openai/gpt-4o',
    messages: convertToModelMessages(messages),
  });

  return result.toUIMessageStreamResponse();
}

Features


  • Automatic scrolling to the bottom when new messages are added
  • Smooth scrolling behavior with configurable animation
  • Scroll button that appears when not at the bottom
  • Responsive design with customizable padding and spacing
  • Flexible content layout with consistent message spacing
  • Accessible with proper ARIA roles for screen readers
  • Customizable styling through className prop
  • Support for any number of child message components

Props


<Conversation />

Prop

Type

contextRef?React.Ref<StickToBottomContext>

instance?StickToBottomInstance

children?((context: StickToBottomContext) => ReactNode) | ReactNode

...props?Omit<React.HTMLAttributes<HTMLDivElement>, "children">

<ConversationContent />

Prop

Type

children?((context: StickToBottomContext) => ReactNode) | ReactNode

...props?Omit<React.HTMLAttributes<HTMLDivElement>, "children">

<ConversationEmptyState />

Prop

Type

title?string

description?string

icon?React.ReactNode

children?React.ReactNode

...props?ComponentProps<"div">

<ConversationScrollButton />

Prop

Type

...props?ComponentProps<typeof Button>

Context

A compound component system for displaying AI model context window usage, token consumption, and cost estimation.
Inline Citation

A hoverable citation component that displays source information and quotes inline with text, perfect for AI-generated content with references.

On this page

Installation Usage with AI SDK Features Props <Conversation /> <ConversationContent /> <ConversationEmptyState /> <ConversationScrollButton />

GitHubEdit this page on GitHub Scroll to topCopy pageOpen in chat