Skip to main content

Playground

The LLM Playground is currently in preview.

Evaluating LLM prompts and responses is challenging. The Weave Playground is designed to simplify the process of iterating on LLM prompts and responses, making it easier to experiment with different models and prompts. With features like prompt editing, message retrying, and model comparison, Playground helps you to quickly test and improve your LLM applications. Playground currently supports models from OpenAI, Anthropic, Google, Groq, Amazon Bedrock, and Microsoft Azure, as well as custom providers.

Features

  • Quick access: Open the Playground from the W&B sidebar for a fresh session or from the Call page to test an existing project.
  • Message controls: Edit, retry, or delete messages directly within the chat.
  • Flexible messaging: Add new messages as either user or system inputs, and send them to the LLM.
  • Customizable settings: Configure your preferred LLM provider and adjust model settings.
  • Multi-LLM support: Switch between models, with team-level API key management.
  • Compare models: Compare how different models respond to prompts.

Get started with the Playground to optimize your LLM interactions and streamline your prompt engineering process and LLM application development.

Prerequisites

Before you can use Playground, you must add provider credentials, and open the Playground UI.

Add provider credentials and information

Playground currently supports models from OpenAI, Anthropic, Google, Groq, Amazon Bedrock, and Microsoft Azure. To use one of the available models, add the appropriate information to your team secrets in W&B settings.

  • OpenAI: OPENAI_API_KEY
  • Anthropic: ANTHROPIC_API_KEY
  • Google: GOOGLE_API_KEY
  • Groq: GEMMA_API_KEY
  • Amazon Bedrock:
    • AWS_ACCESS_KEY_ID
    • AWS_SECRET_ACCESS_KEY
    • AWS_REGION_NAME
  • Azure:
    • AZURE_API_KEY
    • AZURE_API_BASE
    • AZURE_API_VERSION
  • X.AI:
    • XAI_API_KEY
  • Deepseek
    • DEEPSEEK_API_KEY

Access the Playground

There are two ways to access the Playground:

  1. Open a fresh Playground page with a simple system prompt: In the sidebar, select Playground. Playground opens in the same tab.
  2. Open Playground for a specific call:
    1. In the sidebar, select the Traces tab. A list of traces displays.
    2. In the list of traces, click the name of the call that you want to view. The call's details page opens.
    3. Click Open chat in playground. Playground opens in a new tab.

Screenshot of Open in Playground button

Select an LLM

You can switch the LLM using the dropdown menu in the top left. The available models from various providers are listed below:

Amazon Bedrock

  • ai21.j2-mid-v1
  • ai21.j2-ultra-v1
  • amazon.nova-micro-v1:0
  • amazon.nova-lite-v1:0
  • amazon.nova-pro-v1:0
  • amazon.titan-text-lite-v1
  • amazon.titan-text-express-v1
  • mistral.mistral-7b-instruct-v0:2
  • mistral.mixtral-8x7b-instruct-v0:1
  • mistral.mistral-large-2402-v1:0
  • mistral.mistral-large-2407-v1:0
  • anthropic.claude-3-sonnet-20240229-v1:0
  • anthropic.claude-3-5-sonnet-20240620-v1:0
  • anthropic.claude-3-haiku-20240307-v1:0
  • anthropic.claude-3-opus-20240229-v1:0
  • anthropic.claude-v2
  • anthropic.claude-v2:1
  • anthropic.claude-instant-v1
  • cohere.command-text-v14
  • cohere.command-light-text-v14
  • cohere.command-r-plus-v1:0
  • cohere.command-r-v1:0
  • meta.llama2-13b-chat-v1
  • meta.llama2-70b-chat-v1
  • meta.llama3-8b-instruct-v1:0
  • meta.llama3-70b-instruct-v1:0
  • meta.llama3-1-8b-instruct-v1:0
  • meta.llama3-1-70b-instruct-v1:0
  • meta.llama3-1-405b-instruct-v1:0

Anthropic

  • claude-3-5-sonnet-20240620
  • claude-3-5-sonnet-20241022
  • claude-3-haiku-20240307
  • claude-3-opus-20240229
  • claude-3-sonnet-20240229

Azure

  • azure/o1-mini
  • azure/o1-mini-2024-09-12
  • azure/o1
  • azure/o1-preview
  • azure/o1-preview-2024-09-12
  • azure/gpt-4o
  • azure/gpt-4o-2024-08-06
  • azure/gpt-4o-2024-11-20
  • azure/gpt-4o-2024-05-13
  • azure/gpt-4o-mini
  • azure/gpt-4o-mini-2024-07-18

Google

  • gemini/gemini-1.5-flash-001
  • gemini/gemini-1.5-flash-002
  • gemini/gemini-1.5-flash-8b-exp-0827
  • gemini/gemini-1.5-flash-8b-exp-0924
  • gemini/gemini-1.5-flash-exp-0827
  • gemini/gemini-1.5-flash-latest
  • gemini/gemini-1.5-flash
  • gemini/gemini-1.5-pro-001
  • gemini/gemini-1.5-pro-002
  • gemini/gemini-1.5-pro-exp-0801
  • gemini/gemini-1.5-pro-exp-0827
  • gemini/gemini-1.5-pro-latest
  • gemini/gemini-1.5-pro
  • gemini/gemini-pro
  • gemini/gemini-2.0-flash-exp
  • gemini/gemini-2.0-flash-thinking-exp

Groq

  • groq/gemma-7b-it
  • groq/gemma2-9b-it
  • groq/llama-3.1-8b-instant
  • groq/llama3-70b-8192
  • groq/llama3-8b-8192
  • groq/llama3-groq-8b-8192-tool-use-preview
  • groq/mixtral-8x7b-32768

OpenAI

  • gpt-4o-mini
  • gpt-3.5-turbo-0125
  • gpt-3.5-turbo-1106
  • gpt-4-1106-preview
  • gpt-4-32k-0314
  • gpt-4-turbo-2024-04-09
  • gpt-4-turbo-preview
  • gpt-4-turbo
  • gpt-4
  • gpt-4o-2024-05-13
  • gpt-4o-2024-08-06
  • gpt-4o-mini-2024-07-18
  • gpt-4o
  • gpt-4o-2024-11-20
  • o1-mini-2024-09-12
  • o1-mini
  • o1-preview-2024-09-12
  • o1-preview
  • o1-2024-12-17

X.AI

  • xai/grok-beta
  • xai/grok-2-1212
  • xai/grok-2
  • xai/grok-2-latest

Deepseek

  • deepseek/deepseek-reasoner
  • deepseek/deepseek-chat

Adjust LLM parameters

You can experiment with different parameter values for your selected model. To adjust parameters, do the following:

  1. In the upper right corner of the Playground UI, click Chat settings to open the parameter settings dropdown.
  2. In the dropdown, adjust parameters as desired. You can also toggle Weave call tracking on or off, and add a function.
  3. Click Chat settings to close the dropdown and save your changes.

Screenshot of Playground settings

Add a function

You can test how different models use functions based on input it receives from the user. To add a function for testing in Playground, do the following:

  1. In the upper right corner of the Playground UI, click Chat settings to open the parameter settings dropdown.
  2. In the dropdown, click + Add function.
  3. In the pop-up, add your function information.
  4. To save your changes and close the function pop-up, click the x in the upper right corner.
  5. Click Chat settings to close the settings dropdown and save your changes.

Retry, edit, and delete messages

With Playground, you can retry, edit, and delete messages. To use this feature, hover over the message you want to edit, retry, or delete. Three buttons display: Delete, Edit, and Retry.

  • Delete: Remove the message from the chat.
  • Edit: Modify the message content.
  • Retry: Delete all subsequent messages and retry the chat from the selected message.

Screenshot of Playground message buttons Screenshot of Playground editing

Add a new message

To add a new message to the chat, do the following:

  1. In the chat box, select one of the available roles (Assistant or User)
  2. Click + Add.
  3. To send a new message to the LLM, click the Send button. Alternatively, press the Command and Enter keys.

Screenshot of Playground sending a message

Compare LLMs

Playground allows you to compare LLMs. To perform a comparison, do the following:

  1. In the Playground UI, click Compare. A second chat opens next to the original chat.
  2. In the second chat, you can:
  3. In the message box, enter a message that you want to test with both models and press Send.

Adjust the number of trials

Playground allows you to generate multiple outputs for the same input by setting the number of trials. The default setting is 1. To adjust the number of trials, do the following:

  1. In the Playground UI, open the settings sidebar if it is not already open.
  2. Adjust the Number of trials.

Add a custom provider

In addition to the supported providers, you can use the Playground to test OpenAI compatible API endpoints for custom models. Examples include:

  • Older versions of supported model providers
  • Local models

To add a custom provider to the Playground, do the following:

  1. In the upper left corner of the Playground UI, click the Select a model dropdown.

  2. Select + Add AI provider.

  3. In the pop-up modal, enter the provider information:

    • Provider name: For example, openai or ollama.
    • API key: For example, an OpenAI API key.
    • Base URL: For example, https://api.openai.com/v1/ or a ngrok URL https://e452-2600-1700-45f0-3e10-2d3f-796b-d6f2-8ba7.ngrok-free.app.
    • Headers (optional): You can add multiple header keys and values.
    • Models: You can add multiple models for one provider. For example, deepseek-r1 and qwq.
    • Max tokens (optional): For each model, you can specify the max tokens that the model can generate in a response.
  4. Once you've entered your provider information, click Add provider.

  5. Select your new provider and available model(s) from the Select a model dropdown in the upper left corner of the Playground UI.

important

Because of CORS restrictions, you can't call localhost or 127.0.0.1 URLs directly from the Playground. If you're running a local model server (such as Ollama), use a tunneling service like ngrok to expose it securely. For details, see Use ngrok with Ollama.

Now, you can test the custom provider model(s) using standard Playground features. You can also edit or remove the custom provider.

Edit a custom provider

To edit information for a previously created custom provider, do the following:

  1. In the Weave sidebar, navigate to Overview.
  2. From the top navigation menu, select AI Providers.
  3. In the Custom providers table, find the custom provider you want to update.
  4. In the Last Updated column of the entry for your custom provider, click the edit button (the pencil icon).
  5. In the pop-up modal, edit the provider information.
  6. Click Save.

Remove a custom provider

To remove a previously created custom provider, do the following:

  1. In the Weave sidebar, navigate to Overview.
  2. From the top navigation menu, select AI Providers.
  3. In the Custom providers table, find the custom provider you want to update.
  4. In the Last Updated column of the entry for your custom provider, click the delete button (the trashcan icon).
  5. In the pop-up modal, confirm that you want to delete the provider. This action cannot be undone.
  6. Click Delete.

Use ngrok with Ollama

To test a locally running Ollama model in the Playground, use ngrok to create a temporary public URL that bypasses CORS restrictions.

To set it up, do the following:

  1. Install ngrok for your operating system.

  2. Start your Ollama model:

    ollama run <model>
  3. In a separate terminal, create an ngrok tunnel with the required CORS headers:

    ngrok http 11434 --response-header-add "Access-Control-Allow-Origin: *" --host-header rewrite

After ngrok starts, it will display a public URL, such as https://xxxx-xxxx.ngrok-free.app. Use this URL as the base URL when you add Ollama as a custom provider in the Playground.

The following diagram illustrates the data flow between your local environment, the ngrok proxy, and the W&B cloud services: