Skip to content

Commit d9b29f1

Browse files
authored
docs: Move multiple playground messages guide (#779)
1 parent ae697e8 commit d9b29f1

File tree

9 files changed

+14
-13
lines changed

9 files changed

+14
-13
lines changed

docs/evaluation/how_to_guides/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,6 @@ Evaluate and improve your application before deploying it.
3030
- [Run an evaluation from the UI](./how_to_guides/run_evaluation_from_prompt_playground)
3131
- [Run an evaluation via the REST API](./how_to_guides/run_evals_api_only)
3232
- [Run an evaluation with multimodal content](./how_to_guides/evaluate_with_attachments)
33-
- [Set up a multi-turn evaluation](./how_to_guides/multiturn_evaluation)
3433

3534
### Define an evaluator
3635

docs/prompt_engineering/how_to_guides/index.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,12 +16,13 @@ Organize and manage prompts in LangSmith to streamline your LLM development work
1616

1717
Quickly iterate on prompts and models in the LangSmith Playground.
1818

19-
- [Run an evaluation in the Playground](/evaluation?mode=ui)
19+
- [Run an evaluation in the playground](/evaluation?mode=ui)
2020
- [Include multimodal content in a prompt](./how_to_guides/multimodal_content)
2121
- [Manage prompt settings](./how_to_guides/managing_model_configurations)
2222
- [Iterate on your prompts with the Prompt Canvas](./how_to_guides/prompt_canvas)
2323
- [Run the playground against a custom LangServe model server](./how_to_guides/custom_endpoint)
2424
- [Run the playground against an OpenAI-compliant model provider/proxy](./how_to_guides/custom_openai_compliant_model)
25+
- [Use multiple messages in the playground](./how_to_guides/multiple_messages)
2526

2627
## Few shot prompting
2728

docs/evaluation/how_to_guides/multiturn_evaluation.mdx renamed to docs/prompt_engineering/how_to_guides/multiple_messages.mdx

Lines changed: 8 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,7 @@
1-
# How to set up a multi-turn evaluation
1+
# How to use multiple messages in the playground
22

3-
LangSmith makes it easy to evaluate multi-turn conversations in the playground. This allows you to evaluate how changing your system prompt,
4-
the tools available to the model, or the output schema affects a conversation with multiple messages.
5-
6-
This how-to guide walks you through the various ways you can set up the playground for multi-turn evaluation, which will allow you to
7-
test different tool configurations and system prompts to see how they impact your system.
3+
This how-to guide walks you through the various ways you can set up the playground for multi-turn conversations, which will allow you to
4+
test different tool configurations and system prompts against longer threads of messages.
85

96
![](./static/multiturn_diagram.png)
107

@@ -22,7 +19,7 @@ You can then edit the system prompt, tweak the tools and/or output schema and ob
2219
Before starting, make sure you have [set up your dataset](../../evaluation/how_to_guides/manage_datasets_in_application). Since you want to evaluate
2320
multi-turn conversations, make sure there is a key in your inputs that contains a list of messages.
2421

25-
Once you have created your dataset, head to the playground and [load your dataset](/evaluation/how_to_guides/manage_datasets_in_application#from-the-prompt-playground) to evaluate.
22+
Once you have created your dataset, head to the playground and [load your dataset](../../evaluation/how_to_guides/manage_datasets_in_application#from-the-prompt-playground) to evaluate.
2623

2724
Then, add a messages list variable to your prompt, making sure to name it the same as the key in your inputs that contains the list of messages:
2825

@@ -43,11 +40,11 @@ multi-turn conversation there:
4340
![](./static/multiturn_manual_list.gif)
4441

4542
This allows you to just tweak the system prompt or the tools, while allowing any multi-turn conversation to take the place
46-
of the `Messages List` variable, allowing you to reuse this prompt across various evaluations.
43+
of the `Messages List` variable, allowing you to reuse this prompt across various runs.
4744

4845
## Next Steps
4946

50-
Now that you know how to set up your multi-turn evaluation, you can either manually
51-
inspect and judge the outputs, or you can [add evaluators](../how_to_guides#define-an-evaluator) to get repeatable, quantitative results.
47+
Now that you know how to set up the playground for multi-turn interactions, you can either manually
48+
inspect and judge the outputs, or you can [add evaluators](../../evaluation/how_to_guides#define-an-evaluator) to classify results.
5249

53-
You can also read [these how-to guides](../../prompt_engineering/how_to_guides#playground) to learn more about how to use the playground to run evaluations.
50+
You can also read [these how-to guides](../../prompt_engineering/how_to_guides#playground) to learn more about how to use the playground to run evaluations.

vercel.json

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -179,6 +179,10 @@
179179
"source": "/evaluation/how_to_guides/use_langchain_off_the_shelf_evaluators",
180180
"destination": "/evaluation/how_to_guides/prebuilt_evaluators"
181181
},
182+
{
183+
"source": "/evaluation/how_to_guides/multiturn_evaluation",
184+
"destination": "/prompt_engineering/how_to_guides/multiple_messages"
185+
},
182186
{
183187
"source": "/old/tracing/faq/querying_traces",
184188
"destination": "/observability/how_to_guides/export_traces"

0 commit comments

Comments
 (0)