Skip to content

Fix a couple of typos in docs/getstarted/evals.md #2081

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions docs/getstarted/evals.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ choose_evaluator_llm.md
**Evaluation**


Here we will use [AspectCritic](../concepts/metrics/available_metrics/aspect_critic.md), which an LLM based metric that outputs pass/fail given the evaluation criteria.
Here we will use [AspectCritic](../concepts/metrics/available_metrics/aspect_critic.md), which is an LLM based metric that outputs pass/fail given the evaluation criteria.


```python
Expand Down Expand Up @@ -148,8 +148,8 @@ Output
{'summary_accuracy': 0.84}
```

This score shows that out of all the samples in our test data, only 84% of summaries passes the given evaluation criteria. Now, **It
s important to see why is this the case**.
This score shows that out of all the samples in our test data, only 84% of summaries passes the given evaluation criteria. Now, **It's
important to see why is this the case**.

Export the sample level scores to pandas dataframe

Expand Down Expand Up @@ -187,4 +187,4 @@ If you want help with improving and scaling up your AI application using evals.

## Up Next

- [Evaluate a simple RAG application](rag_eval.md)
- [Evaluate a simple RAG application](rag_eval.md)