Sometimes it is helpful to run an evaluation locally without uploading any results to LangSmith. For example, if you’re quickly iterating on a prompt and want to smoke test it on a few examples, or if you’re validating that your target and evaluator functions are defined correctly, you may not want to record these evaluations. You can do this by using the LangSmith Python SDK and passingDocumentation Index
Fetch the complete documentation index at: https://docs.langchain.com/llms.txt
Use this file to discover all available pages before exploring further.
upload_results=False to evaluate() / aevaluate().
This will run you application and evaluators exactly as it always does and return the same output, but nothing will be recorded to LangSmith. This includes not just the experiment results but also the application and evaluator traces.
If you want to upload results to LangSmith but also need to process them in your script (for quality gates, custom aggregations, etc.), refer to Read experiment results locally.
Example
Let’s take a look at an example: Requireslangsmith>=0.2.0. Example also uses pandas.
| inputs.question | outputs.answer | reference.answer | feedback.is_concise | |
|---|---|---|---|---|
| 0 | What is the largest mammal? | What is the largest mammal? is a good question. I don’t know the answer. | The blue whale | False |
| 1 | What do mammals and birds have in common? | What do mammals and birds have in common? is a good question. I don’t know the answer. | They are both warm-blooded | False |
Connect these docs to Claude, VSCode, and more via MCP for real-time answers.

