benchmarking llm performance with langchain auto-evaluator // lance martin //llms in prod con part 2
Published 11 months ago • 1.1K plays • Length 8:25Download video MP4
Download video MP3
Similar videos
-
49:50
evaluating llm-based applications // josh tobin // llms in prod conference part 2
-
5:50
7 popular llm benchmarks explained [openllm leaderboard & chatbot arena]
-
52:07
genai solutions with langchain: lance martin on llms, agents, evals, and more!
-
45:32
a survey of techniques for maximizing llm performance
-
57:55
optimization of llm systems with dspy and langchain/langsmith
-
24:02
langgraph: multi-agent workflows
-
6:00
announcing llm litmus test and sourcegraph labs!
-
8:37
pre-built evaluators | langsmith evaluations - part 5
-
55:06
building llm applications on langchain with lance martin | decoding ai
-
5:30
benchmarking llms with lmsys.org
-
54:56
building llm applications with langchain with lance
-
1:29:45
rag workshop with langchain and llamaindex
-
6:45
why evals matter | langsmith evaluations - part 1
-
44:22
rag time! evaluate rag with llm evals and benchmarking
-
1:18:03
monitoring llms in production using openai, langchain & whylabs
-
1:06:38
monitoring llms in production using langchain and whylabs
-
26:38
automated evaluation of llm apps with the azure-ai-generative sdk | python data science day
-
40:35
testing framework giskard for llm and rag evaluation (bias, hallucination, and more)
-
10:00
open source rag running llms locally with ollama