Skip to content

Releases: relari-ai/continuous-eval

v0.3.13

04 Aug 23:02
a9823db
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.3.11...v0.3.13

v0.3.11

18 Jun 18:44
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.3.10...v0.3.11

v0.3.10

02 Jun 16:12
2618ab6
Compare
Choose a tag to compare

What's Changed

  • add max_tokens field for llm interface in #66
  • fix progress bar always 1 short | updated generator in examples to 4o in #67

Full Changelog: v0.3.9...v0.3.10

v0.3.9

23 May 19:53
8124a22
Compare
Choose a tag to compare

What's Changed

  • Fix contexts attribute name in examples documentation in #58
  • Transition to eval runner in #61
  • Add sql metrics in #63

New Contributors

Full Changelog: v0.3.7...v0.3.9

v0.3.7

25 Apr 01:42
Compare
Choose a tag to compare

What's Changed

  • fixed double counting corner case for precision / average precision in #55
  • Fix required keyword for code string to ground_truth_answers in #56

New Contributors

Full Changelog: v0.3.5...v0.3.7

v0.3.5

27 Mar 06:06
Compare
Choose a tag to compare
  • Add bedrock LLM provider
  • Bug fixing

v0.3.4

20 Mar 05:26
6fca8ab
Compare
Choose a tag to compare

What's Changed

  • Add cohere LLM in #46
  • Feature/llm custom metric in #48
  • Feature: add vllm openai api endpoint in #49

New Contributors

Full Changelog: v0.3.2...v0.3.4

v0.3.2

08 Mar 09:00
Compare
Choose a tag to compare
  • Metrics batch execution now use threads by default
  • Bug fixing

v0.3.1

29 Feb 19:38
1c088ed
Compare
Choose a tag to compare

Key points:

  • Added from_data class method to Dataset class
  • Fixed is_empty method in EvaluationResults, MetricsResults, and TestResults
  • Added error handling in LLM-based metrics

v0.2.7

16 Feb 22:19
696f8d0
Compare
Choose a tag to compare

What's Changed

  • Added Code Evaluation Metrics in #29

Full Changelog: v0.2.6...v0.2.7