Skip to content

Commit

Permalink
Prep version 0.19.0
Browse files Browse the repository at this point in the history
  • Loading branch information
benbrandt committed Nov 28, 2024
1 parent 62ff28e commit 9248906
Show file tree
Hide file tree
Showing 4 changed files with 10 additions and 4 deletions.
6 changes: 6 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,11 @@
# Changelog

## v0.19.0

### Breaking Changes

- Update to tokenizers v0.21

## v0.18.1

### What's New
Expand Down
4 changes: 2 additions & 2 deletions Cargo.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
members = ["bindings/*"]

[workspace.package]
version = "0.18.1"
version = "0.19.0"
authors = ["Ben Brandt <benjamin.j.brandt@gmail.com>"]
edition = "2021"
description = "Split text into semantic chunks, up to a desired chunk size. Supports calculating length by characters and tokens, and is callable from Rust and Python."
Expand Down
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -203,7 +203,7 @@ There are lots of methods of determining sentence breaks, all to varying degrees
| ------------------ | ----------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ |
| `rust_tokenizers` | `^8.0.0` | Enables `(Text/Markdown)Splitter::new` to take any of the provided tokenizers as an argument. |
| `tiktoken-rs` | `^0.6.0` | Enables `(Text/Markdown)Splitter::new` to take `tiktoken_rs::CoreBPE` as an argument. This is useful for splitting text for `OpenAI` models. |
| `tokenizers` | `^0.20.0` | Enables `(Text/Markdown)Splitter::new` to take `tokenizers::Tokenizer` as an argument. This is useful for splitting text models that have a Hugging Face-compatible tokenizer. |
| `tokenizers` | `^0.21.0` | Enables `(Text/Markdown)Splitter::new` to take `tokenizers::Tokenizer` as an argument. This is useful for splitting text models that have a Hugging Face-compatible tokenizer. |

## Inspiration

Expand Down

0 comments on commit 9248906

Please sign in to comment.