Skip to content

Releases: TabbyML/tabby

v0.2.1

04 Oct 00:39
Compare
Choose a tag to compare

πŸš€ Features

Chat Model & Web Interface

We have introduced a new argument, --chat-model, which allows you to specify the model for the chat playground located at http://localhost:8080/playground

To utilize this feature, use the following command in the terminal:

tabby serve --device metal --model TabbyML/StarCoder-1B --chat-model TabbyML/Mistral-7B

ModelScope Model Registry

Mainland Chinese users have been facing challenges accessing Hugging Face due to various reasons. The Tabby team is actively working to address this issue by mirroring models to a hosting provider in mainland China called modelscope.cn.

# Download from the Modelscope registry
TABBY_REGISTRY=modelscope tabby download --model TabbyML/WizardCoder-1B

🧰 Fixes and improvements

  • Implemented more accurate UTF-8 incremental decoding in the GitHub pull request.
  • Fixed the stop words implementation by utilizing RegexSet to isolate the stop word group.
  • Improved model downloading logic; now Tabby will attempt to fetch the latest model version if there's a remote change, and the local cache key becomes stale.
  • set default num_replicas_per_device for ctranslate2 backend to increase parallelism.

πŸ’« New Contributors

Full Changelog: v0.1.2...v0.2.1

v0.2.0-rc.0

03 Oct 02:39
Compare
Choose a tag to compare
v0.2.0-rc.0 Pre-release
Pre-release
v0.2.0-rc.0

v0.2.0

03 Oct 20:56
Compare
Choose a tag to compare
v0.2.0 Pre-release
Pre-release
fix: playground environment misconfig

v0.1.2

19 Sep 06:14
61ade26
Compare
Choose a tag to compare

Patch Release

  • docs: add model spec (unstable) version in #457
  • docs: update vim documentation in #453
  • fix(tabby): fix swagger's local server use local port in #458
  • feat: Update Dockerfile to ctranslate 3.20.0 in #460

New Contributors

Full Changelog: v0.1.1...v0.1.2

v0.1.1

17 Sep 06:19
Compare
Choose a tag to compare

🍺 Homebrew support (Apple M1/M2)

brew install tabbyml/tabby/tabby

# Start with StarCoder-1B
tabby serve --device metal --model TabbyML/StarCoder-1B

πŸ“¦ Artifacts

πŸš€ Features

🧰 Improvements

  • Improve default suffix handling in FIM inference. #400

πŸ’« New Contributors

v0.1.0-rc.1

12 Sep 15:12
30afa19
Compare
Choose a tag to compare
v0.1.0-rc.1 Pre-release
Pre-release
feat: add LLAMA_CPP_LOG_LEVEL to control log level of llama.cpp (#436)

v0.1.0-rc.0

12 Sep 06:52
3af32c8
Compare
Choose a tag to compare
v0.1.0-rc.0 Pre-release
Pre-release

πŸ“¦ Artifacts

πŸš€ Features

llama.cpp on Apple M1/M2 for metal inference backend

  • feat: llama.cpp for metal support [TAB-146] by @wsxiaoys in #391
  • feat: support cancellation in llama backend [TAB-146] by @wsxiaoys in #392
  • feat: tune llama metal backend performance by @wsxiaoys in #393
  • fix: ensure default suffix to be non-empty by @wsxiaoys in #400
  • feat: turn on metal device by default on macosx / aarch64 devices by @wsxiaoys in #398
  • feat: implement input truncation with options.max_input_length by @wsxiaoys in #415
  • feat: implement input truncation for llama-cpp-bindings by @wsxiaoys in #416

Experimental support of http api backend.

🧰 Improvements

  • Improve default suffix handling in FIM inference. #400

πŸ’« New Contributors

Full Changelog: v0.0.1...v0.1.0-rc.0

v0.1.0

12 Sep 16:16
30afa19
Compare
Choose a tag to compare
v0.1.0 Pre-release
Pre-release

πŸ“¦ Artifacts

πŸš€ Features

🧰 Improvements

  • Improve default suffix handling in FIM inference. #400

πŸ’« New Contributors

v0.0.1

31 Aug 02:19
Compare
Choose a tag to compare

πŸ“¦ Artifacts

πŸš€ Features

  • Support FIM inference.
  • Initial support of indexing with tabby scheduler.
  • Support CodeLlama series model.

🧰 Improvements

  • Support early cancellation to reduce GPU workload, increasing capacity of tabby server.

πŸ’« New Contributors

v0.0.1-rc.2

30 Aug 16:18
Compare
Choose a tag to compare
v0.0.1-rc.2 Pre-release
Pre-release

πŸ“¦ Artifacts

πŸš€ Features

  • Support FIM inference.
  • Initial support of indexing with tabby scheduler.
  • Support CodeLlama series model.

🧰 Improvements

  • Support early cancellation to reduce GPU workload, increasing capacity of tabby server.

πŸ’« New Contributors