Breaking Changes
Upgrading
To upgrade from a prior version while still keeping your databases and models that were downloaded, perform the following steps:
- In your directory that contains all of the files currently...delete the
Include
, Lib
, Scripts
folders. (pertain to your virtual environment)
- Delete the
pyvenv.cfg
file. (pertains to your virtual environment)
- At this point, any and all files pertaining to your virtual environment are gone.
- Delete the
Assets
and CSS
folders. (new ones will be provided in their entirety further below)
- Extract all the files from the
.zip
file for this release v7.0.0 and go into the src
folder.
- Copy all files and folders EXCEPT THE CONFIG.YAML FILE into the aforementioned folder where you virtual environment USED TO BE. If there are any prompts indicating that files will be overwritten, CLICK YES since we're only updating necessary files now.
- EXCEPT THE CONFIG.YAML FILE!
- EXCEPT THE CONFIG.YAML FILE!
- last time...EXCEPT THE CONFIG.YAML FILE!
Afterwards, open a command prompt and run the installation instructions on this repository's page, namely:
- Open a command prompt in the directory and run:
python -m venv .
and .\Scripts\activate
- Then based on your Python version (we now support Python 3.11 and 3.12) run:
python setup_windows_cp311.py
or python setup_windows_cp312.py
Python 3.11 and Python 3.12 support
- Just make sure and run the proper installation scripts.
Bug Fixes
Embedding Models.
Change |
Details |
Notes |
Added |
granite - 30m and granite - 125m |
All are excellent for RAG |
Added |
arctic-embed-m and arctic-embed-l |
Also excellent for RAG |
Removed |
all-MiniLM-L12-v2 , all-mpnet-base-v2 , gte-small , gte-base , and gte-large |
All superseded by either granite or arctic-embed models. |
Chat Models
Change |
Details |
Notes |
Added |
granite - 2b and granite - 8b |
All seem to be destined for RAG |
Added |
exaone - 2.4b , exaone - 7.8b , and exaone - 32b |
Great for RAG |
Removed |
Qwen - 3b , Llama 3.2 - 3b , Phi 3.5 Mini - 4b , MiniCPM3 - 4b , Qwen - 7b , Dolphin-Llama 3.1 - 8b , Marco-o1 - 7b |
All superseded in quality, speed, or both. |
Vision Models
Change |
Details |
Notes |
Added |
InternVL 2.5 - 1b and InternVL 2.5 - 4b |
Very good with low resources even |
Removed |
Llava 1.6 - Vicuna and MiniCPM 2.6 - 8b |
Beaten by InternVL 2.5 models. |
Ask Jeeves!
- When Jeeves first loads you now have the ability to choose his "brain," which can be any of the following GGUF models:
EXAONE - 2.4b Q8_0
, EXAONE - 7.8b Q4_K_M
, Llama 3.2 Q8_0
, Qwen - 3b Q8_0
or Zephyr - 3b Q8_0
- Three suggested queries are populated as you type. Click to automatically ask Jeeves that specific question.
- Darker color theme.
Documentation Scraper
- Updated libraries to scrape and the .html file created is more pristine - i.e. less sidebars, menus, etc. - you should now get higher quality .html files, which, in turn, allows for higher quality vector database searches.
- More scraping optimizations to come since there are so many libraries to customize the scraper for.
Updated Dependencies
torch 2.5.1
is now used, which finally allows this program to be compatible with Python 3.12.
CUDA
, triton
, bitsandbytes
, flash attention 2
, langchain
, and a ton of other dependencies updated.
There are a host of other improvements but this is long enough!