-
Notifications
You must be signed in to change notification settings - Fork 557
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When using 'run.bat' an error is shown regarding the 'gpt4all-lora-quantized-ggml.bin' file being 'invalid model file' #53
Comments
uninstall gpt4all from uninstall.bat |
I followed your steps. However 'uninstall.bat' only uninstalls the virtual environment. GPT4ALL-UI folder remains. So I deleted it manually and installed a fresh copy from repository (downloaded a Zip as you said). |
The problem is with this line in the setup script: Asked ChatGPT about it and it seams the -v command isn´t suitable for windows. Mostly used in Unix systems. Got it working by manual skipping that step and doing this: After that the run batch worked fine. |
If this pull request gets accepted, all you have to do is to convert the ggml model with install.bat, and then go to run.bat normally |
Expected Behavior
When using 'run.bat' on Windows 10 machine, the previously downloaded model should be recognized as valid.
Current Behavior
When using 'run.bat' an error is shown regarding the 'gpt4all-lora-quantized-ggml.bin' being 'invalid model file'. Although the newest model file has been downloaded the same day (during the installation of GPT4ALL-UI).
Steps to Reproduce
Please provide detailed steps to reproduce the issue.
"Checking discussions database...
Ok
llama_model_load: loading model from './models/gpt4all-lora-quantized-ggml.bin' - please wait ...
./models/gpt4all-lora-quantized-ggml.bin: invalid model file (bad magic [got 0x67676d66 want 0x67676a74])
you most likely need to regenerate your ggml files
the benefit is you'll get 10-100x faster load times
see Should use
mmap
for model loading ggerganov/llama.cpp#91use convert-pth-to-ggml.py to regenerate from original pth
use migrate-ggml-2023-03-30-pr613.py if you deleted originals
llama_init_from_file: failed to load model
llama_generate: seed = 1680983129
system_info: n_threads = 8 / 12 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 |"
Screenshots
See attached screenshot for an example.
The text was updated successfully, but these errors were encountered: