You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
he following flags have been taken from the environment variable 'OOBABOOGA_FLAGS':
--fkdlsja >nul 2>&1 & python bot.py --token --chat --model-menu
To use the CMD_FLAGS Inside webui.py, unset 'OOBABOOGA_FLAGS'.
bin E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.dll
E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\bitsandbytes\cextension.py:34: UserWarning: The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.
warn("The installed version of bitsandbytes was compiled without GPU support. "
function 'cadam32bit_grad_fp32' not found
The following models are available:
INFO:Loading the extension "gallery"...
[2023-06-13 19:37:45] [INFO ] discord.client: logging in using static token
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 803, in static_login
data = await self.request(Route('GET', '/users/@me'))
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 745, in request
raise HTTPException(response, data)
discord.errors.HTTPException: 401 Unauthorized (error code: 0): 401: Unauthorized
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\text-generation-webui\bot.py", line 544, in
client.run(bot_args.token if bot_args.token else TOKEN)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 860, in run
asyncio.run(runner())
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\asyncio\runners.py", line 44, in run
return loop.run_until_complete(main)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\asyncio\base_events.py", line 649, in run_until_complete
return future.result()
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 849, in runner
await self.start(token, reconnect=reconnect)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 777, in start
await self.login(token)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 612, in login
data = await self.http.static_login(token)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 807, in static_login
raise LoginFailure('Improper token has been passed.') from exc
discord.errors.LoginFailure: Improper token has been passed.
Exception ignored in: <function LlamaCppModel.del at 0x000001CFE86E7AC0>
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\text-generation-webui\modules\llamacpp_model.py", line 23, in del
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\llama_cpp\llama.py", line 1334, in del
TypeError: 'NoneType' object is not callable
Press any key to continue . . .
The script runs fine when using 'python bot.py' directly using the cmd_window.bat on oobabooga_windows.bat but whenever I run it through the bat file it would cause this error.
The text was updated successfully, but these errors were encountered:
he following flags have been taken from the environment variable 'OOBABOOGA_FLAGS':
--fkdlsja >nul 2>&1 & python bot.py --token --chat --model-menu
To use the CMD_FLAGS Inside webui.py, unset 'OOBABOOGA_FLAGS'.
bin E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cpu.dll
E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\bitsandbytes\cextension.py:34: UserWarning: The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.
warn("The installed version of bitsandbytes was compiled without GPU support. "
function 'cadam32bit_grad_fp32' not found
The following models are available:
Which one do you want to load? 1-6
2
INFO:Loading ggml-alpaca-13b-x-gpt-4-q4_0.bin...
INFO:llama.cpp weights detected: models\ggml-alpaca-13b-x-gpt-4-q4_0.bin
INFO:Cache capacity is 0 bytes
llama.cpp: loading model from models\ggml-alpaca-13b-x-gpt-4-q4_0.bin
llama_model_load_internal: format = ggjt v3 (latest)
llama_model_load_internal: n_vocab = 32001
llama_model_load_internal: n_ctx = 2048
llama_model_load_internal: n_embd = 5120
llama_model_load_internal: n_mult = 256
llama_model_load_internal: n_head = 40
llama_model_load_internal: n_layer = 40
llama_model_load_internal: n_rot = 128
llama_model_load_internal: ftype = 2 (mostly Q4_0)
llama_model_load_internal: n_ff = 13824
llama_model_load_internal: n_parts = 1
llama_model_load_internal: model size = 13B
llama_model_load_internal: ggml ctx size = 0.09 MB
llama_model_load_internal: mem required = 9031.71 MB (+ 1608.00 MB per state)
.
llama_init_from_file: kv self size = 1600.00 MB
AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 |
INFO:Loaded the model in 45.33 seconds.
INFO:Loading the extension "gallery"...
[2023-06-13 19:37:45] [INFO ] discord.client: logging in using static token
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 803, in static_login
data = await self.request(Route('GET', '/users/@me'))
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 745, in request
raise HTTPException(response, data)
discord.errors.HTTPException: 401 Unauthorized (error code: 0): 401: Unauthorized
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\text-generation-webui\bot.py", line 544, in
client.run(bot_args.token if bot_args.token else TOKEN)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 860, in run
asyncio.run(runner())
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\asyncio\runners.py", line 44, in run
return loop.run_until_complete(main)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\asyncio\base_events.py", line 649, in run_until_complete
return future.result()
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 849, in runner
await self.start(token, reconnect=reconnect)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 777, in start
await self.login(token)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\client.py", line 612, in login
data = await self.http.static_login(token)
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\discord\http.py", line 807, in static_login
raise LoginFailure('Improper token has been passed.') from exc
discord.errors.LoginFailure: Improper token has been passed.
Exception ignored in: <function LlamaCppModel.del at 0x000001CFE86E7AC0>
Traceback (most recent call last):
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\text-generation-webui\modules\llamacpp_model.py", line 23, in del
File "E:\etc\bot\ChatLLaMA\oobabooga_windows\installer_files\env\lib\site-packages\llama_cpp\llama.py", line 1334, in del
TypeError: 'NoneType' object is not callable
Press any key to continue . . .
The script runs fine when using 'python bot.py' directly using the cmd_window.bat on oobabooga_windows.bat but whenever I run it through the bat file it would cause this error.
The text was updated successfully, but these errors were encountered: