-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Remove non-HF ExLlamaV2 loader (#5431)
- Loading branch information
Showing
9 changed files
with
18 additions
and
220 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
cde000d
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why remove it? This loader provided better performance than the HF one, especially on larger models it was sometimes more than 10% faster.
cde000d
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why did you delete that? even ExLlamav2_HF doesn't work any more...
cde000d
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This was already reverted, see the linked issue. What do you mean by it not working anymore? Both loaders appear to work fine for me on the current state of the dev branch.
cde000d
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
First time i updated it normal clicking the update bottom, then i realized that ExLlamav2_HF doesn't work any more, after you told me about the dev banch i tested the Dev branch and doesn't work either, did a clean install, doesn't work either, it says this here ->
Traceback (most recent call last):
File "C:\AI\oo\text-generation-webui\modules\ui_model_menu.py", line 220, in load_model_wrapper
shared.model, shared.tokenizer = load_model(selected_model, loader)
File "C:\AI\oo\text-generation-webui\modules\models.py", line 87, in load_model
output = load_func_maploader
File "C:\AI\oo\text-generation-webui\modules\models.py", line 387, in ExLlamav2_HF_loader
from modules.exllamav2_hf import Exllamav2HF
File "C:\AI\oo\text-generation-webui\modules\exllamav2_hf.py", line 7, in
from exllamav2 import (
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2_init_.py", line 3, in
from exllamav2.model import ExLlamaV2
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\model.py", line 16, in
from exllamav2.config import ExLlamaV2Config
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\config.py", line 2, in
from exllamav2.fasttensors import STFile
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\fasttensors.py", line 5, in
from exllamav2.ext import exllamav2_ext as ext_c
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\ext.py", line 142, in
exllamav2_ext = load \
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\torch\utils\cpp_extension.py", line 1308, in load
return _jit_compile(
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\torch\utils\cpp_extension.py", line 1736, in _jit_compile
return _import_module_from_library(name, build_directory, is_python_module)
File "C:\AI\oo\text-generation-webui\installer_files\env\Lib\site-packages\torch\utils\cpp_extension.py", line 2136, in _import_module_from_library
module = importlib.util.module_from_spec(spec)
ImportError: DLL load failed while importing exllamav2_ext: The specified module was not found.
cde000d
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hmm, I only run it on linux servers, so I've never seen that error before. Probably best to move that to a separate issue on github to get more eyes on it.