Skip to content

Commit

Permalink
Browse files Browse the repository at this point in the history
  • Loading branch information
MeouSker77 authored and cyita committed Sep 5, 2024
1 parent 1919281 commit ceb839c
Showing 1 changed file with 3 additions and 3 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -151,7 +151,7 @@ def Llama2_7B_gpu_model(self, Name, Model, Tokenizer, model_path):
# currently only compare the output of the last self-attention layer.
layer_norm = "model.layers.31.input_layernorm"
self_attn = "model.layers.31.self_attn"
lower_bound = 8e-3
lower_bound = 2e-2
self.run_optimize_gpu_model(Name, Model, Tokenizer, model_path, self_attn, layer_norm, lower_bound)

def Falcon_7B_gpu_model(self, Name, Model, Tokenizer, model_path):
Expand All @@ -165,7 +165,7 @@ def Chatglm2_gpu_model(self, Name, Model, Tokenizer, model_path):
# currently only need to compare the output of one self-attention layer.
layer_norm = "transformer.encoder.layers.27.input_layernorm"
self_attn = "transformer.encoder.layers.27.self_attention"
lower_bound = 4e-2
lower_bound = 1e-1
self.run_optimize_gpu_model(Name, Model, Tokenizer, model_path, self_attn, layer_norm, lower_bound)

def Mistral_gpu_model(self, Name, Model, Tokenizer, model_path):
Expand All @@ -182,7 +182,7 @@ def Baichuan_gpu_model(self, Name, Model, Tokenizer, model_path):
# currently only need to compare the output of one self-attention layer.
layer_norm = "model.layers.31.input_layernorm"
self_attn = "model.layers.31.self_attn"
lower_bound = 8e-3
lower_bound = 2e-2
self.run_optimize_gpu_model(Name, Model, Tokenizer, model_path, self_attn, layer_norm, lower_bound)

def Qwen_gpu_model(self, Name, Model, Tokenizer, model_path):
Expand Down

0 comments on commit ceb839c

Please sign in to comment.