File tree 1 file changed +5
-3
lines changed
1 file changed +5
-3
lines changed Original file line number Diff line number Diff line change @@ -52,7 +52,7 @@ def prepare_model(self):
52
52
params_file , str
53
53
), "invalid input for --llama_inputs"
54
54
else :
55
- logging .warning (
55
+ logger .warning (
56
56
"Skipping Llama test because of lack of input. To run use --llama_inputs <.pt> <.json>"
57
57
)
58
58
return None , None , None
@@ -61,6 +61,8 @@ def prepare_model(self):
61
61
params_file
62
62
), "Invalid file paths"
63
63
64
+ logger .info ("Running test_llama.py" )
65
+
64
66
# TODO: Enable key value cache
65
67
args = [
66
68
"--disable_dynamic_shape" ,
@@ -112,9 +114,9 @@ def test_llama_tosa_MI(self):
112
114
)
113
115
.export ()
114
116
.to_edge_transform_and_lower ()
115
- .check_count ({"torch.ops.higher_order.executorch_call_delegate" : 14 })
117
+ .check_count ({"torch.ops.higher_order.executorch_call_delegate" : 26 })
116
118
.to_executorch ()
117
119
.run_method_and_compare_outputs (
118
- inputs = llama_inputs , atol = 1.8 , rtol = 0.01 # TODO: decrease tolerance
120
+ inputs = llama_inputs , atol = 4.3 , rtol = 1.1 # TODO: MLETORCH-825 decrease tolerance
119
121
)
120
122
)
You can’t perform that action at this time.
0 commit comments