You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"allow_gpu_fallback": false, # (DLA only) Allow layers unsupported on DLA to run on GPU
202
206
},
203
207
"op_precision": torch.half, # Operating precision set to FP16
208
+
"disable_tf32": False, # Force FP32 layers to use traditional as FP32 format vs the default behavior of rounding the inputs to 10-bit mantissas before multiplying, but accumulates the sum using 23-bit mantissas
204
209
"refit": False, # enable refit
205
210
"debug": False, # enable debuggable engine
206
211
"strict_types": False, # kernels should strictly run in operating precision
Copy file name to clipboardExpand all lines: py/trtorch/_compiler.py
+1Lines changed: 1 addition & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -99,6 +99,7 @@ def convert_method_to_trt_engine(module: torch.jit.ScriptModule, method_name: st
99
99
"allow_gpu_fallback": false, # (DLA only) Allow layers unsupported on DLA to run on GPU
100
100
},
101
101
"op_precision": torch.half, # Operating precision set to FP16
102
+
"disable_tf32": False, # Force FP32 layers to use traditional as FP32 format vs the default behavior of rounding the inputs to 10-bit mantissas before multiplying, but accumulates the sum using 23-bit mantissas
102
103
"refit": false, # enable refit
103
104
"debug": false, # enable debuggable engine
104
105
"strict_types": false, # kernels should strictly run in operating precision
0 commit comments