mirror of
https://github.com/azaion/annotations.git
synced 2026-04-22 10:16:30 +00:00
fix converting model initialization
This commit is contained in:
@@ -100,14 +100,14 @@ cdef class TensorRTEngine(InferenceEngine):
|
||||
return None
|
||||
|
||||
if builder.platform_has_fast_fp16:
|
||||
constants_inf.log('Converting to supported fp16')
|
||||
constants_inf.log(<str>'Converting to supported fp16')
|
||||
config.set_flag(trt.BuilderFlag.FP16)
|
||||
else:
|
||||
constants_inf.log('Converting to supported fp32. (fp16 is not supported)')
|
||||
constants_inf.log(<str>'Converting to supported fp32. (fp16 is not supported)')
|
||||
plan = builder.build_serialized_network(network, config)
|
||||
|
||||
if plan is None:
|
||||
constants_inf.logerror('Conversion failed.')
|
||||
constants_inf.logerror(<str>'Conversion failed.')
|
||||
return None
|
||||
constants_inf.log('conversion done!')
|
||||
return bytes(plan)
|
||||
|
||||
Reference in New Issue
Block a user