diff --git a/models/Model_AMP/Model.py b/models/Model_AMP/Model.py index 28c0420..46d065a 100644 --- a/models/Model_AMP/Model.py +++ b/models/Model_AMP/Model.py @@ -125,7 +125,10 @@ class AMPModel(ModelBase): if ct_mode == 'none': ct_mode = None - use_fp16 = self.is_exporting + use_fp16 = False + if self.is_exporting: + use_fp16 = io.input_bool ("Export quantized?", False, help_message='Makes the exported model faster. If you have problems, disable this option.') + conv_dtype = tf.float16 if use_fp16 else tf.float32 class Downscale(nn.ModelBase): diff --git a/models/Model_SAEHD/Model.py b/models/Model_SAEHD/Model.py index 4dcecd9..dda4a4b 100644 --- a/models/Model_SAEHD/Model.py +++ b/models/Model_SAEHD/Model.py @@ -219,7 +219,10 @@ Examples: df, liae, df-d, df-ud, liae-ud, ... self.set_iter(0) adabelief = self.options['adabelief'] - use_fp16 = False#self.options['use_fp16'] + + use_fp16 = False + if self.is_exporting: + use_fp16 = io.input_bool ("Export quantized?", False, help_message='Makes the exported model faster. If you have problems, disable this option.') self.gan_power = gan_power = 0.0 if self.pretrain else self.options['gan_power'] random_warp = False if self.pretrain else self.options['random_warp']