mirror of
https://github.com/iperov/DeepFaceLab.git
synced 2025-07-14 09:03:41 -07:00
converter: removed 'ebs' color transfer
This commit is contained in:
parent
71a98e2ee4
commit
e65e74ac2d
2 changed files with 4 additions and 4 deletions
|
@ -208,7 +208,7 @@ def ConvertMaskedFace (predictor_func, predictor_input_shape, cfg, frame_info, i
|
||||||
elif cfg.color_transfer_mode == 6: #idt-m
|
elif cfg.color_transfer_mode == 6: #idt-m
|
||||||
prd_face_bgr = imagelib.color_transfer_idt (prd_face_bgr*prd_face_mask_a, dst_face_bgr*prd_face_mask_a)
|
prd_face_bgr = imagelib.color_transfer_idt (prd_face_bgr*prd_face_mask_a, dst_face_bgr*prd_face_mask_a)
|
||||||
|
|
||||||
elif cfg.color_transfer_mode == 7: #ebs
|
elif cfg.color_transfer_mode == 7: #ebs, currently unused
|
||||||
prd_face_bgr = cfg.ebs_ct_func ( np.clip( (dst_face_bgr*255), 0, 255).astype(np.uint8),
|
prd_face_bgr = cfg.ebs_ct_func ( np.clip( (dst_face_bgr*255), 0, 255).astype(np.uint8),
|
||||||
np.clip( (prd_face_bgr*255), 0, 255).astype(np.uint8), )#prd_face_mask_a
|
np.clip( (prd_face_bgr*255), 0, 255).astype(np.uint8), )#prd_face_mask_a
|
||||||
prd_face_bgr = np.clip( prd_face_bgr.astype(np.float32) / 255.0, 0.0, 1.0)
|
prd_face_bgr = np.clip( prd_face_bgr.astype(np.float32) / 255.0, 0.0, 1.0)
|
||||||
|
|
|
@ -107,8 +107,8 @@ half_face_mask_mode_dict = {1:'learned',
|
||||||
4:'FAN-dst',
|
4:'FAN-dst',
|
||||||
7:'learned*FAN-dst'}
|
7:'learned*FAN-dst'}
|
||||||
|
|
||||||
ctm_dict = { 0: "None", 1:"rct", 2:"lct", 3:"mkl", 4:"mkl-m", 5:"idt", 6:"idt-m", 7:"ebs" }
|
ctm_dict = { 0: "None", 1:"rct", 2:"lct", 3:"mkl", 4:"mkl-m", 5:"idt", 6:"idt-m" }
|
||||||
ctm_str_dict = {None:0, "rct":1, "lct":2, "mkl":3, "mkl-m":4, "idt":5, "idt-m":6, "ebs":7 }
|
ctm_str_dict = {None:0, "rct":1, "lct":2, "mkl":3, "mkl-m":4, "idt":5, "idt-m":6 }
|
||||||
|
|
||||||
class ConverterConfigMasked(ConverterConfig):
|
class ConverterConfigMasked(ConverterConfig):
|
||||||
|
|
||||||
|
@ -221,7 +221,7 @@ class ConverterConfigMasked(ConverterConfig):
|
||||||
self.output_face_scale = np.clip (io.input_int ("Choose output face scale modifier [-50..50] (skip:0) : ", 0), -50, 50)
|
self.output_face_scale = np.clip (io.input_int ("Choose output face scale modifier [-50..50] (skip:0) : ", 0), -50, 50)
|
||||||
|
|
||||||
if 'raw' not in self.mode:
|
if 'raw' not in self.mode:
|
||||||
self.color_transfer_mode = io.input_str ("Apply color transfer to predicted face? Choose mode ( rct/lct/ebs skip:None ) : ", None, ctm_str_dict.keys() )
|
self.color_transfer_mode = io.input_str ( f"Apply color transfer to predicted face? Choose mode ( {' / '.join ([str(x) for x in list(ctm_str_dict.keys())])} skip:None ) : ", None, ctm_str_dict.keys() )
|
||||||
self.color_transfer_mode = ctm_str_dict[self.color_transfer_mode]
|
self.color_transfer_mode = ctm_str_dict[self.color_transfer_mode]
|
||||||
|
|
||||||
super().ask_settings()
|
super().ask_settings()
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue