From 3bb8d4f0b29b68293cfb37eacc6cd98ba149833a Mon Sep 17 00:00:00 2001 From: Glenn Jocher Date: Sat, 14 Aug 2021 13:47:20 +0200 Subject: [PATCH] Remove `encoding='ascii'` (#4413) * Remove `encoding='ascii'` * Reinstate `encoding='ascii'` in emojis() --- utils/autoanchor.py | 2 +- utils/datasets.py | 2 +- utils/general.py | 4 ++-- utils/loggers/wandb/wandb_utils.py | 6 +++--- 4 files changed, 7 insertions(+), 7 deletions(-) diff --git a/utils/autoanchor.py b/utils/autoanchor.py index 2571fc99ac89..eca1d5be8ebe 100644 --- a/utils/autoanchor.py +++ b/utils/autoanchor.py @@ -104,7 +104,7 @@ def print_results(k): return k if isinstance(dataset, str): # *.yaml file - with open(dataset, encoding='ascii', errors='ignore') as f: + with open(dataset, errors='ignore') as f: data_dict = yaml.safe_load(f) # model dict from utils.datasets import LoadImagesAndLabels dataset = LoadImagesAndLabels(data_dict['train'], augment=True, rect=True) diff --git a/utils/datasets.py b/utils/datasets.py index 1c780cdbac4b..b402723f9c49 100755 --- a/utils/datasets.py +++ b/utils/datasets.py @@ -931,7 +931,7 @@ def hub_ops(f, max_dim=1920): im.save(im_dir / Path(f).name, quality=75) # save zipped, data_dir, yaml_path = unzip(Path(path)) - with open(check_file(yaml_path), encoding='ascii', errors='ignore') as f: + with open(check_file(yaml_path), errors='ignore') as f: data = yaml.safe_load(f) # data dict if zipped: data['path'] = data_dir # TODO: should this be dir.resolve()? diff --git a/utils/general.py b/utils/general.py index 15111b727f33..c70c21f47636 100755 --- a/utils/general.py +++ b/utils/general.py @@ -112,7 +112,7 @@ def is_pip(): def emojis(str=''): # Return platform-dependent emoji-safe version of string - return str.encode().decode('ascii', 'ignore') if platform.system() == 'Windows' else str + return str.encode().decode(encoding='ascii', errors='ignore') if platform.system() == 'Windows' else str def file_size(file): @@ -250,7 +250,7 @@ def check_dataset(data, autodownload=True): # Read yaml (optional) if isinstance(data, (str, Path)): - with open(data, encoding='ascii', errors='ignore') as f: + with open(data, errors='ignore') as f: data = yaml.safe_load(f) # dictionary # Parse yaml diff --git a/utils/loggers/wandb/wandb_utils.py b/utils/loggers/wandb/wandb_utils.py index 3f2684a7f3e3..019aebf094e1 100644 --- a/utils/loggers/wandb/wandb_utils.py +++ b/utils/loggers/wandb/wandb_utils.py @@ -62,7 +62,7 @@ def check_wandb_resume(opt): def process_wandb_config_ddp_mode(opt): - with open(check_file(opt.data), encoding='ascii', errors='ignore') as f: + with open(check_file(opt.data), errors='ignore') as f: data_dict = yaml.safe_load(f) # data dict train_dir, val_dir = None, None if isinstance(data_dict['train'], str) and data_dict['train'].startswith(WANDB_ARTIFACT_PREFIX): @@ -152,7 +152,7 @@ def __init__(self, opt, run_id, job_type='Training'): self.wandb_artifact_data_dict = self.check_and_upload_dataset(opt) elif opt.data.endswith('_wandb.yaml'): # When dataset is W&B artifact - with open(opt.data, encoding='ascii', errors='ignore') as f: + with open(opt.data, errors='ignore') as f: data_dict = yaml.safe_load(f) self.data_dict = data_dict else: # Local .yaml dataset file or .zip file @@ -186,7 +186,7 @@ def check_and_upload_dataset(self, opt): opt.single_cls, 'YOLOv5' if opt.project == 'runs/train' else Path(opt.project).stem) print("Created dataset config file ", config_path) - with open(config_path, encoding='ascii', errors='ignore') as f: + with open(config_path, errors='ignore') as f: wandb_data_dict = yaml.safe_load(f) return wandb_data_dict