当前位置:网站首页>[deep learning] [original] let yolov6-0.1.0 support the txt reading dataset mode of yolov5
[deep learning] [original] let yolov6-0.1.0 support the txt reading dataset mode of yolov5
2022-07-05 16:29:00 【FL1623863129】
Meituan gave a yolov6 The framework looks very good at present , Because I didn't come out for long , Many are not perfect . Today, I specially trained my data set and found that this framework can only be placed according to this pattern :
custom_dataset
├── images
│ ├── train
│ │ ├── train0.jpg
│ │ └── train1.jpg
│ ├── val
│ │ ├── val0.jpg
│ │ └── val1.jpg
│ └── test
│ ├── test0.jpg
│ └── test1.jpg
└── labels
├── train
│ ├── train0.txt
│ └── train1.txt
├── val
│ ├── val0.txt
│ └── val1.txt
└── test
├── test0.txt
└── test1.txtAnd I prefer yolov5 The pattern of , Of course yolov5 It also supports the above placement mode .
images-
1.jpg
2.jpg
......
labels-
1.txt
2.txt
.......
Then put the split data set txt Inside
train.txt
/home/fut/data/images/1.jpg
/home/fut/data/images/2.jpg
....
val.txt
/home/fut/data/images/6.jpg
/home/fut/data/images/7.jpg
....
Configure in the configuration file :
train: myproj/config/train.txt
val: myproj/config/val.txt
nc: 2
# whether it is coco dataset, only coco dataset should be set to True.
is_coco: False
# class names
names: ['dog','cat']So you don't have to cut four folders at a time . Don't talk much, start changing the code , We turn on YOLOv6-0.1.0/yolov6/data/datasets.py modify
def get_imgs_labels(self, img_dir): This function loads the mode . Here is the modified complete code of this function
def get_imgs_labels(self, img_dir):
NUM_THREADS = min(8, os.cpu_count())
if os.path.isdir(img_dir):
valid_img_record = osp.join(
osp.dirname(img_dir), "." + osp.basename(img_dir) + ".json"
)
img_paths = glob.glob(osp.join(img_dir, "*"), recursive=True)
img_paths = sorted(
p for p in img_paths if p.split(".")[-1].lower() in IMG_FORMATS
)
assert img_paths, f"No images found in {img_dir}."
else:
with open(img_dir,'r') as f:
img_paths = f.read().rstrip('\n').split('\n')
valid_img_record = os.path.dirname(img_dir)+os.sep+'.'+osp.basename(img_dir)[:-4] + ".json"
img_hash = self.get_hash(img_paths)
if osp.exists(valid_img_record):
with open(valid_img_record, "r") as f:
cache_info = json.load(f)
if "image_hash" in cache_info and cache_info["image_hash"] == img_hash:
img_info = cache_info["information"]
else:
self.check_images = True
else:
self.check_images = True
# check images
if self.check_images and self.main_process:
img_info = {}
nc, msgs = 0, [] # number corrupt, messages
LOGGER.info(
f"{self.task}: Checking formats of images with {NUM_THREADS} process(es): "
)
with Pool(NUM_THREADS) as pool:
pbar = tqdm(
pool.imap(TrainValDataset.check_image, img_paths),
total=len(img_paths),
)
for img_path, shape_per_img, nc_per_img, msg in pbar:
if nc_per_img == 0: # not corrupted
img_info[img_path] = {"shape": shape_per_img}
nc += nc_per_img
if msg:
msgs.append(msg)
pbar.desc = f"{nc} image(s) corrupted"
pbar.close()
if msgs:
LOGGER.info("\n".join(msgs))
cache_info = {"information": img_info, "image_hash": img_hash}
# save valid image paths.
with open(valid_img_record, "w") as f:
json.dump(cache_info, f)
# # check and load anns
# label_dir = osp.join(
# osp.dirname(osp.dirname(img_dir)), "coco", osp.basename(img_dir)
# )
# assert osp.exists(label_dir), f"{label_dir} is an invalid directory path!"
img_paths = list(img_info.keys())
label_dir = os.path.dirname(img_paths[0]).replace('images', 'labels')
label_paths = sorted(
osp.join(label_dir, osp.splitext(osp.basename(p))[0] + ".txt")
for p in img_paths
)
label_hash = self.get_hash(label_paths)
if "label_hash" not in cache_info or cache_info["label_hash"] != label_hash:
self.check_labels = True
if self.check_labels:
cache_info["label_hash"] = label_hash
nm, nf, ne, nc, msgs = 0, 0, 0, 0, [] # number corrupt, messages
LOGGER.info(
f"{self.task}: Checking formats of labels with {NUM_THREADS} process(es): "
)
with Pool(NUM_THREADS) as pool:
pbar = pool.imap(
TrainValDataset.check_label_files, zip(img_paths, label_paths)
)
pbar = tqdm(pbar, total=len(label_paths)) if self.main_process else pbar
for (
img_path,
labels_per_file,
nc_per_file,
nm_per_file,
nf_per_file,
ne_per_file,
msg,
) in pbar:
if nc_per_file == 0:
img_info[img_path]["labels"] = labels_per_file
else:
img_info.pop(img_path)
nc += nc_per_file
nm += nm_per_file
nf += nf_per_file
ne += ne_per_file
if msg:
msgs.append(msg)
if self.main_process:
pbar.desc = f"{nf} label(s) found, {nm} label(s) missing, {ne} label(s) empty, {nc} invalid label files"
if self.main_process:
pbar.close()
with open(valid_img_record, "w") as f:
json.dump(cache_info, f)
if msgs:
LOGGER.info("\n".join(msgs))
if nf == 0:
LOGGER.warning(
f"WARNING: No labels found in {osp.dirname(self.img_paths[0])}. "
)
if self.task.lower() == "val":
if self.data_dict.get("is_coco", False): # use original json file when evaluating on coco dataset.
assert osp.exists(self.data_dict["anno_path"]), "Eval on coco dataset must provide valid path of the annotation file in config file: data/coco.yaml"
else:
assert (
self.class_names
), "Class names is required when converting labels to coco format for evaluating."
save_dir = osp.join(osp.dirname(osp.dirname(img_dir)), "annotations")
if not osp.exists(save_dir):
os.mkdir(save_dir)
save_path = osp.join(
save_dir, "instances_" + osp.basename(img_dir) + ".json"
)
TrainValDataset.generate_coco_format_labels(
img_info, self.class_names, save_path
)
img_paths, labels = list(
zip(
*[
(
img_path,
np.array(info["labels"], dtype=np.float32)
if info["labels"]
else np.zeros((0, 5), dtype=np.float32),
)
for img_path, info in img_info.items()
]
)
)
self.img_info = img_info
LOGGER.info(
f"{self.task}: Final numbers of valid images: {len(img_paths)}/ labels: {len(labels)}. "
)
return img_paths, labels
边栏推荐
- Exception com alibaba. fastjson. JSONException: not match : - =
- Data Lake (XIV): spark and iceberg integrated query operation
- 求解汉诺塔问题【修改版】
- 利用GrayLog告警功能实现钉钉群机器人定时工作提醒
- 漫画:什么是八皇后问题?
- 《21天精通TypeScript-3》-安装搭建TypeScript开发环境.md
- Cartoon: what is the eight queens problem?
- The list set is summed up according to a certain attribute of the object, the maximum value, etc
- Boost the development of digital economy and consolidate the base of digital talents - the digital talent competition was successfully held in Kunming
- 事务回滚异常
猜你喜欢

Subclasses and superclasses of abstract classes

ES6深入—ES6 Class 类

单商户 V4.4,初心未变,实力依旧!

Replknet: it's not that large convolution is bad, but that convolution is not large enough. 31x31 convolution. Let's have a look at | CVPR 2022

vant tabbar遮挡内容的解决方式

数据湖(十四):Spark与Iceberg整合查询操作

List de duplication and count the number

超分辨率技术在实时音视频领域的研究与实践

HiEngine:可媲美本地的云原生内存数据库引擎

Domestic API management artifact used by the company
随机推荐
[js] 技巧 简化if 判空
单商户 V4.4,初心未变,实力依旧!
怎样在电脑上设置路由器的WiFi密码
Dataarts studio data architecture - Introduction to data standards
RLock锁的使用
极坐标扇图使用场景与功能详解
abstract关键字和哪些关键字会发生冲突呢
10 minutes to help you get ZABBIX monitoring platform alarm pushed to nail group
One click installation script enables rapid deployment of graylog server 4.2.10 stand-alone version
list去重并统计个数
ES6 drill down - Async functions and symbol types
记一次'非常诡异'的云安全组规则问题排查过程
vulnhub-FirstBlood
Exception com alibaba. fastjson. JSONException: not match : - =
tf.sequence_mask函数讲解案例
给自己打打气
漫画:什么是分布式事务?
事务回滚异常
详解SQL中Groupings Sets 语句的功能和底层实现逻辑
Use of RLOCK lock