| import os |
| from tqdm import tqdm |
| import json |
|
|
| meta_file = '/etc/ssd1/jiangzhongtao/baai_embedding_tune/data/task_meta/casual/all.json' |
| with open(meta_file, 'r') as f: |
| meta_info = json.load(f) |
|
|
|
|
| task_file = "/etc/ssd1/jiangzhongtao/baai_embedding_tune/data/all_collect/task_files.txt" |
| unsample_to_be_divisible_by = None |
| with open(task_file, 'r') as f: |
| for line in f: |
| file, _ = line.strip().split('\t') |
| try: |
| assert os.path.exists(file) |
| except Exception: |
| print(file) |
|
|
| tasktype2batchsize = { |
| 'classification': 2048, |
| 'clustering': 2048, |
| 'duplication': 512, |
| 'nli': 512, |
| 'retrieval': 512, |
| 'sts': 256, |
| 'super-ni': 2048, |
| 'unk': 2048 |
| } |
|
|
|
|
| cnt = 0 |
| type2cnt = dict() |
| steps = 0 |
| with open(task_file, 'r') as f: |
| for line in tqdm(f): |
| file, train_size = line.strip().split('\t') |
| _, task_name, _ = os.path.split(file)[-1].split('_') |
| task_type = meta_info[task_name]['task_type'] |
| if task_type != 'retrieval': |
| total_size = 0 |
| with open(file, 'r') as f_file: |
| for line in f_file: |
| total_size += 1 |
| else: |
| queries = set() |
| with open(file, 'r') as f_file: |
| for line in f_file: |
| queries.add(json.loads(line)['query']) |
| total_size = len(queries) |
| train_size = int(train_size) |
|
|
| if train_size > 0: |
| if total_size < train_size: |
| raise ValueError(file) |
| else: |
| train_size = total_size |
| batch_size = tasktype2batchsize[task_type] |
| train_size = train_size // batch_size * batch_size |
| steps += train_size // batch_size |
| |
| cnt += train_size |
| type2cnt[task_type] = type2cnt.get(task_type, 0) + train_size |
|
|
| print('steps', steps) |
| print(type2cnt) |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| print(cnt) |