| |
| |
|
|
| import random |
| import sys, os, pdb |
| import json, math |
| import datasets |
| from datasets import load_dataset |
| import csv |
|
|
| random.seed(42) |
|
|
| DATASETS = { |
| "natural_language_understanding": [ |
| "ATIS", "ATIS-NER", "BANKING77", "BANKING77-OOS", "CLINC-Single-Domain-OOS-banking", |
| "CLINC-Single-Domain-OOS-credit_cards", "CLINC150", "DSTC8-SGD", "HWU64", "MIT-Movie", |
| "MIT-Restaurant", "RESTAURANTS8K", "SNIPS", "SNIPS-NER", "TOP", "TOP-NER" |
| ], |
| "task_oriented": [ |
| "ABCD", "AirDialogue", "BiTOD", "CaSiNo", "CraigslistBargains", |
| "Disambiguation", "DSTC2-Clean", "FRAMES", "GECOR", "HDSA-Dialog", |
| "KETOD", "KVRET", "MetaLWOZ", "MS-DC", "MuDoCo", |
| "MulDoGO", "MultiWOZ_2.1", "MULTIWOZ2_2", "SGD", "SimJointGEN", |
| "SimJointMovie", "SimJointRestaurant", "STAR", "Taskmaster1", "Taskmaster2", |
| "Taskmaster3", "WOZ2_0" |
| ], |
| "dialogue_summarization": [ |
| "AMI", "CRD3", "DialogSum", "ECTSum", "ICSI", |
| "MediaSum", "QMSum", "SAMSum", "TweetSumm", "ConvoSumm", |
| "SummScreen_ForeverDreaming", "SummScreen_TVMegaSite" |
| ], |
| "conversational_recommendation": [ |
| "Redial", "DuRecDial-2.0", "OpenDialKG", "SalesBot", |
| ], |
| "open_domain": [ |
| "chitchat-dataset", "ConvAI2", "AntiScam", "Empathetic", "HH-RLHF", |
| "PLACES3.5", "Prosocial", "SODA" |
| ], |
| "knowledge_grounded": [ |
| "CompWebQ", "CoQA", "CoSQL", "DART", "FeTaQA", |
| "GrailQA", "HybridQA", "MTOP", "MultiModalQA", "SParC", |
| "Spider", "SQA", "ToTTo", "WebQSP", "WikiSQL", |
| "WikiTQ", "wizard_of_internet", "wizard_of_wikipedia" |
| ], |
| } |
|
|
| class Test(object): |
| def __init__(self): |
| pass |
|
|
| def test_single_dataset(self, data_name): |
| |
| dataset = load_dataset("Salesforce/dialogstudio", data_name, revision="download") |
| dataset_size = { |
| "train":0, |
| "validation": 0, |
| "test": 0, |
| } |
| for split in dataset: |
| dataset_size[split] = len(dataset[split]) |
| print(dataset_size) |
| |
| return dataset_size |
|
|
| def test_all(self): |
| with open("dataset_stats.csv", "w", newline="") as tf: |
| writer = csv.writer(tf) |
| writer.writerow(["Category", "Data_name", "train", "val", "test"]) |
| for cat, dataset_list in DATASETS.items(): |
| for data_name in dataset_list: |
| dataset_size = self.test_single_dataset(data_name=data_name) |
| writer.writerow([cat, data_name] + list(dataset_size.values())) |
|
|
|
|
| def main(): |
| test = Test() |
| |
| |
| |
| |
| test.test_single_dataset("Taskmaster2") |
|
|
| if __name__ == "__main__": |
| main() |
|
|