File size: 4,369 Bytes
2e12b82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
96ec482
2e12b82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
96ec482
 
 
2e12b82
 
96ec482
2e12b82
 
 
 
96ec482
2e12b82
 
 
 
96ec482
2e12b82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107

from datasets import Value, ClassLabel,Sequence
import datasets


_AMZ20_CITATION = """\

"""

_AMZ20_DESCRIPTION = """\
GLUE, the General Language Understanding Evaluation benchmark
(https://gluebenchmark.com/) is a collection of resources for training,
evaluating, and analyzing natural language understanding systems.
"""

class AMZ20Config(datasets.BuilderConfig):

    def __init__(
        self,
        text_features,
        label_column,
        data_url,
        data_dir,
        citation,
        url,
        label_classes=None,
        process_label=lambda x: x,
        **kwargs,
    ):
        super(AMZ20Config, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
        
        self.text_features = text_features
        self.label_column = label_column
        self.label_classes = label_classes
        self.data_url = data_url
        self.data_dir = data_dir
        self.citation = citation
        self.url = url
        self.process_label = process_label

class AMZ20(datasets.GeneratorBasedBuilder):
    domain_list = ['Sandal', 'Magazine_Subscriptions', 'RiceCooker', 'Flashlight', 'Jewelry', 'CableModem', 'GraphicsCard',
            'GPS', 'Projector', 'Keyboard', 'Video_Games', 'AlarmClock', 'HomeTheaterSystem', 'Vacuum', 'Gloves',
            'Baby', 'Bag', 'Movies_TV', 'Dumbbell', 'Headphone']

    BUILDER_CONFIGS  = [
        AMZ20Config(name=domain_name, 
                    description= f'comments of JD {domain_name}.', 
                    text_features={'sentence':'sentence', 'domain':'domain'}, 
                    label_classes=['POS','NEG', 'NEU'], 
                    label_column='label',
                    citation="",
                    data_dir= "",
                    data_url = "https://huggingface.co/datasets/kuroneko3578/amz20/resolve/main/",
                    url='https://github.com/ws719547997/LNB-DA')
        for domain_name in domain_list
    ]

    def _info(self):
        features = {'id':Value(dtype='int32', id=None),
            'domain':Value(dtype='string', id=None),
            'label':ClassLabel(num_classes=3, names=['POS', 'NEG', 'NEU'], names_file=None, id=None),
            'rank':Value(dtype='int32', id=None),
            'sentence':Value(dtype='string', id=None)}

        return datasets.DatasetInfo(
            description=_AMZ20_DESCRIPTION,
            features=datasets.Features(features),
            homepage=self.config.url,
            citation=self.config.citation + "\n" + _AMZ20_CITATION,
        )

    def _split_generators(self, dl_manager):

        test_file = rf'{self.config.data_url}test/{self.config.name}.txt'
        dev_file = rf'{self.config.data_url}dev/{self.config.name}.txt'
        train_file = rf'{self.config.data_url}train/{self.config.name}.txt'
        return [datasets.SplitGenerator(name=datasets.Split.TEST,
                                        gen_kwargs={
                                            "data_file": dl_manager.download(test_file),
                                            "split": "test",
                                        },), 
                datasets.SplitGenerator(name=datasets.Split.VALIDATION,
                                        gen_kwargs={
                                            "data_file": dl_manager.download(dev_file),
                                            "split": "dev",
                                        },), 
                datasets.SplitGenerator(name=datasets.Split.TRAIN,
                                        gen_kwargs={
                                            "data_file": dl_manager.download(train_file),
                                            "split": "train",
                                        },)]

    def _generate_examples(self, data_file, split):
        with open(data_file, 'r', encoding='utf-8') as f:
            for line in f:
                lin = line.strip()
                if not lin:
                    continue
                lin_sp = lin.split('\t')
                if len(lin_sp) < 5:
                    continue
                # 列表头也被随机打散到数据集里面了
                if lin_sp[0] == 'index':
                    continue
                # id, {example}
                yield lin_sp[0], {'sentence':lin_sp[4],'domain':lin_sp[1], 'label':lin_sp[2], 'id':lin_sp[0], 'rank':lin_sp[3]}