-
Notifications
You must be signed in to change notification settings - Fork 0
/
setting.json
113 lines (93 loc) · 2.21 KB
/
setting.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
{
"path": {
"data": "data",
"code": "code",
"log": "log",
"model": "model",
"state_dict": "state_dict",
"submit": "submit",
"train": "train",
"valid": "valid",
"ensemble": "ensemble",
"best_model": "best_model"
},
"file_name": {
"train": "2023-05-22_train_data",
"test": "2023-05-22_test_data"
},
"save_name": "",
"num_kfolds": 3,
"seed": 42,
"device": "cuda",
"fixed_train_columns": ["user_id", "question_id", "answer_code"],
"fixed_embedding_columns": ["question_id"],
"selective_train_columns": ["test_id", "knowledge_tag"],
"selective_embedding_columns": ["test_id", "knowledge_tag"],
"predict_column": "answer_code",
"train_valid_split": 0.9,
"max_train_length": 100,
"extra_split": false,
"model_name": "lstm_attn",
"loss_fn": "BCEWLL",
"optimizer": "ADAM",
"scheduler": "plateau",
"epoch": 10,
"batch_size": 16,
"num_workers": 1,
"wandb_activate" : true,
"patience" : 10,
"best_model_activate": true,
"adam": {
"learn_rate": 0.01,
"weight_decay": 0
},
"sgd": {
"learn_rate": 0.0001,
"weight_decay": 0.01
},
"rmsprop": {
"learn_rate": 0.0001,
"weight_decay": 0.01
},
"adadelta": {
"learn_rate": 0.0001,
"weight_decay": 0.01
},
"plateau": {
"patience": 10,
"factor": 0.5,
"mode": "max",
"verbose": true
},
"lstm": {
"embedding_dim": 21,
"input_dim": 64,
"output_dim": 64,
"max_seq_len": 20,
"n_layers": 2,
"dense_layer_dim": [16, 4]
},
"lstm_attn": {
"embedding_dim": 21,
"input_dim": 64,
"output_dim": 64,
"max_seq_len": 20,
"n_layers": 2,
"n_heads": 2,
"drop_out": 0.2,
"dense_layer_dim": [16, 4]
},
"bert": {
"embedding_dim": 21,
"input_dim": 64,
"max_seq_len": 20,
"n_layers": 2,
"n_heads": 2,
"dense_layer_dim": [16, 4]
},
"lgcn": {
"embedding_dim": 64,
"num_layers": 1,
"alpha": 0
}
}