-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathcontrastive_experiments.py
More file actions
118 lines (110 loc) · 3.32 KB
/
contrastive_experiments.py
File metadata and controls
118 lines (110 loc) · 3.32 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
from models import *
from contrastive import *
ARTIFACT_DIR = {
'dpi': f'./artifact/exp1',
'scope': f'./artifact/exp2',
}
for dir in ARTIFACT_DIR.values():
os.makedirs(dir, exist_ok=True)
def run_contrastive(**kwargs):
set_seed(kwargs["seed"])
save_dir = f'{ARTIFACT_DIR[kwargs["dataset"]]}/{kwargs["dataset"]}_contrastive_{kwargs["emb_model"]}_{kwargs["pooling_mode"]}'
if kwargs['pooling_mode'] == 'bom':
save_dir += f'_k{kwargs["k"]}_s{kwargs["stride"]}'
save_dir += f'_seed{kwargs["seed"]}.pt'
dist_net_config = {
'qnet': MLP([PLM_dim[kwargs['emb_model']], 256, 1024]),
'knet': MLP([PLM_dim[kwargs['emb_model']], 256, 1024]),
'vnet': MLP([PLM_dim[kwargs['emb_model']], 256, 1024])
}
dist_net = CrossAttentionKernel(**dist_net_config) if kwargs['pooling_mode'] == 'bom' else MultiheadLinearKernel(**dist_net_config)
if kwargs['dataset'] == 'dpi':
data = DPIDataset(
emb_model=kwargs['emb_model'],
save='load_pool',
pooling_mode=kwargs['pooling_mode'],
k=kwargs.get('k', None),
stride=kwargs.get('stride', None)
)
model = DPI(
data=data,
contrastive_loss=TripletLoss(dist_net, margin=kwargs.get('margin', 1.)),
)
if kwargs['dataset'] == 'scope':
data = SCOPeDataset(
emb_model=kwargs['emb_model'],
save='load_pool',
cls='all',
pooling_mode=kwargs['pooling_mode'],
k=kwargs.get('k', None),
stride=kwargs.get('stride', None)
)
model = SCOPe(
data=data,
contrastive_loss=TripletLoss(dist_net, margin=kwargs.get('margin', 1.)),
)
model.reset_history()
model.train(
n_epochs=kwargs.get('n_epochs', 401),
interval=kwargs.get('interval', 20),
batch_size=kwargs.get('batch_size', 128),
lr=kwargs.get('lr', 4e-4)
)
if kwargs['save']:
torch.save(model.history, save_dir)
return model
def dpi_experiments(dev):
torch.cuda.set_device(dev)
run_experiments(
common_kwargs = {
'dataset': 'dpi',
'save': True,
'k': 40,
'stride': 8,
'n_epochs': 2001,
'interval': 100,
'margin': 0.6
},
lr={'contrastive': 1e-4},
)
def scope_experiments(dev):
torch.cuda.set_device(dev)
run_experiments(
common_kwargs = {
'dataset': 'scope',
'save': True,
'k': 100,
'stride': 80,
'n_epochs': 201,
'interval': 10,
},
lr={'contrastive': 5e-4},
)
def run_experiments(common_kwargs, lr):
pooling_method = [
'bom',
# 'cls',
# 'avg',
# 'sep'
]
emb_models = [
# 'prottrans',
# 'protbert',
'esm2-35M',
# 'esm2-150M',
# 'esm2-650M'
]
seeds = [
261,
# 2602,
# 26003,
# 2604,
# 265
]
for seed in seeds:
common_kwargs['seed'] = seed
for pm in pooling_method:
for em in emb_models:
run_contrastive(**common_kwargs, pooling_mode=pm, emb_model=em, lr=lr['contrastive'])
if __name__ == '__main__':
scope_experiments(dev=0)