Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
B
BiTransDPI
概览
概览
详情
活动
周期分析
版本库
存储库
文件
提交
分支
标签
贡献者
分支图
比较
统计图
问题
0
议题
0
列表
看板
标记
里程碑
CI / CD
CI / CD
流水线
日程表
维基
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
提交
问题看板
Open sidebar
杨志辉
BiTransDPI
Commits
da2551fa
提交
da2551fa
authored
5月 27, 2021
作者:
朱学凯
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
change data
上级
9d55deb2
隐藏空白字符变更
内嵌
并排
正在显示
21 个修改的文件
包含
56 行增加
和
53 行删除
+56
-53
ER_ic50
data/ER/ER_ic50
+0
-0
ER_smile
data/ER/ER_smile
+0
-0
ER_sps
data/ER/ER_sps
+0
-0
GPCR_ic50
data/GPCR/GPCR_ic50
+0
-0
GPCR_smile
data/GPCR/GPCR_smile
+0
-0
GPCR_sps
data/GPCR/GPCR_sps
+0
-0
channel_ic50
data/Ion_channel/channel_ic50
+0
-0
channel_smile
data/Ion_channel/channel_smile
+0
-0
channel_sps
data/Ion_channel/channel_sps
+0
-0
kinase_ic50
data/Tyrosine_kinase/kinase_ic50
+0
-0
kinase_smile
data/Tyrosine_kinase/kinase_smile
+0
-0
kinase_sps
data/Tyrosine_kinase/kinase_sps
+0
-0
test_ic50
data/test/test_ic50
+0
-0
test_smile
data/test/test_smile
+0
-0
test_sps
data/test/test_sps
+0
-0
train_ic50
data/train/train_ic50
+0
-0
train_smile
data/train/train_smile
+0
-0
train_sps
data/train/train_sps
+0
-0
dataset.py
dataset.py
+52
-0
run_interaction.py
run_interaction.py
+3
-52
test.py
test.py
+1
-1
没有找到文件。
data/ER_ic50
→
data/ER
/ER
_ic50
浏览文件 @
da2551fa
File moved
data/ER_smile
→
data/ER
/ER
_smile
浏览文件 @
da2551fa
File moved
data/ER_sps
→
data/ER
/ER
_sps
浏览文件 @
da2551fa
File moved
data/GPCR_ic50
→
data/GPCR
/GPCR
_ic50
浏览文件 @
da2551fa
File moved
data/GPCR_smile
→
data/GPCR
/GPCR
_smile
浏览文件 @
da2551fa
File moved
data/GPCR_sps
→
data/GPCR
/GPCR
_sps
浏览文件 @
da2551fa
File moved
data/channel_ic50
→
data/
Ion_channel/
channel_ic50
浏览文件 @
da2551fa
File moved
data/channel_smile
→
data/
Ion_channel/
channel_smile
浏览文件 @
da2551fa
File moved
data/channel_sps
→
data/
Ion_channel/
channel_sps
浏览文件 @
da2551fa
File moved
data/kinase_ic50
→
data/
Tyrosine_kinase/
kinase_ic50
浏览文件 @
da2551fa
File moved
data/kinase_smile
→
data/
Tyrosine_kinase/
kinase_smile
浏览文件 @
da2551fa
File moved
data/kinase_sps
→
data/
Tyrosine_kinase/
kinase_sps
浏览文件 @
da2551fa
File moved
data/test_ic50
→
data/test
/test
_ic50
浏览文件 @
da2551fa
File moved
data/test_smile
→
data/test
/test
_smile
浏览文件 @
da2551fa
File moved
data/test_sps
→
data/test
/test
_sps
浏览文件 @
da2551fa
File moved
data/train_ic50
→
data/train
/train
_ic50
浏览文件 @
da2551fa
File moved
data/train_smile
→
data/train
/train
_smile
浏览文件 @
da2551fa
File moved
data/train_sps
→
data/train
/train
_sps
浏览文件 @
da2551fa
File moved
dataset.py
浏览文件 @
da2551fa
...
@@ -146,6 +146,58 @@ class Data_Encoder(data.Dataset):
...
@@ -146,6 +146,58 @@ class Data_Encoder(data.Dataset):
# return len(d), len(p)
# return len(d), len(p)
def
get_task
(
task_name
):
tokenizer_config
=
{
"vocab_file"
:
'./config/vocab.txt'
,
"vocab_pair"
:
'./config/drug_codes_chembl.txt'
,
"begin_id"
:
'[CLS]'
,
"separate_id"
:
"[SEP]"
,
"max_len"
:
256
}
if
task_name
.
lower
()
==
'train'
:
df_train
=
{
"sps"
:
'./data/train/train_sps'
,
"smile"
:
'./data/train/train_smile'
,
"affinity"
:
'./data/train/train_ic50'
,
}
return
df_train
,
tokenizer_config
elif
task_name
.
lower
()
==
'test'
:
df_test
=
{
"sps"
:
'./data/test/test_sps'
,
"smile"
:
'./data/test/test_smile'
,
"affinity"
:
'./data/test/test_ic50'
,
}
return
df_test
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_1'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_1_ic50'
,
}
return
df
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_10'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_10_ic50'
,
}
return
df
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_100'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_100_ic50'
,
}
return
df
,
tokenizer_config
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
# local test
# local test
# dataFolder = './IC50/SPS/train_smile'
# dataFolder = './IC50/SPS/train_smile'
...
...
run_interaction.py
浏览文件 @
da2551fa
from
argparse
import
ArgumentParser
from
argparse
import
ArgumentParser
from
dataset
import
Data_Encoder
from
dataset
import
Data_Encoder
,
get_task
import
torch
import
torch
from
torch.utils.data
import
DataLoader
from
torch.utils.data
import
DataLoader
from
configuration_bert
import
BertConfig
from
configuration_bert
import
BertConfig
...
@@ -9,55 +9,6 @@ import os
...
@@ -9,55 +9,6 @@ import os
from
tqdm
import
tqdm
from
tqdm
import
tqdm
torch
.
set_default_tensor_type
(
torch
.
DoubleTensor
)
torch
.
set_default_tensor_type
(
torch
.
DoubleTensor
)
def
get_task
(
task_name
):
tokenizer_config
=
{
"vocab_file"
:
'./config/vocab.txt'
,
"vocab_pair"
:
'./config/drug_codes_chembl.txt'
,
"begin_id"
:
'[CLS]'
,
"separate_id"
:
"[SEP]"
,
"max_len"
:
256
}
if
task_name
.
lower
()
==
'train'
:
df_train
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_ic50'
,
}
return
df_train
,
tokenizer_config
elif
task_name
.
lower
()
==
'test'
:
df_test
=
{
"sps"
:
'./data/test_sps'
,
"smile"
:
'./data/test_smile'
,
"affinity"
:
'./data/test_ic50'
,
}
return
df_test
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_1'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_1_ic50'
,
}
return
df
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_10'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_10_ic50'
,
}
return
df
,
tokenizer_config
elif
task_name
.
lower
()
==
'train_z_100'
:
df
=
{
"sps"
:
'./data/train_sps'
,
"smile"
:
'./data/train_smile'
,
"affinity"
:
'./data/train_z_100_ic50'
,
}
return
df
,
tokenizer_config
...
@@ -144,8 +95,8 @@ def test(args, model, dataset):
...
@@ -144,8 +95,8 @@ def test(args, model, dataset):
for
res
in
pred_affinity
:
for
res
in
pred_affinity
:
f
.
write
(
str
(
res
)
+
'
\n
'
)
f
.
write
(
str
(
res
)
+
'
\n
'
)
if
args
.
do_eval
:
#
if args.do_eval:
os
.
system
(
'python eval.py'
)
#
os.system('python eval.py')
def
main
(
args
):
def
main
(
args
):
...
...
test.py
浏览文件 @
da2551fa
...
@@ -5,7 +5,7 @@ import torch
...
@@ -5,7 +5,7 @@ import torch
tokenizer
=
BertTokenizer
.
from_pretrained
(
'bert-base-uncased'
)
tokenizer
=
BertTokenizer
.
from_pretrained
(
'bert-base-uncased'
)
model
=
BertForMaskedLM
.
from_pretrained
(
'bert-base-uncased'
)
model
=
BertForMaskedLM
.
from_pretrained
(
'bert-base-uncased'
)
inputs
=
tokenizer
(
"The capital of France is [MASK]."
,
return_tensors
=
"pt"
)
inputs
=
tokenizer
(
"The capital of France is [MASK]."
,
return_tensors
=
"pt"
)
labels
=
tokenizer
(
"The capital of France is Paris."
,
return_tensors
=
"pt"
)[
"input_ids"
]
labels
=
tokenizer
(
"The capital of France is Paris."
,
return_tensors
=
"pt"
)
#
["input_ids"]
outputs
=
model
(
**
inputs
,
labels
=
labels
)
outputs
=
model
(
**
inputs
,
labels
=
labels
)
loss
=
outputs
.
loss
loss
=
outputs
.
loss
logits
=
outputs
.
logits
logits
=
outputs
.
logits
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论