-
Notifications
You must be signed in to change notification settings - Fork 2
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #100 from jyaacoub/development
Development
- Loading branch information
Showing
13 changed files
with
360 additions
and
132 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,29 +1,148 @@ | ||
#%% | ||
from src.data_prep.downloaders import Downloader | ||
#%% 1.Gather data for davis,kiba and pdbbind datasets | ||
import os | ||
import pandas as pd | ||
import matplotlib.pyplot as plt | ||
from src.analysis.utils import combine_dataset_pids | ||
from src import config as cfg | ||
df_prots = combine_dataset_pids(dbs=[cfg.DATA_OPT.davis, cfg.DATA_OPT.PDBbind], # just davis and pdbbind for now | ||
subset='test') | ||
|
||
Downloader.download_SDFs(['CHEMBL245769', 'CHEMBL55788'], save_dir='./') | ||
|
||
#%% | ||
import pandas as pd | ||
#%% 2. Load TCGA data | ||
df_tcga = pd.read_csv('../downloads/TCGA_ALL.maf', sep='\t') | ||
|
||
df = pd.read_csv("../data/TCGA_BRCA_Mutations.csv") | ||
df = df.loc[df['SWISSPROT'].dropna().index] | ||
df[['Gene', 'SWISSPROT']].head() | ||
#%% 3. Pre filtering | ||
df_tcga = df_tcga[df_tcga['Variant_Classification'] == 'Missense_Mutation'] | ||
df_tcga['seq_len'] = pd.to_numeric(df_tcga['Protein_position'].str.split('/').str[1]) | ||
df_tcga = df_tcga[df_tcga['seq_len'] < 5000] | ||
df_tcga['seq_len'].plot.hist(bins=100, title="sequence length histogram capped at 5K") | ||
plt.show() | ||
df_tcga = df_tcga[df_tcga['seq_len'] < 1200] | ||
df_tcga['seq_len'].plot.hist(bins=100, title="sequence length after capped at 1.2K") | ||
|
||
# %% | ||
from src.utils.pdb import pdb2uniprot | ||
df2 = pd.read_csv("../data/PlatinumDataset/nomsa_binary_original_binary/full/cleaned_XY.csv", index_col=0) | ||
df2['pdb_id'] = df2.prot_id.str.split("_").str[0] | ||
#%% 4. Merging df_prots with TCGA | ||
df_tcga['uniprot'] = df_tcga['SWISSPROT'].str.split('.').str[0] | ||
|
||
dfm = df_tcga.merge(df_prots[df_prots.db != 'davis'], | ||
left_on='uniprot', right_on='prot_id', how='inner') | ||
|
||
# for davis we have to merge on HUGO_SYMBOLS | ||
dfm_davis = df_tcga.merge(df_prots[df_prots.db == 'davis'], | ||
left_on='Hugo_Symbol', right_on='prot_id', how='inner') | ||
|
||
dfm = pd.concat([dfm,dfm_davis], axis=0) | ||
|
||
del dfm_davis # to save mem | ||
|
||
# %% 5. Post filtering step | ||
# 5.1. Filter for only those sequences with matching sequence length (to get rid of nonmatched isoforms) | ||
# seq_len_x is from tcga, seq_len_y is from our dataset | ||
tmp = len(dfm) | ||
# allow for some error due to missing amino acids from pdb file in PDBbind dataset | ||
# - assumption here is that isoforms will differ by more than 50 amino acids | ||
dfm = dfm[(dfm.seq_len_y <= dfm.seq_len_x) & (dfm.seq_len_x<= dfm.seq_len_y+50)] | ||
print(f"Filter #1 (seq_len) : {tmp:5d} - {tmp-len(dfm):5d} = {len(dfm):5d}") | ||
|
||
# 5.2. Filter out those that dont have the same reference seq according to the "Protein_position" and "Amino_acids" col | ||
|
||
# Extract mutation location and reference amino acid from 'Protein_position' and 'Amino_acids' columns | ||
dfm['mt_loc'] = pd.to_numeric(dfm['Protein_position'].str.split('/').str[0]) | ||
dfm = dfm[dfm['mt_loc'] < dfm['seq_len_y']] | ||
dfm[['ref_AA', 'mt_AA']] = dfm['Amino_acids'].str.split('/', expand=True) | ||
|
||
dfm['db_AA'] = dfm.apply(lambda row: row['prot_seq'][row['mt_loc']-1], axis=1) | ||
|
||
# Filter #2: Match proteins with the same reference amino acid at the mutation location | ||
tmp = len(dfm) | ||
dfm = dfm[dfm['db_AA'] == dfm['ref_AA']] | ||
print(f"Filter #2 (ref_AA match): {tmp:5d} - {tmp-len(dfm):5d} = {len(dfm):5d}") | ||
print('\n',dfm.db.value_counts()) | ||
|
||
|
||
# %% final seq len distribution | ||
n_bins = 25 | ||
lengths = dfm.seq_len_x | ||
fig, ax = plt.subplots(1, 1, figsize=(10, 5)) | ||
|
||
# Plot histogram | ||
n, bins, patches = ax.hist(lengths, bins=n_bins, color='blue', alpha=0.7) | ||
ax.set_title('TCGA final filtering for db matches') | ||
|
||
# Add counts to each bin | ||
for count, x, patch in zip(n, bins, patches): | ||
ax.text(x + 0.5, count, str(int(count)), ha='center', va='bottom') | ||
|
||
uniprots = pdb2uniprot(df2.pdb_id.unique()) | ||
df_pid = pd.DataFrame(list(uniprots.items()), columns=['pdbID', 'uniprot']) | ||
ax.set_xlabel('Sequence Length') | ||
ax.set_ylabel('Frequency') | ||
|
||
plt.tight_layout() | ||
plt.show() | ||
|
||
# %% Getting updated sequences | ||
def apply_mut(row): | ||
ref_seq = list(row['prot_seq']) | ||
ref_seq[row['mt_loc']-1] = row['mt_AA'] | ||
return ''.join(ref_seq) | ||
|
||
dfm['mt_seq'] = dfm.apply(apply_mut, axis=1) | ||
|
||
|
||
# %% | ||
dfm.to_csv("/cluster/home/t122995uhn/projects/data/tcga/tcga_maf_davis_pdbbind.csv") | ||
# %% | ||
# find overlap between uniprots and swissprot ids | ||
uniprots = set(uniprots) # convert to set for faster lookup | ||
df['uniprot'] = df['SWISSPROT'].str.split('.').str[0] | ||
# Merge the original df with uni_to_pdb_df | ||
df = df.merge(df_pid, on='uniprot', how='left') | ||
from src.utils.seq_alignment import MSARunner | ||
from tqdm import tqdm | ||
import pandas as pd | ||
import os | ||
|
||
DATA_DIR = '/cluster/home/t122995uhn/projects/data/tcga' | ||
CSV = f'{DATA_DIR}/tcga_maf_davis_pdbbind.csv' | ||
N_CPUS= 6 | ||
NUM_ARRAYS = 10 | ||
array_idx = 0#${SLURM_ARRAY_TASK_ID} | ||
|
||
df = pd.read_csv(CSV, index_col=0) | ||
df.sort_values(by='seq_len_y', inplace=True) | ||
|
||
|
||
# %% | ||
df[~df['pdbID'].isna()] | ||
for DB in df.db.unique(): | ||
print('DB', DB) | ||
RAW_DIR = f'{DATA_DIR}/{DB}' | ||
# should already be unique if these are proteins mapped form tcga! | ||
unique_df = df[df['db'] == DB] | ||
########################## Get job partition | ||
partition_size = len(unique_df) / NUM_ARRAYS | ||
start, end = int(array_idx*partition_size), int((array_idx+1)*partition_size) | ||
|
||
unique_df = unique_df[start:end] | ||
|
||
#################################### create fastas | ||
fa_dir = os.path.join(RAW_DIR, f'{DB}_fa') | ||
fasta_fp = lambda idx,pid: os.path.join(fa_dir, f"{idx}-{pid}.fasta") | ||
os.makedirs(fa_dir, exist_ok=True) | ||
for idx, (prot_id, pro_seq) in tqdm( | ||
unique_df[['prot_id', 'prot_seq']].iterrows(), | ||
desc='Creating fastas', | ||
total=len(unique_df)): | ||
with open(fasta_fp(idx,prot_id), "w") as f: | ||
f.write(f">{prot_id},{idx},{DB}\n{pro_seq}") | ||
|
||
##################################### Run hhblits | ||
aln_dir = os.path.join(RAW_DIR, f'{DB}_aln') | ||
aln_fp = lambda idx,pid: os.path.join(aln_dir, f"{idx}-{pid}.a3m") | ||
os.makedirs(aln_dir, exist_ok=True) | ||
|
||
# finally running | ||
for idx, (prot_id, pro_seq) in tqdm( | ||
unique_df[['prot_id', 'mt_seq']].iterrows(), | ||
desc='Running hhblits', | ||
total=len(unique_df)): | ||
in_fp = fasta_fp(idx,prot_id) | ||
out_fp = aln_fp(idx,prot_id) | ||
|
||
if not os.path.isfile(out_fp): | ||
print(MSARunner.hhblits(in_fp, out_fp, n_cpus=N_CPUS, return_cmd=True)) | ||
break | ||
|
||
# %% |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.