| |
|
|
| |
| import pandas as pd |
| import numpy as np |
| import urllib.request |
| import rdkit |
| from rdkit import Chem |
| import os |
| import molvs |
| import csv |
| import json |
| import tqdm |
|
|
| standardizer = molvs.Standardizer() |
| fragment_remover = molvs.fragment.FragmentRemover() |
|
|
|
|
| |
| |
| |
|
|
|
|
| |
| df = pd.read_csv('all_molecular_data.csv') |
|
|
| missing_SMILES = df[df.iloc[:, 0].isna()] |
|
|
| print(f'There are {len(missing_SMILES)} rows with missing SMILES.') |
|
|
|
|
| |
| |
| quarter_df_1 = df.iloc[:len(df)//4] |
|
|
| quarter_df_1['X'] = [ \ |
| rdkit.Chem.MolToSmiles( |
| fragment_remover.remove( |
| standardizer.standardize( |
| rdkit.Chem.MolFromSmiles( |
| smiles)))) |
| for smiles in quarter_df_1['smiles']] |
|
|
| problems = [] |
| for index, row in tqdm.tqdm(quarter_df_1.iterrows()): |
| result = molvs.validate_smiles(row['X']) |
| if len(result) == 0: |
| continue |
| problems.append((row['X'], result)) |
|
|
| |
| for result, alert in problems: |
| print(f"SMILES: {result}, problem: {alert[0]}") |
|
|
| quarter_df_1.to_csv('MolData_sanitized_0.25.csv') |
|
|
|
|
|
|
| quarter_df_2 = df.iloc[len(df)//4 : len(df)//2] |
|
|
| quarter_df_2['X'] = [ \ |
| rdkit.Chem.MolToSmiles( |
| fragment_remover.remove( |
| standardizer.standardize( |
| rdkit.Chem.MolFromSmiles( |
| smiles)))) |
| for smiles in quarter_df_2['smiles']] |
|
|
| problems = [] |
| for index, row in tqdm.tqdm(quarter_df_2.iterrows()): |
| result = molvs.validate_smiles(row['X']) |
| if len(result) == 0: |
| continue |
| problems.append((row['X'], result)) |
|
|
| |
| for result, alert in problems: |
| print(f"SMILES: {result}, problem: {alert[0]}") |
|
|
| quarter_df_2.to_csv('MolData_sanitized_0.5.csv') |
|
|
|
|
| quarter_df_3 = df.iloc[len(df)//2 : 3 *len(df)//4] |
|
|
| quarter_df_3['X'] = [ \ |
| rdkit.Chem.MolToSmiles( |
| fragment_remover.remove( |
| standardizer.standardize( |
| rdkit.Chem.MolFromSmiles( |
| smiles)))) |
| for smiles in quarter_df_3['smiles']] |
|
|
| problems = [] |
| for index, row in tqdm.tqdm(quarter_df_3.iterrows()): |
| result = molvs.validate_smiles(row['X']) |
| if len(result) == 0: |
| continue |
| problems.append((row['X'], result)) |
|
|
| |
| for result, alert in problems: |
| print(f"SMILES: {result}, problem: {alert[0]}") |
|
|
| quarter_df_3.to_csv('MolData_sanitized_0.75.csv') |
|
|
|
|
|
|
| quarter_df_4 = df.iloc[3 *len(df)//4 :len(df)] |
|
|
| quarter_df_4['X'] = [ \ |
| rdkit.Chem.MolToSmiles( |
| fragment_remover.remove( |
| standardizer.standardize( |
| rdkit.Chem.MolFromSmiles( |
| smiles)))) |
| for smiles in quarter_df_4['smiles']] |
|
|
| problems = [] |
| for index, row in tqdm.tqdm(quarter_df_4.iterrows()): |
| result = molvs.validate_smiles(row['X']) |
| if len(result) == 0: |
| continue |
| problems.append((row['X'], result)) |
|
|
| |
| for result, alert in problems: |
| print(f"SMILES: {result}, problem: {alert[0]}") |
|
|
| quarter_df_4.to_csv('MolData_sanitized_1.0.csv') |
|
|
|
|
| |
| sanitized1 = pd.read_csv('MolData_sanitized_0.25.csv') |
| sanitized2 = pd.read_csv('MolData_sanitized_0.5.csv') |
| sanitized3 = pd.read_csv('MolData_sanitized_0.75.csv') |
| sanitized4 = pd.read_csv('MolData_sanitized_1.0.csv') |
|
|
| smiles_concatenated = pd.concat([sanitized1, sanitized2, sanitized3, sanitized4], ignore_index=True) |
|
|
| smiles_concatenated.to_csv('MolData_sanitized_concatenated.csv', index = False) |
|
|
|
|
|
|
|
|
| |
| |
| chunk_size = 10**5 |
| input_file = 'MolData_sanitized_concatenated.csv' |
| output_prefix = 'MolData_long_form_' |
|
|
| column_names = pd.read_csv(input_file, nrows=1).columns |
| column_names = column_names.tolist() |
|
|
| column_names = ['SMILES' if col == 'X' else col for col in column_names] |
|
|
| var_name_list = [col for col in column_names if col.startswith('activity_')] |
|
|
| with pd.read_csv(input_file, chunksize=chunk_size) as reader: |
| for i, chunk in enumerate(reader): |
| chunk.columns = column_names |
|
|
| long_df = pd.melt(chunk, id_vars=['SMILES', 'PUBCHEM_CID', 'split'], |
| value_vars=var_name_list, var_name='AID', value_name='score') |
| |
| long_df = long_df.dropna(subset=['score']) |
| long_df['score'] = long_df['score'].astype('Int64') |
|
|
| output_file = f"{output_prefix}{i+1}.csv" |
| long_df.to_csv(output_file, index=False) |
|
|
| print(f"Saved: {output_file}") |
|
|
|
|
|
|
| |
| chunk_size = 10**5 |
| input_files = [f'MolData_long_form_{i+1}.csv' for i in range(15)] |
|
|
| output_train_file = 'MolData_train.csv' |
| output_test_file = 'MolData_test.csv' |
| output_valid_file = 'MolData_validation.csv' |
|
|
| train_data = [] |
| test_data = [] |
| valid_data = [] |
|
|
| for input_file in input_files: |
| with pd.read_csv(input_file, chunksize=chunk_size) as reader: |
| for chunk in reader: |
| train_chunk = chunk[chunk['split'] == 'train'] |
| test_chunk = chunk[chunk['split'] == 'test'] |
| valid_chunk = chunk[chunk['split'] == 'validation'] |
|
|
| train_data.append(train_chunk) |
| test_data.append(test_chunk) |
| valid_data.append(valid_chunk) |
|
|
| train_df = pd.concat(train_data, ignore_index=True) |
| test_df = pd.concat(test_data, ignore_index=True) |
| valid_df = pd.concat(valid_data, ignore_index=True) |
|
|
| train_df.to_csv(output_train_file, index=False) |
| test_df.to_csv(output_test_file, index=False) |
| valid_df.to_csv(output_valid_file, index=False) |
|
|
|
|
| def fix_cid_column(df): |
| df['PUBCHEM_CID'] = df['PUBCHEM_CID'].astype(str).apply(lambda x: x.split(',')[0]) |
| df['PUBCHEM_CID'] = df['PUBCHEM_CID'].astype('Int64') |
| df = df.rename(columns = {'score' : 'Y'}) |
| return df |
|
|
| train_csv = fix_cid_column(pd.read_csv('MolData_train.csv')) |
| test_csv = fix_cid_column(pd.read_csv('MolData_test.csv')) |
| valid_csv = fix_cid_column(pd.read_csv('MolData_validation.csv')) |
|
|
| train_csv.to_parquet('MolData_train.parquet', index=False) |
| test_csv.to_parquet('MolData_test.parquet', index=False) |
| valid_csv.to_parquet('MolData_validation.parquet', index=False) |
|
|
|
|