diff options
author | Hui Lan <lanhui@zjnu.edu.cn> | 2024-08-06 15:59:45 +0800 |
---|---|---|
committer | Hui Lan <lanhui@zjnu.edu.cn> | 2024-08-06 15:59:45 +0800 |
commit | 2f1b7bc8d9542e3d65007b6658691ae9e5c5e870 (patch) | |
tree | 2e094a688d494dd8bfc1db71d338f13a3a4cb04d | |
parent | 1ba96fd1fc6fc27d24ed1ae2776a8b85ee967fc5 (diff) |
merge_edges.py: add some comments
-rwxr-xr-x | Code/#update_network.py# | 828 | ||||
l--------- | Code/.#update_network.py | 1 | ||||
-rw-r--r-- | Code/merge_edges.py | 6 |
3 files changed, 833 insertions, 2 deletions
diff --git a/Code/#update_network.py# b/Code/#update_network.py# new file mode 100755 index 0000000..4e46fdc --- /dev/null +++ b/Code/#update_network.py# @@ -0,0 +1,828 @@ +#! /usr/bin/python3 +# Usage: python3 update_network.py +# Put this script under directory Code/. +# IMPORTANT: Run this script under directory Code/. +# Execute the above command regularly, or +# Cron job this command to make it run everyday at 5am: +# +# 1. crontab -e. +# 2. Add this line: 01 05 * * * cd /home/hui/network/v03/Code && python3 update_network.py +# +# IMPORTANT: Make sure that you execute this script (update_network.py) under the directory Code. +# +# Purpose: periodically (e.g., per week) run this script to see if the network needs update. If yes, update it. +# +# Set HOLDON=NO in parameter_for_buildCmatrix.txt, +# parameter_for_buildRmatrix.txt and parameter_for_net.txt to make +# changes in these file effective. +# +# parameter_for_buildRmatrix.txt will be updated automatically (I +# hope). However, we need to update parameter_for_buildCmatrix.txt +# manually. +# +# Revision history: +# +# Last modified: 26 Feb 2017 +# Last modified: 17 Mar 2017 +# Last modified: 04 Apr 2017 +# Last modified: 05 Apr 2017 +# Last modified: 10 Apr 2017 +# Last modified: 19 Apr 2017 +# Last modified: 20 Apr 2017 [addded create_edges0B.py which calls correlation_per_tissue.R] +# Last modified: 21 Jun 2017 [added correlation_per_group.R and wedge.R] +# Last modified: 30 Jun 2017 [added get_sample_size so that we have sample size for correlations of type all, added in ll_dict ] +# Last modified: 23 Jan 2018 [edited a few print-out messages] +# Last modified: 25 Jan 2018 [updated function compute_metric(), set S=365.0 and modified return statement] +# Last modified: 24 Aug 2018 [updated function from get_sample_size(d, sorted_keys, day) to get_sample_size(d, sorted_keys, day, rcond_string)] +# Last modified: 03 Feb 2019 +# Last modified: 08 Aug 2019, hui +# Last modified: 10 Aug 2019, hui <lanhui@zjnu.edu.cn> +# Last modified: 23 Aug 2019, hui <lanhui@zjnu.edu.cn> [correlation_mixtools(num_component)] +# Last modified: 10 Sep 2019, hui <lanhui@zjnu.edu.cn> [correlation_mixtools, check the previous R session has finished before starting a new one.] + +import os, sys +import numpy as np +import glob +import time +import subprocess +from datetime import datetime +from param4net import make_global_param_dict, get_key_value +from log import write_log_file +from configure import HISTORY_DIR, HISTORY_DIR2, FILE_TIMESTAMP, SAMPLE_SIZE_FILE, TEMP_DIR, \ + PARAMETER_FOR_BUILDCMATRIX, PARAMETER_FOR_BUILDRMATRIX, \ + PARAMETER_FOR_NET, PARAMETER_FOR_NET_TRAVADB_STRESS, PARAMETER_FOR_NET_TRAVADB_MAP, PARAMETER_FOR_NET_MILD_DROUGHT, PARAMETER_FOR_NET_WIGGELAB_DIURNAL, \ + BINDING_FILE, TPM_FILE, \ + BUILDRMATRIX_RENEW_INTERVAL, MIN_RNA_SEQ_INCREASE, UPDATE_NETWORK_LOG_FILE, NEW_OR_UPDATED_CHIP_FILE, \ + RNA_SEQ_INFO_DATABASE, RNA_SEQ_INFO_DATABASE_JSON, GENE_ID_FIRST_TWO_LETTERS, MEMORY_STRENGTH, \ + MAPPED_RDATA_DIR, MAPPED_CDATA_DIR, \ + EDGE_POOL_DIR, MERGED_EDGE_FILE, \ + TARGET_TF_FILE + + + +## Helper functions + +def get_value(s, delimit): + lst = s.split(delimit, 1) # only split at the first delimit + return lst[1].strip() + + +def validate_webapp_dir(para_for_net): + ''' Make sure this function is executed under the directory Code. ''' + glb_param_dict = make_global_param_dict(para_for_net) + # if genes.json is not present, create one + if not os.path.exists('../Webapp/static/json/genes.json'): + print('[update_network.py]: cannot find genes.json, make one ...') + cmd = 'python3 text2json.py %s > ../Webapp/static/json/genes.json' % (glb_param_dict['GENE_ID_AND_GENE_NAME']) + os.system(cmd) + + +def make_paths(s): + if not os.path.isdir(s): + os.makedirs(s) + + +def make_important_dirs(): + make_paths('../Data/history/edges/many_targets') + make_paths('../Data/history/edges/one_target') + make_paths('../Data/log') + make_paths('../Data/information') + make_paths('../Data/temp') + make_paths('../Data/upload') + make_paths('../Data/parameter') + make_paths('../Data/R/Mapped') + make_paths('../Data/R/Mapped/public') + make_paths('../Data/R/Mapped/inhouse') + make_paths('../Data/R/Mapped/other') + make_paths('../Data/R/Raw') + make_paths('../Data/C/Mapped') + make_paths('../Data/C/Raw') + make_paths('../Data/history/edges') + make_paths(EDGE_POOL_DIR) + make_paths('../Data/history/bind') + make_paths('../Data/history/expr') + make_paths('../Webapp/static/json') + make_paths('../Webapp/static/edges') + make_paths('../Webapp/templates') + + +def num_line(fname): + ''' Return number of lines in file fname. ''' + if not os.path.exists(fname): + return 0 + f = open(fname) + lines = f.readlines() + f.close() + return len(lines) + + +def num_ids(fname): + ''' Return number of IDs in fname. ''' + f = open(fname) + lines = f.readlines() + f.close() + return len(lines[0].split('\t')) - 1 + + +def write_sample_size_file(sample_size_file, curr_date, tpm_sample_size): + if not os.path.exists(sample_size_file): + f = open(sample_size_file, 'w') + else: + f = open(sample_size_file, 'a') + f.write('%s\t%s\n' % (curr_date, tpm_sample_size)) + f.close() + + +def age_of_file_in_days(fname): + ''' Return age of fname in days. ''' + st = os.stat(fname) + days = (time.time() - st.st_mtime)/(3600*24.0) + return days + + +def age_of_file_in_seconds(fname): + ''' Return age of fname in days. ''' + st = os.stat(fname) + seconds = time.time() - st.st_mtime + return seconds + + +def hold_on(fname): + f = open(fname) + lines = f.readlines() + f.close() + for line in lines[:100]: # check the first 100 lines for HOLDON + line = line.strip() + if line.startswith('%%HOLDON=YES'): + return True + return False + + +def all_files_present(lst): + missing_file_lst = [] + for path in lst: # lst is a list of file names to check + if not os.path.exists(path): + if 'edges.txt' in path: + write_log_file('[update_network.py] WARNING: must have %s to update network. Call create_edges*.py to create edge files.' % (path), UPDATE_NETWORK_LOG_FILE) + missing_file_lst.append(path) + return missing_file_lst + + +def record_file_time(lst, fname): + ''' + lst - a list of files + fname - a recorder file + ''' + f = open(fname, 'w') + s = '' + for x in lst: + if os.path.exists(x): + s += '%s\t%d\n' % (os.path.basename(x), int(os.stat(x).st_mtime)) + else: + s += '%s\t%d\n' % (os.path.basename(x), 0) + f.write(s) + f.close() + + +def read_file_timestamp(ftimestamp): + d = {} + f = open(ftimestamp) + for line in f: + line = line.strip() + lst = line.split() + fname = lst[0] + t = lst[1] + d[fname] = int(t) + + f.close() + return d + + +def file_updated(fname, d): + ft = int(os.stat(fname).st_mtime) + k = os.path.basename(fname) + return ft > d[k] + + +def get_updated_files(lst, d): + result = [] + for x in lst: + if file_updated(x, d): + result.append(os.path.basename(x)) + return result + + +def get_sample_size(d, sorted_keys, day, rcond_string): + + if rcond_string.isdigit(): + return int(rcond_string) + + if len(d) == 0: + return 1200 # a default number of sample size, CHANGE + + for x in sorted_keys: + if x >= day: + return d[x] + + k = sorted_keys[-1] # last key, latest date + return d[k] + + +def number_rnaseq_id(tpm_file): + f = open(tpm_file) + first_line = f.readlines()[0] + f.close() + first_line = first_line.strip() + return len(first_line.split()) - 1 + + +def number_rnaseq_diff(para_file, tpm_file): + ''' count the number @ in para_file, and count the number of columns in tpm_file, return their difference ''' + a = 0 + f = open(para_file) + for line in f: + line = line.strip() + if line.startswith('@'): + a += 1 + f.close() + + b = number_rnaseq_id(tpm_file) + + return a - b + + +def validate_gene_file(fname): + f = open(fname) + lines = f.readlines() + f.close() + for line in lines: # check all lines + line = line.strip() + lst = line.split('\t') + if len(lst) < 6: + print('[update_network.py]:Not enought fields: %s. Only %d are given. Each line must have gene_id, gene_name, chr, start, end, strand, description (optional). See prepare_gene_file.py in the documentation on how to prepare this file.' % (line, len(lst))) + sys.exit() + + +def validate_parameter_for_buildcmatrix(fname): + # first the file must exist + if not os.path.exists(fname): + print('[update_network.py]:CANNOT FIND %s.' % (fname)) + sys.exit() + f = open(fname) + lines = f.readlines() + f.close() + d = {} + location_count = 0 + for line in lines: + line = line.strip() + if line.startswith('%%'): + k, v = get_key_value(line[2:]) + d[k] = v + if k == 'GENE_FILE' or k == 'CHR_INFO': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'GENE_FILE': + validate_gene_file(v) + if k == 'DESTINATION': + if not os.path.isdir(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'TARGET_RANGE': + if int(v) <= 0: + print('[update_network.py]:Target range (%d) must be greater than 0.' % (v)) + sys.exit() + if line.startswith('LOCATION:'): + v = get_value(line, ':') + location_count += 1 + if not os.path.exists(v): + print('[Warning] update_network.py: Location %s does not exists.' % (v)) + #sys.exit() + + if not 'GENE_FILE' in d: + print('[update_network.py]:Must specify GENE_FILE.') + sys.exit() + if not 'DESTINATION' in d: + print('[update_network.py]:Must specify DESTINATION.') + sys.exit() + if not 'CHR_INFO' in d: + print('[update_network.py]:Must specify CHR_INFO.') + sys.exit() + if location_count == 0: + print('[update_network.py]:Must contain at least one ChIP-seq.') + sys.exit() + + +def validate_parameter_for_buildrmatrix(fname): + # first the file must exist + if not os.path.exists(fname): + print('[update_network.py]:CANNOT FIND %s.' % (fname)) + sys.exit() + f = open(fname) + lines = f.readlines() + f.close() + d = {} + location_count = 0 + for line in lines: + line = line.strip() + if line.startswith('%%'): + k, v = get_key_value(line[2:]) + d[k] = v + if k == 'GENE_LIST': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if line.startswith('LOCATION:'): + v = get_value(line, ':') + location_count += 1 + if not os.path.exists(v): + print('[update_network.py]:Location %s does not exists.' % (v)) + #sys.exit() + + if not 'GENE_LIST' in d: + print('[update_network.py]:Must specify GENE_LIST.') + sys.exit() + if location_count == 0: + print('[update_network.py]:Must contain at least one RNA-seq.') + sys.exit() + + +def validate_parameter_for_net(fname): + # first the file must exist + if not os.path.exists(fname): + print('[update_network.py]:CANNOT FIND %s.' % (fname)) + sys.exit() + f = open(fname) + lines = f.readlines() + f.close() + d = {} + location_count = 0 + for line in lines: + line = line.strip() + if line.startswith('%%'): + k, v = get_key_value(line[2:]) + d[k] = v + if k == 'GENE_LIST': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'GENE_ID_AND_GENE_NAME': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'BINDING_INFO': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'EXPRESSION_INFO': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + sys.exit() + if k == 'BINDING_MATRIX': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + print('[update_network.py]:Use python3 buildCmatrix.py paramter_for_buildCmatrix.txt > binding.txt to create binding.txt.') + if k == 'EXPRESSION_MATRIX': + if not os.path.exists(v): + print('[update_network.py]:%s not exists.' % (v)) + print('[update_network.py]:Use python3 buildRmatrix.py paramter_for_buildRmatrix.txt to create TPM.txt.') + + if not 'GENE_LIST' in d: + print('[update_network.py]:Must specify GENE_FILE.') + sys.exit() + if not 'GENE_ID_AND_GENE_NAME' in d: + print('[update_network.py]:Must specify GENE_ID_AND_GENE_NAME.') + sys.exit() + if not 'BINDING_INFO' in d: + print('[update_network.py]:Must specify BINDING_INFO.') + sys.exit() + if not 'EXPRESSION_INFO' in d: + print('[update_network.py]:Must specify EXPRESSION_INFO.') + sys.exit() + if not 'BINDING_MATRIX' in d: + print('[update_network.py]:%s not exists.' % (v)) + print('[update_network.py]:Use python3 buildCmatrix.py paramter_for_buildCmatrix.txt > binding.txt to create binding.txt.') + if not 'EXPRESSION_MATRIX' in d: + print('[update_network.py]:%s not exists.' % (v)) + print('[update_network.py]:Use python3 buildRmatrix.py paramter_for_buildRmatrix.txt to create TPM.txt.') + + + +def need_update_parameter_file(param_file, dirs): + ''' Make sure param_file is consistent with dirs (a list of directories to check against). ''' + result = [] + + files_in_parameter = {} + f = open(param_file) + for line in f: + line = line.strip() + if line.startswith('LOCATION:'): + lst = line.split(':') + k = os.path.abspath(lst[1]) + files_in_parameter[k] = 1 + f.close() + param_modification_time = os.path.getmtime(param_file) + + files_in_dirs = {} + for directory in dirs: + for root, dirnames, filenames in os.walk(os.path.abspath(directory)): + for filename in filenames: + k = os.path.join(root, filename) + files_in_dirs[k] = 1 + if 'narrowPeak' in k or '_quant' in k: + if not k in files_in_parameter and os.path.getmtime(k) > param_modification_time: + result.append('%s is not in %s' % (k, param_file)) + + return result + + + +def validate_binding_file(fname): + f = open(fname) + lines = f.readlines() + for line in lines: + line = line.strip() + if 'buildCmatrix: ChIP-seq ID list is empty.' in line: + return False + f.close() + return True + + +def lines_with_10_fields(s): + result = [] + for line in s.split('\n'): + line = line.strip() + if len(line.split('\t')) == 10: + result.append(line) + return result + + +def concatenate_edge_files(fname_lst, fname_out): + fout = open(fname_out, 'w') + for fname in fname_lst: + f = open(fname) + s = f.read() + f.close() + # Make sure each edge has 10 fields before writing. + lines = lines_with_10_fields(s) + if lines != []: + write_log_file('[update_network.py] In function concatenate_edge_files. File %s has %d rows with 10 columns.' % (fname, len(lines)), UPDATE_NETWORK_LOG_FILE) + fout.write('\n'.join(lines) + '\n') + else: + write_log_file('[update_network.py] In function concatenate_edge_files. Check file %s. It has no rows with 10 fields.' % (fname), UPDATE_NETWORK_LOG_FILE) + fout.close() + + +def delete_edge_files(fname_lst): + for fname in fname_lst: + # Before we delete, we should make sure it is not being written. Make sure it is old enough. Otherwise, don't delete. + if age_of_file_in_seconds(fname) > 12*60*60: # 10 minutes + os.remove(fname) + else: + write_log_file('[update_network.py] In function delete_edge_files. Check file %s. It is probably still being written. So I don\'t delete it.' % (fname), UPDATE_NETWORK_LOG_FILE) + + +def create_edges0(): + if os.path.exists(PARAMETER_FOR_NET): + write_log_file('[update_network.py] Create simple edges.txt using create_edges0.py with %s' % (PARAMETER_FOR_NET), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0.py %s' % (PARAMETER_FOR_NET) + os.system(cmd) + + # The following commands are optional. For example, if a user wants to run it locally, he don't have to provide these TPM tables. + if os.path.exists(PARAMETER_FOR_NET_TRAVADB_STRESS): + #write_log_file('[update_network.py] Create simple edges.txt using create_edges0.py with %s' % (PARAMETER_FOR_NET_TRAVADB_STRESS), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0.py %s' % (PARAMETER_FOR_NET_TRAVADB_STRESS) + #os.system(cmd) + + if os.path.exists(PARAMETER_FOR_NET_TRAVADB_MAP): + #write_log_file('[update_network.py] Create simple edges.txt using create_edges0.py with %s' % (PARAMETER_FOR_NET_TRAVADB_MAP), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0.py %s' % (PARAMETER_FOR_NET_TRAVADB_MAP) + #os.system(cmd) + + if os.path.exists(PARAMETER_FOR_NET_MILD_DROUGHT): + #write_log_file('[update_network.py] Create simple edges.txt using create_edges0.py with %s' % (PARAMETER_FOR_NET_MILD_DROUGHT), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0.py %s' % (PARAMETER_FOR_NET_MILD_DROUGHT) + #os.system(cmd) + + if os.path.exists(PARAMETER_FOR_NET_WIGGELAB_DIURNAL): + #write_log_file('[update_network.py] Create simple edges.txt using create_edges0.py with %s' % (PARAMETER_FOR_NET_WIGGELAB_DIURNAL), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0.py %s' % (PARAMETER_FOR_NET_WIGGELAB_DIURNAL) + #os.system(cmd) + + +def create_edges0B(): + if os.path.exists(PARAMETER_FOR_NET): + write_log_file('[update_network.py] Create tissue-specific edges.txt using new binding.txt (size=%d). create_edges0B.py' % (num_ids(BINDING_FILE)), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 create_edges0B.py %s' % (PARAMETER_FOR_NET) # call correlation_per_tissue.R + os.system(cmd) + + +def wedge(): + if os.path.exists(PARAMETER_FOR_NET): + write_log_file('[update_network.py] Create edges using wedge shapes. wedge.R', UPDATE_NETWORK_LOG_FILE) + cmd = 'Rscript wedge.R' + os.system(cmd) + + +def correlation_per_group(): + # For 3,130 RNA-seq samples and 30,000 pairs, need at least 10 hours. + if os.path.exists(PARAMETER_FOR_NET): + write_log_file('[update_network.py] Create group-specific edges.txt using new TPM.txt (size=%d). correlation_per_group.R' % (number_rnaseq_id(TPM_FILE)), UPDATE_NETWORK_LOG_FILE) + cmd = 'Rscript correlation_per_group.R' + os.system(cmd) + + +def correlation_per_group_fixed_number(): + if os.path.exists(PARAMETER_FOR_NET): + write_log_file('[update_network.py] Create group-specific (fixed) edges.txt using new TPM.txt (size=%d). correlation_per_group_fixed_number.R' % (number_rnaseq_id(TPM_FILE)), UPDATE_NETWORK_LOG_FILE) + cmd = 'Rscript correlation_per_group_fixed_number.R' + os.system(cmd) + + +def correlation_mixtools(num_component): + if os.system('pidof R') != 0: # since it take long time (several days) to run create_edges_mixtool.R, so we make sure the previous R computing has finished before we start a new one. os.system returns 0 if R is running. + write_log_file('[update_network.py] Create edges.txt using TPM.txt (size=%d). create_edges_mixtool.R with %d components.' % (number_rnaseq_id(TPM_FILE), num_component), UPDATE_NETWORK_LOG_FILE) + cmd = 'Rscript create_edges_mixtool.R %d' % (num_component) + os.system(cmd) + + +def check_rnaseq_info(): + # check rnaseq_info_database.txt and rnaseq_info_database.json. If they are outdated, then remind us to update it in log file. + if os.path.exists(RNA_SEQ_INFO_DATABASE): + if age_of_file_in_days(RNA_SEQ_INFO_DATABASE) > 120: # older than 120 days + write_log_file('[update_network.py] Need update %s. It is %d days old.' % (RNA_SEQ_INFO_DATABASE, age_of_file_in_days(RNA_SEQ_INFO_DATABASE)), UPDATE_NETWORK_LOG_FILE) + else: + write_log_file('[update_network.py] [MISSING] Must create %s.' % (RNA_SEQ_INFO_DATABASE), UPDATE_NETWORK_LOG_FILE) + + if os.path.exists(RNA_SEQ_INFO_DATABASE_JSON): + if age_of_file_in_days(RNA_SEQ_INFO_DATABASE_JSON) > 120: + write_log_file('[update_network.py] Need update %s. It is %d days old.' % (RNA_SEQ_INFO_DATABASE_JSON, age_of_file_in_days(RNA_SEQ_INFO_DATABASE_JSON)), UPDATE_NETWORK_LOG_FILE) + else: + write_log_file('[update_network.py] [MISSING] Must create %s.' % (RNA_SEQ_INFO_DATABASE_JSON), UPDATE_NETWORK_LOG_FILE) + + +# main +FILE_LIST_TO_CHECK = [PARAMETER_FOR_BUILDCMATRIX, PARAMETER_FOR_BUILDRMATRIX, PARAMETER_FOR_NET, \ + MERGED_EDGE_FILE, BINDING_FILE, TPM_FILE] # a list of important files + +make_important_dirs() # make important directories (if non-existent) for holding various kinds of files, must be put after os.chdir(CODE_DIR) +#validate_webapp_dir(PARAMETER_FOR_NET) # make sure the directory Webapp contains necessary files, e.g., genes.json. + +check_rnaseq_info() # rnaseq informtion is useful for displaying scatterplots + +# Make sure all necessary files are present, if not, make them if possible +miss_lst = all_files_present(FILE_LIST_TO_CHECK) # check if any of them are missing +if miss_lst != []: # miss_lst is non-empty in the beginning. + print('These mandatory files are missing: %s.\nPrepare them first.' % (' '.join(miss_lst))) + write_log_file('[update_network.py] Cannot find these required files:%s' % (' '.join(miss_lst)), UPDATE_NETWORK_LOG_FILE) + + # initially, we (at most) only have three parameter files, no binding.txt, TPM.txt or edges.txt ... + important_miss_number = 0 + if PARAMETER_FOR_BUILDCMATRIX in miss_lst: + print('[update_network.py]: must prepare %s first.' % (PARAMETER_FOR_BUILDCMATRIX)) + important_miss_number += 1 + + if PARAMETER_FOR_BUILDRMATRIX in miss_lst: + print('[update_network.py]: must prepare %s first.' % (PARAMETER_FOR_BUILDRMATRIX)) + important_miss_number += 1 + + if PARAMETER_FOR_NET in miss_lst: + print('[update_network.py]: must prepare %s first.' % (PARAMETER_FOR_NET)) + important_miss_number += 1 + + if important_miss_number > 0: + sys.exit() # need to provide all the above three files; otherwise cannot proceed + + if BINDING_FILE in miss_lst: + print('[update_network.py]: make initial binding.txt ... wait') + write_log_file('[update_network.py] Make initial binding.txt', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 get_binding.py %s' % (PARAMETER_FOR_BUILDCMATRIX) + #os.system(cmd) + cmd = 'python3 buildCmatrix.py %s > %s' % (PARAMETER_FOR_BUILDCMATRIX, BINDING_FILE) + #os.system(cmd) + print('[update_network.py]: IMPORATNT: make sure BINDING_MATRIX in %s was set %s and rerun update_network.py.' % (PARAMETER_FOR_NET, BINDING_FILE)) + sys.exit() + + if TPM_FILE in miss_lst: + print('[update_network.py]: make initial TPM.txt ... wait') + write_log_file('[update_network.py] Make initial TPM.txt', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 buildRmatrix.py %s' % (PARAMETER_FOR_BUILDRMATRIX) # produce TPM.txt + os.system(cmd) + print('[update_network.py]:IMPORTANT: make sure EXPRESSION_MATRIX in %s was set %s and rerun update_network.py.' % (PARAMETER_FOR_NET, TPM_FILE)) + sys.exit() + + miss_lst2 = all_files_present(FILE_LIST_TO_CHECK) # check files again + if len(miss_lst2) == 1 and miss_lst2[0] == MERGED_EDGE_FILE: # all other files are ready except edges.txt, make one. + print('[update_network.py]: make initial edges.txt ... wait') + create_edgeds0() + + +# Make json2 (sliced binding.txt) if it does not exist. Copy json2 to +# the web application folder static/edges [do it manually] for displaying +# binding strength plots. +if not os.path.isdir('../Data/history/bind/json2') and os.path.exists(BINDING_FILE): + write_log_file('Make directory ../Data/history/bind/json2. Don\'t forget to copy json2 to static/edges in the web application.', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 slice_binding_to_JSON.py %s' % (PARAMETER_FOR_NET) + os.system(cmd) + + +# Make json (sliced TPM.txt) if it does not exist. Copy json to the +# web application folder static/edges [manual] for displaying gene +# expression scatterplots. +if not os.path.isdir('../Data/history/expr/json') and os.path.exists(TPM_FILE): + write_log_file('Make directory ../Data/history/expr/json. Don\'t forget to copy json to static/edges in the web application.', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 slice_TPM_to_JSON.py %s' % (PARAMETER_FOR_NET) + os.system(cmd) + + +# Make sure parameter files are present and valid (rudimentary check but important) +validate_parameter_for_buildcmatrix(PARAMETER_FOR_BUILDCMATRIX) +validate_parameter_for_buildrmatrix(PARAMETER_FOR_BUILDRMATRIX) +validate_parameter_for_net(PARAMETER_FOR_NET) + + +# If the file timestamp does not exist, create one +if not os.path.exists(FILE_TIMESTAMP): + record_file_time(FILE_LIST_TO_CHECK, FILE_TIMESTAMP) + +# Get update time of mandatory files +timestamp_dict = read_file_timestamp(FILE_TIMESTAMP) + + + +################## binding.txt stuff ##################################### +# Check parameter_for_buildCmatrix.txt +updated_file_list = get_updated_files(FILE_LIST_TO_CHECK, timestamp_dict) +if 'parameter_for_buildCmatrix.txt' in updated_file_list and not hold_on(PARAMETER_FOR_BUILDCMATRIX): + write_log_file('[update_network.py] Parameter file %s has been updated.' % (PARAMETER_FOR_BUILDCMATRIX), UPDATE_NETWORK_LOG_FILE) + write_log_file('[update_network.py] Make binding column files', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 get_binding.py %s' % (PARAMETER_FOR_BUILDCMATRIX) # won't re-compute existing binding columns unless updated + os.system(cmd) + + +updated_file_list = get_updated_files(FILE_LIST_TO_CHECK, timestamp_dict) +if 'binding.txt' in updated_file_list: + write_log_file('[update_network.py] binding.txt has been updated. This update will take effect next time TPM.txt is updated.', UPDATE_NETWORK_LOG_FILE) + # create_edges0() + # create_edges0B() + # wedge() + # correlation_per_group() + # correlation_per_group_fixed_number() + # correlation_mixtools(2) + # correlation_mixtools(3) + + ## TODO mixtool stuff, forget it for now. + #cmd = 'nohup python3 create_edges4.py %s &' % (temp_file_name) + #os.system(cmd) + + + + +################## TPM.txt stuff ##################################### + +# update parameter_for_buildRmatrix.txt periodically and automatically. +if datetime.now().day % BUILDRMATRIX_RENEW_INTERVAL == 0: # check if need to update parameter_for_buildRmatrix.txt bi-weekly + curr_time = datetime.now().strftime('%Y%m%d%H%M') + new_parameter_file = '../Data/temp/parameter_for_buildRmatrix.%s' % (curr_time) + cmd = 'python3 make_parameter_rnaseq.py > %s' % (new_parameter_file) # new_parameter_file will not be updated unless download_and_map.py has finished. + os.system(cmd) + num = number_rnaseq_diff(new_parameter_file, TPM_FILE) + if num >= MIN_RNA_SEQ_INCREASE: # sufficient number of RNA-seq samples have been added + write_log_file('[update_network.py] Update %s' % (PARAMETER_FOR_BUILDRMATRIX), UPDATE_NETWORK_LOG_FILE) + cmd = 'cp %s %s' % (new_parameter_file, PARAMETER_FOR_BUILDRMATRIX) + os.system(cmd) + else: + write_log_file('[update_network.py] You have downloaded %d RNA-seq since last build of TPM.txt. TPM.txt will be rebuilt if this number reaches %d.' % (num, MIN_RNA_SEQ_INCREASE), UPDATE_NETWORK_LOG_FILE) + + +# Check if parameter_for_buildRmatrix.txt has been updated +updated_file_list = get_updated_files(FILE_LIST_TO_CHECK, timestamp_dict) +# TODO To simplify things, I will provide TPM.txt directly. So set the +# HOLDON option to YES in parameter_for_buildRmatrix.txt to prevent +# the following from being True. +if 'parameter_for_buildRmatrix.txt' in updated_file_list and not hold_on(PARAMETER_FOR_BUILDRMATRIX): + write_log_file('[update_network.py] Parameter file %s has been updated.' % (PARAMETER_FOR_BUILDRMATRIX), UPDATE_NETWORK_LOG_FILE) + write_log_file('[update_network.py] Rebuild TPM.txt ...', UPDATE_NETWORK_LOG_FILE) + curr_time = datetime.now().strftime('%Y%m%d%H%M%S') + if os.path.exists(TPM_FILE): + backup_file_name = '../Data/history/expr/TPM.txt.backup.at.%s' % (curr_time) + cmd = 'cp %s %s' % (TPM_FILE, backup_file_name) + os.system(cmd) + cmd = 'gzip %s' % (backup_file_name) + os.system(cmd) + + cmd = 'python3 buildRmatrix.py %s' % (PARAMETER_FOR_BUILDRMATRIX) # produce TPM.txt, whose location is specified in TPM_TABLE in buidlRmatrix.py + os.system(cmd) + + curr_date = datetime.now().strftime('%Y%m%d') + tpm_sample_size = number_rnaseq_id(TPM_FILE) + write_sample_size_file(SAMPLE_SIZE_FILE, curr_date, tpm_sample_size) + + + +# Create edges using all RNA-seq experiments +updated_file_list = get_updated_files(FILE_LIST_TO_CHECK, timestamp_dict) +if 'TPM.txt' in updated_file_list: # we could _touch_ TPM.txt to make it recent. We will recompute edges using the full binding.txt. + # Make a full binding.txt since we are going to use the new TPM.txt to recompute all edges + write_log_file('[update_network.py] Build full binding matrix for the new TPM.txt.', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 buildCmatrix.py %s include-all > %s' % (PARAMETER_FOR_BUILDCMATRIX, BINDING_FILE) # include all ChIP-seq IDs. Pay attention to include-all in the command-line argument. + os.system(cmd) + + # target_tf.txt + write_log_file('[update_network.py] Make target_tf.txt.', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 make_target_tf.py %s > %s' % (PARAMETER_FOR_NET, TARGET_TF_FILE) + os.system(cmd) + + write_log_file('[update_network.py] Update ../Data/history/expr/json using the new TPM.txt. Don\'t forget to update the static/edges/json folder in the web application.', UPDATE_NETWORK_LOG_FILE) + ## json -- make/renew json directory for displaying scatterplots + cmd = 'python3 slice_TPM_to_JSON.py %s' % (PARAMETER_FOR_NET) + ## os.system(cmd) # turn this on if we are going to use this TPM.txt for displaying scatterplots + write_log_file('[update_network.py] Update directory ../Data/history/bind/json2. Don\'t forget to copy json2 to static/edges in the web application.', UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 slice_binding_to_JSON.py %s' % (PARAMETER_FOR_NET) + #os.system(cmd) # turn this on if we are going to use this binding.txt for displaying bar charts of binding strengths + ## copy ../Data/history/bind/json2 and ../Data/history/expr/json to the web application folder 'static/edges' [manual] + + if False: # TODO For now I will always use travadb's TPM.txt (138 columns) to display scatterplots. Simpler and faster. + write_log_file('Assign tissue, refine tissue and update rnaseq_info_database.json', UPDATE_NETWORK_LOG_FILE) + os.environ["PYTHONIOENCODING"] = "UTF-8" # for non-ascii letters in ENA RNA-sample description. If this statement does not work, try 'export PYTHONIOENCODING=UTF-8' in the command line instead. The export command can be put in crontab -e before running this script + cmd = 'python3 assign_tissue.py' + os.system(cmd) + cmd = 'python3 refine_tissue.py > ../Data/information/experiment.and.tissue.2.txt' + os.system(cmd) + cmd = 'python3 update_rnaseq_info_json.py' + os.system(cmd) + + + + # Compute edges. This could take a lot of time so update FILE_TIMESTAMP first. + record_file_time(FILE_LIST_TO_CHECK, FILE_TIMESTAMP) + create_edges0() + create_edges0B() + wedge() + correlation_per_group() + correlation_per_group_fixed_number() + correlation_mixtools(2) # two components + #correlation_mixtools(3) + + +########## Merge edges ####################### +# update edges.txt, a merged file from two sources, HISTORY_DIR and HISTORY_DIR2. Some new edge files are being generated ... +time.sleep(5) +edge_file_lst = [] # collect edge files. +most_recent_edge_modification_time = 0 +write_log_file('[update_network.py] Look at edge files in %s.' % (HISTORY_DIR), UPDATE_NETWORK_LOG_FILE) +for fname in glob.glob(os.path.join(HISTORY_DIR, 'edges.txt.*')): # many small edges.txt.* are to be merged + edge_file_lst.append(fname) + if os.path.getmtime(fname) > most_recent_edge_modification_time: + most_recent_edge_modification_time = os.path.getmtime(fname) + +write_log_file('[update_network.py] Look at edge files in %s.' % (HISTORY_DIR2), UPDATE_NETWORK_LOG_FILE) +for fname in glob.glob(os.path.join(HISTORY_DIR2, 'edges.txt.*')): # edges.txt.* are to be merged + edge_file_lst.append(fname) + if os.path.getmtime(fname) > most_recent_edge_modification_time: + most_recent_edge_modification_time = os.path.getmtime(fname) + + +if edge_file_lst == []: + write_log_file('[update_network.py] No edge files to merge in %s and %s.' % (HISTORY_DIR, HISTORY_DIR2), UPDATE_NETWORK_LOG_FILE) +elif os.path.getmtime(MERGED_EDGE_FILE) < most_recent_edge_modification_time: # update edges.txt only if there are newer edges to add. + # concatenate edge files into one + write_log_file('[update_network.py] Concatenate edge files in %s and %s into one file.' % (HISTORY_DIR, HISTORY_DIR2), UPDATE_NETWORK_LOG_FILE) + curr_time = datetime.now().strftime('%Y%m%d_%H%M') + concatenate_edge_files(edge_file_lst, os.path.join(EDGE_POOL_DIR, 'edges.txt.many.one.targets.' + curr_time)) + delete_edge_files(edge_file_lst) + +if os.path.getmtime(MERGED_EDGE_FILE) < os.path.getmtime(EDGE_POOL_DIR): # edge pool directory has been updated, create new edges.txt + write_log_file('[update_network.py] Make a new edges.txt from edge files in %s.' % (EDGE_POOL_DIR), UPDATE_NETWORK_LOG_FILE) + write_log_file('[update_network.py] Number of lines in the old edges.txt: %d.' % (num_line(MERGED_EDGE_FILE)), UPDATE_NETWORK_LOG_FILE) + cmd = 'python3 merge_edges.py' + os.system(cmd) + write_log_file('[update_network.py] Number of lines in the new edges.txt: %d.' % (num_line(MERGED_EDGE_FILE)), UPDATE_NETWORK_LOG_FILE) + manual_copy_commands = 'Please copy files to the web application: sudo cp /home/lanhui/brain/Data/temp/edges.txt /var/www/brain/brain/static/edges/edges.txt sudo cp /home/lanhui/brain/Data/temp/html_edges/edges.sqlite /var/www/brain/brain/static/edges' + write_log_file('[update_network.py] %s' % (manual_copy_commands), UPDATE_NETWORK_LOG_FILE) + + +# exclude edges as suggested by Phil Wigge. +# write_log_file('Exclude edges (now ineffective)', UPDATE_NETWORK_LOG_FILE) +# cmd = 'python3 exclude_edges.py %s' % (EDGE_FILE) +#os.system(cmd) + +# # check if parameter_for_net.txt, or TPM.txt is updated, if yes, create edges. +# updated_file_list = get_updated_files(FILE_LIST_TO_CHECK, timestamp_dict) +# if ('parameter_for_net.txt' in updated_file_list or 'TPM.txt' in updated_file_list) and not hold_on(PARAMETER_FOR_NET): +# write_log_file('Create edges.txt using new TPM.txt (size=%d) ...' % (number_rnaseq_id(TPM_FILE)), UPDATE_NETWORK_LOG_FILE) +# time.sleep(7200) # wait one hour for the previous create_edges4.py (if any) to finish creating JSON_DIR and target_tf_fname +# cmd = 'nohup python3 create_edges4.py %s &' % (PARAMETER_FOR_NET) # put process to background +# os.system(cmd) +# time.sleep(60) + + +# remove .R files in ../Data/temp. Files older than 3 days will be removed +cmd = 'find %s -mtime +2 -name \"*.R\" -delete' % (TEMP_DIR) +os.system(cmd) + +# update time stamp file +record_file_time(FILE_LIST_TO_CHECK, FILE_TIMESTAMP) + +write_log_file('[update_network.py] Update done at %s.\n\n' % (datetime.now().strftime('%Y-%m-%d %H:%M:%S')), UPDATE_NETWORK_LOG_FILE) + diff --git a/Code/.#update_network.py b/Code/.#update_network.py new file mode 120000 index 0000000..66ec7a4 --- /dev/null +++ b/Code/.#update_network.py @@ -0,0 +1 @@ +lanhui@VM-0-14-ubuntu.13592:1688234417
\ No newline at end of file diff --git a/Code/merge_edges.py b/Code/merge_edges.py index 2fc034c..780da95 100644 --- a/Code/merge_edges.py +++ b/Code/merge_edges.py @@ -188,7 +188,7 @@ for fname in sorted(edges_file_lst, key=lambda t: -os.stat(t).st_mtime): # sort f.close() # use the most recent 365 edge files because they might be the best - if file_count >= 300: + if file_count >= 365: break write_log_file('[merge_edges.py] BRAIN has collected edges from %d files.' % (file_count) , UPDATE_NETWORK_LOG_FILE) @@ -199,6 +199,7 @@ if not os.path.isdir(folder_path): os.mkdir(folder_path) +# Write edge information to a text file called edges.txt write_log_file('[merge_edges.py] Make text edge file %s ...' % (MERGED_EDGE_FILE), UPDATE_NETWORK_LOG_FILE) fout = open(MERGED_EDGE_FILE, 'w') for k in d: @@ -207,10 +208,11 @@ for k in d: fout.close() +# Write edge information to SQLite database file, which could be opened using software called "DB Browser for SQLite" +# Biologists may use this file for filtering the information they need. db_fname = os.path.join(folder_path, 'edges.sqlite') write_log_file('[merge_edges.py] Make SQLite database file %s.' % (db_fname), UPDATE_NETWORK_LOG_FILE) - if os.path.exists(db_fname): os.remove(db_fname) |