4
4
5
5
import os
6
6
import argparse
7
- from lightdock .constants import DEFAULT_CLUSTER_FOLDER , GSO_OUTPUT_FILE , EVALUATION_FILE , SCORING_FILE , \
7
+ from lightdock .constants import DEFAULT_SWARM_FOLDER , GSO_OUTPUT_FILE , EVALUATION_FILE , SCORING_FILE , \
8
8
LIGHTDOCK_PDB_FILE , CLUSTER_REPRESENTATIVES_FILE
9
9
from lightdock .util .logger import LoggingManager
10
10
from lightdock .util .analysis import read_rmsd_and_contacts_data , read_lightdock_output , write_ranking_to_file , \
11
11
read_cluster_representatives_file
12
12
13
13
14
- log = LoggingManager .get_logger ('lightdock_rank ' )
14
+ log = LoggingManager .get_logger ('lgd_rank ' )
15
15
16
16
17
17
def parse_command_line ():
18
- parser = argparse .ArgumentParser (prog = 'lightdock_rank ' )
18
+ parser = argparse .ArgumentParser (prog = 'lgd_rank ' )
19
19
parser .add_argument ("num_clusters" , help = "number of clusters to consider" , type = int , metavar = "num_clusters" )
20
20
parser .add_argument ("steps" , help = "steps to consider" , type = int , metavar = "steps" )
21
21
parser .add_argument ("-c" , "--clashes_cutoff" , help = "clashes cutoff" , dest = "clashes_cutoff" , type = float )
@@ -39,18 +39,18 @@ def parse_command_line():
39
39
num_clusters_found = 0
40
40
for cluster_id in range (args .num_clusters ):
41
41
if args .result_file :
42
- result_file_name = os .path .join (DEFAULT_CLUSTER_FOLDER + str (cluster_id ), args .result_file )
42
+ result_file_name = os .path .join (DEFAULT_SWARM_FOLDER + str (cluster_id ), args .result_file )
43
43
else :
44
- result_file_name = os .path .join (DEFAULT_CLUSTER_FOLDER + str (cluster_id ),
44
+ result_file_name = os .path .join (DEFAULT_SWARM_FOLDER + str (cluster_id ),
45
45
(GSO_OUTPUT_FILE % args .steps ))
46
46
47
- cluster_representatives_file = os .path .join (DEFAULT_CLUSTER_FOLDER + str (cluster_id ),
47
+ cluster_representatives_file = os .path .join (DEFAULT_SWARM_FOLDER + str (cluster_id ),
48
48
CLUSTER_REPRESENTATIVES_FILE )
49
49
clusters = []
50
50
if os .path .isfile (cluster_representatives_file ) and not args .ignore_clusters :
51
51
clusters = read_cluster_representatives_file (cluster_representatives_file )
52
52
53
- scoring_file_name = os .path .join (DEFAULT_CLUSTER_FOLDER + str (cluster_id ), SCORING_FILE )
53
+ scoring_file_name = os .path .join (DEFAULT_SWARM_FOLDER + str (cluster_id ), SCORING_FILE )
54
54
try :
55
55
results = read_lightdock_output (result_file_name )
56
56
num_clusters_found += 1
0 commit comments