@@ -67,7 +67,6 @@ def build_dataset(step, word_dict, article_max_len, summary_max_len, toy=False):
67
67
title_list = get_text_list (train_title_path , toy )
68
68
elif step == "valid" :
69
69
article_list = get_text_list (valid_article_path , toy )
70
- title_list = get_text_list (valid_title_path , toy )
71
70
else :
72
71
raise NotImplementedError
73
72
@@ -76,11 +75,13 @@ def build_dataset(step, word_dict, article_max_len, summary_max_len, toy=False):
76
75
x = list (map (lambda d : d [:article_max_len ], x ))
77
76
x = list (map (lambda d : d + (article_max_len - len (d )) * [word_dict ["<padding>" ]], x ))
78
77
79
- y = list (map (lambda d : word_tokenize (d ), title_list ))
80
- y = list (map (lambda d : list (map (lambda w : word_dict .get (w , word_dict ["<unk>" ]), d )), y ))
81
- y = list (map (lambda d : d [:(summary_max_len - 1 )], y ))
82
-
83
- return x , y
78
+ if step == "valid" :
79
+ return x
80
+ else :
81
+ y = list (map (lambda d : word_tokenize (d ), title_list ))
82
+ y = list (map (lambda d : list (map (lambda w : word_dict .get (w , word_dict ["<unk>" ]), d )), y ))
83
+ y = list (map (lambda d : d [:(summary_max_len - 1 )], y ))
84
+ return x , y
84
85
85
86
86
87
def batch_iter (inputs , outputs , batch_size , num_epochs ):
0 commit comments