You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
importtqdm, nltk, torch, time, numpyasnp, argparse, os
7
5
fromutils_logplotimportLogPlot
8
-
fromcoverageimportKeywordCoverage
9
-
importutils_hdf5
6
+
frommodel_coverageimportKeywordCoverage
7
+
fromutils_datasetimportSQLDataset
8
+
importutils_misc
10
9
11
10
parser=argparse.ArgumentParser()
12
11
parser.add_argument("--experiment", type=str, required=True, help="Experiment name. Will be used to save a model file and a log file.")
12
+
parser.add_argument("--dataset_file", type=str, required=True, help="Which dataset file to use. Can be full path or the root folder will be attached.")
parser.add_argument("--experiment", type=str, required=True, help="Experiment name. Will be used to save a model file and a log file.")
15
12
parser.add_argument("--dataset_file", type=str, required=True, help="Which dataset file to use.")
16
-
parser.add_argument("--task", type=str, help="Which generation task to perform. Can be: `cgen` (conditionally generate), lm` (language modeling) or `copy`")
13
+
parser.add_argument("--task", type=str, required=True, help="Which generation task to perform. Can be: `cgen` (conditionally generate), lm` (language modeling) or `copy`. `cgen` is useful to train a supervised model, when data is available (for example a headline generator, summarizer, etc). `lm` is an unconditional language model, such as the GPT2 model, can be used to train a Fluency model. `copy` can be used to pretrain the generator for the summary_loop, this speeds up training of the summary_loop as the generator already starts with the strong baseline of copying the first K words of the input.")
17
14
parser.add_argument("--max_output_length", required=True, type=int, help="Maximum output length. Saves time if the sequences are short.")
0 commit comments