-
Notifications
You must be signed in to change notification settings - Fork 62
/
Copy pathMakefile
402 lines (341 loc) · 14.6 KB
/
Makefile
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
#
# Experiment management for Tgen
#
# Environment settings
SHELL = /bin/bash
TGEN = ../run_tgen.py
MEM = 8g
#QP = -q *@a*,*@c*,*@f*,*@hyp*,*@i*,*@l*,*@o*,*@p*,*@t*# avoid Hydra and Belzebub
#QP = --exclude-comp hyperion5
QP = --queue troja-all.q
QSUBMIT = qsubmit --logdir '$(TRY_DIR)' --mem $(MEM) --jobname T.$(TRY_NUM).$(RUN_NAME) $(QP) $(QHOLD)# using OB's qsubmit script
ifdef PYPY
ACTIVATE = source ~/work/tools/virtualenv-pypy-x86_64/bin/activate
else
ACTIVATE = echo -n
endif
ifdef DEBUG
DEBUG_LOG_TRAIN = -d $(TRY_DIR)/percrank.debug-log.txt.gz
DEBUG_LOG_GEN = -d $(TRY_DIR)/asearch.debug-log.txt
endif
ifdef DEBUG_GEN
DEBUG_LOG_GEN = -d $(TRY_DIR)/asearch.debug-log.txt
endif
MTEVAL = ../util/mteval-v11b.pl
AVERAGE_SCORES = ../util/average_scores.pl
GREP_SCORES = ../util/grep_scores.pl
DESC_EXP = ../util/describe_experiment.pl
LOG_CUTOFF= 10000
NUM_EXP = 20 # number of experiments to list in `make desc'
LOGFILES = $$file/T.* $$file/{asearch_gen.,gen.,seq2seq_gen.,}log.txt # logfile pattern to use for score printing
SCOREFILE = SCORE
# Runs directories
RUNS_DIR := runs# main directory for experiment outputs
TRY_NUM := $(shell perl -e '$$m=0; for(<$(RUNS_DIR)/*>){/\/(\d+)_/ and $$1 > $$m and $$m=$$1;} printf "%03d", $$m+1;')# experiment number
RUN_NAME := experiment# default name, to be overridden by targets
DATE := $(shell date +%Y-%m-%d_%H-%M-%S)
TRY_DIR = $(RUNS_DIR)/$(TRY_NUM)_$(DATE)_$(RUN_NAME)/$(CV_NUM)# experiment output directory
TRAINING_SET := training2
CV_NUMS := cv00 cv01 cv02 cv03 cv04 cv05 cv06 cv07 cv08 cv09
ifdef D # Shortcut D -> DESC
DESC := $(D)
endif
ifdef J # Shortcut J -> JOBS
JOBS := $(J)
endif
ifdef JOBS
PARALLEL := -j $(JOBS) -e "T.$(TRY_NUM)"
endif
# Input data file defaults
CANDGEN_MODEL = model/$(TRAINING_SET)/$(CV_NUM)/candgen.pickle.gz
#PERCRANK_MODEL = model/percrank.pickle.gz
PERCRANK_CONFIG = config/percrank.py
SEQ2SEQ_CONFIG = config/seq2seq.py
ASEARCH_CONFIG = config/asearch.py
CLASSIF_CONFIG = config/classif.py
TRAIN_PORTION = 1.0
TRAIN_DAS = data/$(TRAINING_SET)/$(CV_NUM)/train-das.txt
TRAIN_TREES = data/$(TRAINING_SET)/$(CV_NUM)/train-text.pickle.gz
TRAIN_ABSTR = data/$(TRAINING_SET)/$(CV_NUM)/train-abstr.txt
TRAIN_CONC = data/$(TRAINING_SET)/$(CV_NUM)/train-conc.sgm
GRAMMATEMES = data/$(TRAINING_SET)/$(CV_NUM)/grammatemes.tsv
TEST_DAS = data/$(TRAINING_SET)/$(CV_NUM)/test-das.txt
TEST_REF = data/$(TRAINING_SET)/$(CV_NUM)/test-das.sgm
TEST_TREES = data/$(TRAINING_SET)/$(CV_NUM)/test-text.pickle.gz
TEST_ABSTR = data/$(TRAINING_SET)/$(CV_NUM)/test-abstr.txt
TEST_CONC = data/$(TRAINING_SET)/$(CV_NUM)/test-conc.sgm
SURFACE_SCEN = surface/en_synthesis.scen
XS_INSTEAD = 0
EVAL_SELECTOR =
TARGET_SELECTOR = gen
ifdef CV_NUM
TEST_CONC = data/$(TRAINING_SET)/$(CV_NUM)/test-text.sgm
XS_INSTEAD = 1
ifneq (,$(findstring norep,$(TRAINING_SET)))
XS_INSTEAD = "\#"
endif
EVAL_SELECTOR = ref0
TEST_ABSTR = data/$(TRAINING_SET)/$(CV_NUM)/test.0-abstr.txt
endif
# Help text
define HELP_MESSAGE
Tgen experiments
================
make desc [NUM_EXP=20] [REFRESH=1]
* List experiments with descriptions and (possibly) scores.
* Change NUM_EXP if you need to see more than 20 latest results.
* Refresh scores to obtain new results by setting REFRESH=1 (makes it slower).
make percrank_train [J=XX] D='Description'
* Run a new percrank training experiment
- set TRAIN_PORTION to a fraction of the training
data to be used
- set J/JOBS to the number of parallel jobs to be used
in a parallel training setup
make asearch_gen D='Description'
* Run A*search generation
(set PERCRANK_MODEL to the model you want to use).
make asearch_gen-XXX
* Directly test the model trained from experiment number
XXX.
make rerun-XXX
* Re-run experiment number XXX.
- Will find the last log file and select the experiment type
accordingly.
make norm_log LOGFILE=path/to/debug.log
* Normalize debug log files for diffing (stripping dates)
- Set LOG_CUTOFF to adjust the number of lines where the logs
will be cut off.
make cv_run J=[XX] D='Description'
* Run a 10-fold cross-validation experiment where percrank
training is directly followed by A*search generation.
- set SEEDS=1 to use 5 different random seeds for each CV run.
make cv_score-XXX
* Retrieve a cross-validation run score.
make compare_output TRY_DIR=runs/XXX/cvXX
* Compare the output of the given run to gold standard
---
Use QP='--console' if you want the job to run interactively instead
of submitting to the cluster.
Use ACTIVATE='source path/to/virtualenv/bin/activate' to run
within a Virtualenv instead of the default Python.
Use PYPY=1 to activate the preset PyPy Virtualenv.
Use DEBUG=1 to activate debugging outputs. Use DEBUG_GEN=1 to activate
debugging only for testing (in CV runs).
endef
export HELP_MESSAGE
#
# Targets
#
# Auxiliary targets
help:
@echo "$$HELP_MESSAGE" | egrep --color '^(\s*make.*|)'
# List last NUM_EXP experiments, find scores in logs (of various versions)
desc:
@ls -d $(RUNS_DIR)/* | \
perl -e '@lns = <>; print join "", sort { ($$na) = ( $$a =~ /\/([0-9]+)_/ ); ($$nb) = ( $$b =~ /\/([0-9]+)_/ ); $$na <=> $$nb } @lns;' | \
tail -n $(NUM_EXP) | while read file; do \
if [[ -f $$file/$(SCOREFILE) && -z "$(REFRESH)" ]]; then \
cat $$file/$(SCOREFILE) ; \
continue ; \
fi ; \
echo -ne $$file ":\t" | sed 's/runs\///;s/_/\t/;s/_/ /;s/_.*/ /;' > $$file/$(SCOREFILE); \
$(GREP_SCORES) $(LOGFILES) >> $$file/$(SCOREFILE) ; \
echo -ne ' ' >> $$file/$(SCOREFILE) ; \
cat $$file/ABOUT | tr '\n' ' ' >> $$file/$(SCOREFILE); \
echo >> $$file/$(SCOREFILE); \
cat $$file/$(SCOREFILE); \
done
train_desc:
make desc SCOREFILE=TRAIN_SCORE LOGFILES='$$$$file/percrank_train.log.txt' NUM_EXP=$(NUM_EXP) REFRESH=$(REFRESH)
printvars:
$(foreach V, $(sort $(.VARIABLES)), $(if $(filter-out environment% default automatic, $(origin $V)), $(info $V=$($V) ($(value $V)))))
printgit:
@git status
@echo -e "\n*** *** ***\n"
@git log --pretty=format:"%h - %an, %ar : %s" -1
@echo -e "\n*** *** ***\n"
@git diff
prepare_dir:
# create the experiment directory, print description and git version
mkdir -p $(TRY_DIR)
if [[ "$(RUN_TYPE)" == "seq2seq" ]]; then \
CONFIG=$(SEQ2SEQ_CONFIG) ; \
else \
CONFIG=$(PERCRANK_CONFIG) ; \
fi; \
[[ -n "$(DEBUG)" || -n "$(DEBUG_GEN)" ]] && DEBUG_SETTING=-d ; \
[[ -n "$(RANDS)" ]] && RANDS_SETTING=-r ; \
$(DESC_EXP) -t "$(TRAINING_SET)" -p "$(TRAIN_PORTION)" -j "$(JOBS)" -c "$(CV_NUMS)" $$DEBUG_SETTING $$RANDS_SETTING $$CONFIG > $(TRY_DIR)/ABOUT ; \
echo "$(DESC)" >> $(TRY_DIR)/ABOUT
make printvars > $(TRY_DIR)/VARS
make printgit > $(TRY_DIR)/GIT_VERSION
# Main targets
asearch_cv_run: RUN_TYPE := asearch
asearch_cv_run: cv_run
seq2seq_cv_run: RUN_TYPE := seq2seq
seq2seq_cv_run: cv_run
cv_run: RUN_NAME := cv_run
cv_run: prepare_dir
cv_run:
for CV_NUM in $(CV_NUMS); do \
if [[ -n "$(RANDS)" ]]; then \
SEEDS=(s0 s1 s2 s3 s4); \
else \
SEEDS=(""); \
fi; \
for SEED in "$${SEEDS[@]}"; do \
if [[ "$(RUN_TYPE)" == "seq2seq" ]]; then \
TRAIN_RUN=seq2seq_train ; \
TEST_RUN=seq2seq_gen_cv ; \
else \
TRAIN_RUN=percrank_train ; \
TEST_RUN=asearch_gen_cv ; \
fi; \
make $$TRAIN_RUN TRY_DIR=$(TRY_DIR)/$$CV_NUM$$SEED SEED=$$SEED TRY_NUM=$(TRY_NUM) QP="$(QP)" CV_NUM=$$CV_NUM 2>&1 | tee -a $(TRY_DIR)/cv.log.txt ; \
JOB_NUM=`cat $(TRY_DIR)/cv.log.txt | grep '^Your job' | tail -n 1 | sed 's/^Your job \([0-9]\+\).*/\1/;' ` ; \
make $$TEST_RUN TRY_DIR=$(TRY_DIR)/$$CV_NUM$$SEED TRY_NUM=$(TRY_NUM) CV_NUM=$$CV_NUM QP="$(QP)" QHOLD="--hold $$JOB_NUM"; \
done; \
done
cv_score-%:
$(eval TRY_DIR := $(shell ls -d $(RUNS_DIR)/$**))
$(eval TRY_NUM := $*)
make cv_score TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM)
cv_score:
$(AVERAGE_SCORES) $(TRY_DIR)/cv*/T.*.*_gen.o* | tee $(TRY_DIR)/gen.log.txt
cv_train_stats-%:
$(eval TRY_DIR := $(shell ls -d $(RUNS_DIR)/$**))
$(eval TRY_NUM := $*)
make cv_train_stats TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM)
cv_train_stats:
$(AVERAGE_SCORES) $(TRY_DIR)/cv*/T.*.percrank_train.o* | tee $(TRY_DIR)/percrank_train.log.txt
percrank_train: RUN_NAME := percrank_train
percrank_train: prepare_dir
percrank_train:
# copy needed files (to ensure replication)
cp $(CANDGEN_MODEL) $(TRY_DIR)/candgen.pickle.gz
cp $(PERCRANK_CONFIG) $(TRY_DIR)/percrank_config.py
cp $(TRAIN_DAS) $(TRY_DIR)/train-das.txt
cp $(TRAIN_TREES) $(TRY_DIR)/train-text.pickle.gz
# run the experiment
$(QSUBMIT) '$(ACTIVATE); \
$(TGEN) percrank_train $(DEBUG_LOG_TRAIN) -c $(TRY_DIR)/candgen.pickle.gz \
-s $(TRAIN_PORTION) $(PARALLEL) -r "$(SEED)" \
$(TRY_DIR)/percrank_config.py $(TRY_DIR)/train-das.txt $(TRY_DIR)/train-text.pickle.gz \
$(TRY_DIR)/percrank.pickle.gz' 2>&1 | tee $(TRY_DIR)/percrank_train.log.txt
seq2seq_train: RUN_NAME := seq2seq_train
seq2seq_train: prepare_dir
seq2seq_train:
# copy needed files (to ensure replication)
cp $(SEQ2SEQ_CONFIG) $(TRY_DIR)/seq2seq_config.py
cp $(TRAIN_DAS) $(TRY_DIR)/train-das.txt
cp $(TRAIN_TREES) $(TRY_DIR)/train-text.pickle.gz
# run the experiment
$(QSUBMIT) '$(ACTIVATE); \
$(TGEN) seq2seq_train $(DEBUG_LOG_TRAIN) \
-s $(TRAIN_PORTION) $(PARALLEL) -r "$(SEED)" \
$(TRY_DIR)/seq2seq_config.py $(TRY_DIR)/train-das.txt $(TRY_DIR)/train-text.pickle.gz \
$(TRY_DIR)/seq2seq.pickle.gz' 2>&1 | tee $(TRY_DIR)/seq2seq_train.log.txt
asearch_gen: RUN_NAME := asearch_gen
asearch_gen: prepare_dir
asearch_gen:
cp $(CANDGEN_MODEL) $(TRY_DIR)/candgen.pickle.gz
cp $(PERCRANK_MODEL) $(TRY_DIR)/percrank.pickle.gz
cp $(ASEARCH_CONFIG) $(TRY_DIR)/asearch_config.py
make prepare_test_data TRY_DIR=$(TRY_DIR)
make asearch_gen_process TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM) QP="$(QP)" QHOLD="$(QHOLD)" CV_NUM=$(CV_NUM)
asearch_gen_cv:
asearch_gen_cv:
cp $(ASEARCH_CONFIG) $(TRY_DIR)/asearch_config.py
make prepare_test_data TRY_DIR=$(TRY_DIR)
make asearch_gen_process TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM) QP="$(QP)" QHOLD="$(QHOLD)" CV_NUM=$(CV_NUM)
asearch_gen-%:
$(eval TRY_DIR := $(shell ls -d $(RUNS_DIR)/$**))
$(eval TRY_NUM := $*)
cp $(ASEARCH_CONFIG) $(TRY_DIR)/asearch_config.py
make prepare_test_data TRY_DIR=$(TRY_DIR)
make asearch_gen_process TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM) QP="$(QP)" QHOLD="$(QHOLD)" CV_NUM=$(CV_NUM)
seq2seq_gen_cv:
make prepare_test_data TRY_DIR=$(TRY_DIR)
make seq2seq_gen_process TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM) QP="$(QP)" QHOLD="$(QHOLD)" CV_NUM=$(CV_NUM)
prepare_test_data:
cp $(TEST_DAS) $(TRY_DIR)/test-das.txt
cp $(TEST_REF) $(TRY_DIR)/test-das.sgm
cp $(TEST_TREES) $(TRY_DIR)/test-text.pickle.gz
cp $(TEST_CONC) $(TRY_DIR)/test-conc.sgm
cp $(TEST_ABSTR) $(TRY_DIR)/test-abstr.txt
cp $(GRAMMATEMES) $(TRY_DIR)/grammatemes.tsv
asearch_gen_process: RUN_NAME := asearch_gen
asearch_gen_process: GEN_MODEL := $(TRY_DIR)/candgen.pickle.gz $(TRY_DIR)/percrank.pickle.gz
asearch_gen_process: SELECTOR_CONFIG := -s "$(EVAL_SELECTOR)"
asearch_gen_process: GEN_CONFIG := -c $(TRY_DIR)/asearch_config.py
asearch_gen_process: gen_process
seq2seq_gen_process: RUN_NAME := seq2seq_gen
seq2seq_gen_process: GEN_MODEL := $(TRY_DIR)/seq2seq.pickle.gz
seq2seq_gen_process: SELECTOR_CONFIG := -r "$(EVAL_SELECTOR)" -t "$(TARGET_SELECTOR)"
seq2seq_gen_process: GEN_CONFIG :=
seq2seq_gen_process: gen_process
gen_process:
# run
$(QSUBMIT) '$(ACTIVATE); \
$(TGEN) $(RUN_NAME) -e $(TRY_DIR)/test-text.pickle.gz $(DEBUG_LOG_GEN) \
$(SELECTOR_CONFIG) \
-w $(TRY_DIR)/out-trees.yaml.gz $(GEN_CONFIG) \
$(GEN_MODEL) $(TRY_DIR)/test-das.txt; \
treex -Len -Sgen Read::YAML from=$(TRY_DIR)/out-trees.yaml.gz \
T2T::AssignDefaultGrammatemes grammateme_file=$(TRY_DIR)/grammatemes.tsv \
Misc::DeabstractDialogueSlots abstraction_file=$(TRY_DIR)/test-abstr.txt xs_instead=$(XS_INSTEAD) \
$(SURFACE_SCEN) Write::Treex \
Util::Eval document="$$.set_path(\"\"); $$.set_file_stem(\"test\");" \
Write::SgmMTEval to=$(TRY_DIR)/out-text.sgm set_id=BAGEL sys_id=TGEN add_header=tstset; \
$(MTEVAL) -s $(TRY_DIR)/test-das.sgm -r $(TRY_DIR)/test-conc.sgm -t $(TRY_DIR)/out-text.sgm' \
2>&1 | tee $(TRY_DIR)/$(RUN_NAME).log.txt
rerun-%:
$(eval TRY_DIR := $(shell ls -d $(RUNS_DIR)/$**))
$(eval TRY_NUM := $*)
$(eval RUN_NAME := $(shell ls -t $(TRY_DIR)/T.* | head -n 1 | sed 's/.*\/T\.[0-9]\+\.//;s/\..*//;'))
# select last log file, extract command from it and run it
if ls -d $(TRY_DIR)/cv* >/dev/null 2>&1; then \
for DIR in `find $(TRY_DIR)/cv* -type d`; do \
make rerun TRY_DIR=$$DIR TRY_NUM=$(TRY_NUM) RUN_NAME=percrank_train 2>&1 | tee -a $(TRY_DIR)/cv.log.txt ; \
JOB_NUM=`cat $(TRY_DIR)/cv.log.txt | grep '^Your job' | tail -n 1 | sed 's/^Your job \([0-9]\+\).*/\1/;' ` ; \
if `ls $$DIR/T.*.asearch_gen.* >/dev/null 2>&1`; then \
make rerun TRY_NUM=$(TRY_NUM) TRY_DIR=$$DIR RUN_NAME=asearch_gen QP="--hold $$JOB_NUM" ; \
else \
make asearch_gen_cv TRY_DIR=$$DIR TRY_NUM=$(TRY_NUM) CV_NUM=`basename $$DIR | sed 's/s.$$//'` QP="--hold $$JOB_NUM"; \
fi ; \
done ; \
else \
make rerun TRY_DIR=$(TRY_DIR) TRY_NUM=$(TRY_NUM) RUN_NAME=$(RUN_NAME) ; \
fi
# Use only through rerun-XXX (TRY_DIR, RUN_NAME must be set)
rerun:
LAST_LOGFILE=`ls -t $(TRY_DIR)/T.*.$(RUN_NAME).* | head -n 1` ; \
COMMAND=`cat $$LAST_LOGFILE | sed '1,/^== Directory/d;/^== Hard res/,$$d;s/^== Command:\s*//;'` ; \
echo "Rerunning command: $$COMMAND"; \
$(QSUBMIT) "$$COMMAND" 2>&1 | tee -a $(TRY_DIR)/$(RUN_NAME).log.txt ;
norm_log:
if [[ -z "$(LOGFILE)" ]]; then \
echo "LOGFILE must be set." ; \
exit 1; \
fi; \
cat $(LOGFILE) | head -n $(LOG_CUTOFF) | sed -r 's/^[A-Za-z]{3} [A-Za-z]{3} [0-9: ]{16} //' > $(LOGFILE).norm ; \
candgen_train:
cp $(TRAIN_DAS) $(TRY_DIR)/train-das.txt
cp $(TRAIN_TREES) $(TRY_DIR)/train-text.pickle.gz
echo -e "Training set: $(TRAINING_SET)\nParams: $(CANDGEN_PARAMS)" > $(TRY_DIR)/PARAMS
PARAMS="$(CANDGEN_PARAMS)" ; \
if [[ -n `echo $$PARAMS | grep -- '-t'` ]]; then \
CL_CONFIG=`echo "$$PARAMS" | sed 's/.*-t //'`; \
PARAMS=`echo "$$PARAMS" | sed 's/-t .*/-t /'`"$$CL_CONFIG" ; \
cp $$CL_CONFIG $(TRY_DIR)/classif_config.py; \
fi; \
../run_tgen.py candgen_train $$PARAMS $(TRY_DIR)/train-das.txt $(TRY_DIR)/train-text.pickle.gz $(TRY_DIR)/candgen.pickle.gz 2>&1 | tee $(TRY_DIR)/training.log
candgen_train_cv:
for CV in cv0{0,1,2,3,4,5,6,7,8,9}; do \
mkdir -p model/$(TRAINING_SET)/$$CV; \
make candgen_train TRY_DIR="model/$(TRAINING_SET)/$$CV" CANDGEN_PARAMS="$(CANDGEN_PARAMS)" TRAINING_SET="$(TRAINING_SET)" CV_NUM="$$CV"; \
done;
compare_output:
cat $(TRY_DIR)/test-conc.sgm | sed 's/<\/doc>/***/;s/<[^>]*>//g;s/ \././g;/^$$/d;4,24s/^/* /;28,$$s/^/+ /;' | awk '{print tolower($$0)}' > $(TRY_DIR)/test-conc.txt
cat $(TRY_DIR)/out-text.sgm | sed 's/<\/doc>/***/;s/<[^>]*>//g;s/ \././g;/^$$/d;4,24s/^/* /;' | awk '{print tolower($$0)}' > $(TRY_DIR)/out-text.txt
vimdiff $(TRY_DIR)/test-conc.txt $(TRY_DIR)/out-text.txt