From 49b2a2e00d7fbec9f970aea0cd1befa4c1526b1d Mon Sep 17 00:00:00 2001 From: dpressel Date: Thu, 7 May 2020 01:26:03 -0400 Subject: [PATCH] remove for now --- mead/config/ptb-kim-adam.yml | 54 --------------------------- mead/config/ptb-transformer-chars.yml | 46 ----------------------- 2 files changed, 100 deletions(-) delete mode 100644 mead/config/ptb-kim-adam.yml delete mode 100644 mead/config/ptb-transformer-chars.yml diff --git a/mead/config/ptb-kim-adam.yml b/mead/config/ptb-kim-adam.yml deleted file mode 100644 index 2a1f40fb5..000000000 --- a/mead/config/ptb-kim-adam.yml +++ /dev/null @@ -1,54 +0,0 @@ -basedir: ptb-med -task: lm -backend: tensorflow -dataset: ptb -unif: 0.05 -preproc: - mxwlen: 50 -features: -- name: x - vectorizer: - type: token1d - embeddings: - dsz: 512 - -- name: xch - vectorizer: - type: char2d - mxwlen: 50 - embeddings: - dsz: 16 - type: "char-conv" - filtsz: - - [1, 32] - - [2, 32] - - [3, 64] - - [4, 128] - - [5, 256] - - [6, 512] - - [7, 1024] - gating: highway - num_gates: 2 - cmotsz: 30 - -reader: - reader_type: default - tgt_key: x - src_keys: xch - nbptt: 35 - -model: - model_type: default - hsz: 512 - skip_conn: 1 - projsz: 512 - layers: 1 - pdrop: 0.5 - -train: - epochs: 25 - optim: adam - eta: 0.001 - nsteps: 100 - clip: 1.0 - batchsz: 20 diff --git a/mead/config/ptb-transformer-chars.yml b/mead/config/ptb-transformer-chars.yml deleted file mode 100644 index 2ee2689b5..000000000 --- a/mead/config/ptb-transformer-chars.yml +++ /dev/null @@ -1,46 +0,0 @@ -task: lm -basedir: ptb-transformer-chars -batchsz: 10 -unif: 0.1 -nbptt: 120 -preproc: {} -backend: pytorch -dataset: ptb - -loader: - reader_type: default - src_keys: [word_chars] - tgt_key: word - -features: - - name: word - vectorizer: - type: token1d - fields: text - embeddings: - dsz: 300 - type: default - - name: word_chars - vectorizer: - type: char2d - fields: text - embeddings: - dsz: 300 - wsz: 300 - #type: learned-positional-char-conv - type: positional-char-conv - -model: - model_type: transformer - hsz: 300 - layers: 2 - pdrop: 0.65 - -train: - epochs: 40 - optim: adam - #weight_decay: 1.0e-8 - eta: 4.0e-4 - do_early_stopping: true - clip: 0.5 -