speechbrainteam commited on
Commit
598163b
·
1 Parent(s): ace1f98

Update hyperparams.yaml

Browse files
Files changed (1) hide show
  1. hyperparams.yaml +0 -106
hyperparams.yaml CHANGED
@@ -10,89 +10,6 @@
10
  # Basic parameters
11
  # Seed needs to be set at top of yaml, before objects with parameters are made
12
  #
13
- seed: 1234
14
- __set_seed: !apply:torch.manual_seed [1234]
15
-
16
- # Data params
17
-
18
- # the data folder for the wham dataset
19
- # data_folder needs to follow the format: /yourpath/whamr.
20
- # make sure to use the name whamr at your top folder for the dataset!
21
- data_folder: /network/tmp1/subakany/whamr
22
- task: enhancement
23
- dereverberate: false
24
-
25
- # the path for wsj0/si_tr_s/ folder -- only needed if dynamic mixing is used
26
- # e.g. /yourpath/wsj0-processed/si_tr_s/
27
- # you need to convert the original wsj0 to 8k
28
- # you can do this conversion with the script ../meta/preprocess_dynamic_mixing.py
29
- base_folder_dm: /network/tmp1/subakany/wsj0-processed/si_tr_s/
30
-
31
- experiment_name: sepformer-whamr-enhancement-DM
32
- output_folder: results/sepformer-whamr-enhancement-DM/1234
33
- train_log: results/sepformer-whamr-enhancement-DM/1234/train_log.txt
34
- save_folder: results/sepformer-whamr-enhancement-DM/1234/save
35
-
36
- # the file names should start with whamr instead of whamorg
37
- train_data: results/sepformer-whamr-enhancement-DM/1234/save/whamr_tr.csv
38
- valid_data: results/sepformer-whamr-enhancement-DM/1234/save/whamr_cv.csv
39
- test_data: results/sepformer-whamr-enhancement-DM/1234/save/whamr_tt.csv
40
- skip_prep: false
41
-
42
- # Experiment params
43
- auto_mix_prec: true # Set it to True for mixed precision
44
- test_only: false
45
- num_spks: 1 # set to 3 for wsj0-3mix
46
- progressbar: true
47
- save_audio: true # Save estimated sources on disk
48
- sample_rate: 8000
49
- n_audio_to_save: 20
50
-
51
- # Training parameters
52
- N_epochs: 200
53
- batch_size: 1
54
- lr: 0.00015
55
- clip_grad_norm: 5
56
- loss_upper_lim: 999999 # this is the upper limit for an acceptable loss
57
- # if True, the training sequences are cut to a specified length
58
- limit_training_signal_len: false
59
- # this is the length of sequences if we choose to limit
60
- # the signal length of training sequences
61
- training_signal_len: 32000000
62
-
63
- # Set it to True to dynamically create mixtures at training time
64
- dynamic_mixing: true
65
-
66
- # Parameters for data augmentation
67
-
68
- # rir_path variable points to the directory of the room impulse responses
69
- # e.g. /miniscratch/subakany/rir_wavs
70
- # If the path does not exist, it is created automatically.
71
- rir_path: /network/scratch/s/subakany/whamr_rirs_wavs_8k/
72
-
73
- use_wavedrop: false
74
- use_speedperturb: true
75
- use_speedperturb_sameforeachsource: false
76
- use_rand_shift: false
77
- min_shift: -8000
78
- max_shift: 8000
79
-
80
- speedperturb: !new:speechbrain.lobes.augment.TimeDomainSpecAugment
81
- perturb_prob: 1.0
82
- drop_freq_prob: 0.0
83
- drop_chunk_prob: 0.0
84
- sample_rate: 8000
85
- speeds: [95, 100, 105]
86
-
87
- wavedrop: !new:speechbrain.lobes.augment.TimeDomainSpecAugment
88
- perturb_prob: 0.0
89
- drop_freq_prob: 1.0
90
- drop_chunk_prob: 1.0
91
- sample_rate: 8000
92
-
93
- # loss thresholding -- this thresholds the training loss
94
- threshold_byloss: true
95
- threshold: -30
96
 
97
  # Encoder parameters
98
  N_encoder_out: 256
@@ -100,11 +17,6 @@ out_channels: 256
100
  kernel_size: 16
101
  kernel_stride: 8
102
 
103
- # Dataloader options
104
- dataloader_opts:
105
- batch_size: 1
106
- num_workers: 3
107
-
108
  # Specifying the network
109
  Encoder: &id003 !new:speechbrain.lobes.models.dual_path.Encoder
110
  kernel_size: 16
@@ -149,20 +61,6 @@ Decoder: &id004 !new:speechbrain.lobes.models.dual_path.Decoder
149
  stride: 8
150
  bias: false
151
 
152
- optimizer: !name:torch.optim.Adam
153
- lr: 0.00015
154
- weight_decay: 0
155
-
156
- loss: !name:speechbrain.nnet.losses.get_si_snr_with_pitwrapper
157
-
158
- lr_scheduler: &id007 !new:speechbrain.nnet.schedulers.ReduceLROnPlateau
159
-
160
- factor: 0.5
161
- patience: 2
162
- dont_halve_until_epoch: 85
163
-
164
- epoch_counter: &id006 !new:speechbrain.utils.epoch_loop.EpochCounter
165
- limit: 200
166
 
167
  modules:
168
  encoder: *id003
@@ -175,10 +73,6 @@ checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
175
  encoder: *id003
176
  decoder: *id004
177
  masknet: *id005
178
- counter: *id006
179
- lr_scheduler: *id007
180
- train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
181
- save_file: results/sepformer-whamr-enhancement-DM/1234/train_log.txt
182
 
183
  pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
184
  loadables:
 
10
  # Basic parameters
11
  # Seed needs to be set at top of yaml, before objects with parameters are made
12
  #
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  # Encoder parameters
15
  N_encoder_out: 256
 
17
  kernel_size: 16
18
  kernel_stride: 8
19
 
 
 
 
 
 
20
  # Specifying the network
21
  Encoder: &id003 !new:speechbrain.lobes.models.dual_path.Encoder
22
  kernel_size: 16
 
61
  stride: 8
62
  bias: false
63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
64
 
65
  modules:
66
  encoder: *id003
 
73
  encoder: *id003
74
  decoder: *id004
75
  masknet: *id005
 
 
 
 
76
 
77
  pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
78
  loadables: