enh_inference_streaming.py
Less than 1 minute
enh_inference_streaming.py
Frontend inference
usage: enh_inference_streaming.py [-h] [--config CONFIG]
[--log_level {CRITICAL,ERROR,WARNING,INFO,DEBUG,NOTSET}]
--output_dir OUTPUT_DIR [--ngpu NGPU] [--seed SEED]
[--dtype {float16,float32,float64}] [--fs FS]
[--num_workers NUM_WORKERS] --data_path_and_name_and_type
DATA_PATH_AND_NAME_AND_TYPE [--key_file KEY_FILE]
[--allow_variable_data_keys ALLOW_VARIABLE_DATA_KEYS]
[--output_format OUTPUT_FORMAT] [--train_config TRAIN_CONFIG]
[--model_file MODEL_FILE] [--model_tag MODEL_TAG]
[--inference_config INFERENCE_CONFIG]
[--enh_s2t_task ENH_S2T_TASK] [--batch_size BATCH_SIZE]
[--ref_channel REF_CHANNEL]