enh_inference_streaming.py
Less than 1 minute
enh_inference_streaming.py
Frontend inference
usage: enh_inference_streaming.py [-h] [--config CONFIG] [--log_level {CRITICAL,ERROR,WARNING,INFO,DEBUG,NOTSET}] --output_dir OUTPUT_DIR [--ngpu NGPU] [--seed SEED]
[--dtype {float16,float32,float64}] [--fs FS] [--num_workers NUM_WORKERS] --data_path_and_name_and_type DATA_PATH_AND_NAME_AND_TYPE [--key_file KEY_FILE]
[--allow_variable_data_keys ALLOW_VARIABLE_DATA_KEYS] [--output_format OUTPUT_FORMAT] [--train_config TRAIN_CONFIG] [--model_file MODEL_FILE]
[--model_tag MODEL_TAG] [--inference_config INFERENCE_CONFIG] [--enh_s2t_task ENH_S2T_TASK] [--batch_size BATCH_SIZE] [--ref_channel REF_CHANNEL]