superb_asv#
(s3prl.problem.asv.superb_asv)
The setting of Superb ASV
- Authors
Po-Han Chi 2021
Leo 2021
Haibin Wu 2022
Leo 2022
SuperbASV#
- class s3prl.problem.asv.superb_asv.SuperbASV[source][source]#
Bases:
ASV
- default_config()[source][source]#
The default arguments for
run
in yaml. Note that for the fields with inner values, likebuild_model
, the outer field name corresponds to a method name, so you can find the methodbuild_model
. Furthermore, the values inside that field will be directly passed into the method. So by changing these inner values, you can directly affect the behavior of the corresponding method. See the method documentation for all the supported arguments and their meanings.The methods affected by the following config are:
prepare_data
build_dataset
build_batch_sampler
build_upstream
build_featurizer
build_model
build_task
build_optimizer
build_scheduler
train
target_dir: ??? cache_dir: null test_ckpt_steps: null prepare_data: dataset_root: ??? build_dataset: train: min_secs: 2.0 max_secs: 8.0 build_batch_sampler: train: batch_size: 10 shuffle: true test: batch_size: 1 build_upstream: name: ??? build_featurizer: layer_selections: null normalize: false build_model: upstream_trainable: false build_task: loss_type: amsoftmax loss_conf: margin: 0.4 scale: 30 build_optimizer: name: AdamW conf: lr: 0.0001 build_scheduler: name: ExponentialLR gamma: 0.9 train: total_steps: 200000 log_step: 500 eval_step: 1.0e+20 save_step: 10000 gradient_clipping: 1000.0 gradient_accumulate: 5 valid_metric: null valid_higher_better: null auto_resume: true resume_ckpt_dir: null keep_num_ckpts: null
- prepare_data(prepare_data: dict, target_dir: str, cache_dir: str, get_path_only: bool)[source][source]#
Prepare the task-specific data metadata (path, labels…). By default call
prepare_voxceleb1_for_sv
with**prepare_data
- Parameters:
prepare_data (dict) – same in
default_config
, support arguments inprepare_voxceleb1_for_sv
target_dir (str) – Parse your corpus and save the csv file into this directory
cache_dir (str) – If the parsing or preprocessing takes too long time, you can save the temporary files into this directory. This directory is expected to be shared across different training sessions (different hypers and
target_dir
)get_path_only (bool) – Directly return the filepaths no matter they exist or not.
- Returns:
tuple
train_path (str)
test_trial_paths (List[str])
The
train_path
should be a csv file containing the following columns:column
description
id
(str) - the unique id for this utterance
wav_path
(str) - the absolute path of the waveform file
spk
(str) - a string speaker label
Each
test_trial_path
should be a csv file containing the following columns:column
description
id1
(str) - the unique id of the first utterance
id2
(str) - the unique id of the second utterance
wav_path1
(str) - the absolute path of the first utterance
wav_path2
(str) - the absolute path of the second utterance
label
(int) - 0 when two utterances are from different speakers, 1 when same speaker
- build_encoder(build_encoder: dict, target_dir: str, cache_dir: str, train_csv: str, test_csvs: list, get_path_only: bool)[source][source]#
Build the encoder (for the labels) given the data metadata, and return the saved encoder path. By default generate and save a
s3prl.dataio.encoder.CategoryEncoder
from thelabel
column of the train csv.- Parameters:
build_encoder (dict) – same in
default_config
, no argument supported for nowtarget_dir (str) – Save your encoder into this directory
cache_dir (str) – If the preprocessing takes too long time, you can save the temporary files into this directory. This directory is expected to be shared across different training sessions (different hypers and
target_dir
)train_csv_path (str) – the train path from
prepare_data
valid_csv_path (str) – the valid path from
prepare_data
test_csv_paths (List[str]) – the test paths from
prepare_data
get_path_only (bool) – Directly return the filepaths no matter they exist or not
- Returns:
str
encoder_path: The encoder should be saved in the pickle format
- build_dataset(build_dataset: dict, target_dir: str, cache_dir: str, mode: str, data_csv: str, encoder_path: str)[source][source]#
Build the dataset for train/valid/test.
- Parameters:
build_dataset (dict) –
same in
default_config
, havetrain
andtest
keys, each is a dictionary, fortrain
dictionary:key
description
min_secs
(float) - Drop a waveform if it is not longer than
min_secs
max_secs
(float) - If a waveform is longer than
max_secs
seconds, randomly crop the waveform intomax_secs
seconds. Default: None, no croppingfor
test
dictionary, no argument supported yettarget_dir (str) – Current experiment directory
cache_dir (str) – If the preprocessing takes too long time, you can save the temporary files into this directory. This directory is expected to be shared across different training sessions (different hypers and
target_dir
)mode (str) – train/valid/test
data_csv (str) – The metadata csv file for the specific
mode
encoder_path (str) – The pickled encoder path for encoding the labels
- Returns:
torch Dataset
For train mode, the dataset should return each item as a dictionary containing the following keys:
key
description
x
(torch.FloatTensor) - the waveform in (seq_len, 1)
x_len
(int) - the waveform length
seq_len
class_id
(str) - the label class id encoded by
encoder_path
unique_name
(str) - the unique id for this datapoint
For test mode:
x (torch.FloatTensor) - the waveform in (seq_len, 1) x_len (int) - the waveform length
seq_len
unique_name (str) - the unique id for this datapoint
- build_batch_sampler(build_batch_sampler: dict, target_dir: str, cache_dir: str, mode: str, data_csv: str, dataset)[source][source]#
Return the batch sampler for torch DataLoader.
- Parameters:
build_batch_sampler (dict) –
same in
default_config
key
description
train
(dict) - arguments for
FixedBatchSizeBatchSampler
test
(dict) - arguments for
FixedBatchSizeBatchSampler
Note that ASV does not support valid
target_dir (str) – Current experiment directory
cache_dir (str) – If the preprocessing takes too long time, save the temporary files into this directory. This directory is expected to be shared across different training sessions (different hypers and
target_dir
)mode (str) – train/valid/test
data_csv (str) – the
mode
specific csv fromprepare_data
dataset – the dataset from
build_dataset
- Returns:
batch sampler for torch DataLoader
- build_downstream(build_downstream: dict, downstream_input_size: int, downstream_output_size: int, downstream_input_stride: int)[source][source]#
Return the task-specific downstream model. By default build the
SuperbXvector
model- Parameters:
build_downstream (dict) – same in
default_config
, support arguments ofSuperbXvector
downstream_input_size (int) – the required input size of the model
downstream_output_size (int) – the required output size of the model
downstream_input_stride (int) – the input feature’s stride (from 16 KHz)
- Returns:
- build_collate_fn(build_collate_fn: dict, mode: str)[source]#
By default returns
s3prl.dataset.base.default_collate_fn
- Parameters:
build_collate_fn (dict) – same in
default_config
, no argument supported for nowmode (str) – train, valid, or test
- Returns:
callable
the collate_fn for torch DataLoader in train/valid/test
mode
- build_featurizer(build_featurizer: dict, upstream)[source]#
By default build the featurizer with
s3prl.nn.Featurizer
- Parameters:
build_featurizer (dict) – same in
default_config
, arguments fors3prl.nn.Featurizer
upstream (
AbsUpstream
) – the upstream model built bybuild_upstream
- Returns:
s3prl.nn.interface.AbsFeaturizer
Return the featurizer model. The featurizer is used to reduce the multiple hidden states returned from the upstream model (built by
build_upstream
) into a single hidden state, so can be easliy fed into the downstream model
- build_model(build_model: dict, model_output_size: int, build_upstream: dict, build_featurizer: dict, build_downstream: dict)[source]#
By default build model with
s3prl.nn.upstream.UpstreamDownstreamModel
- Parameters:
build_model (dict) – same in
default_config
, arguments fors3prl.nn.upstream.UpstreamDownstreamModel
model_output_size (int) – the required model’s output hidden size
build_upstream (dict) – same in
default_config
, refer tobuild_upstream
build_featurizer (dict) – same in
default_config
, refer tobuild_featurizer
build_downstream (dict) – same in
default_config
, refer tobuild_downstream
- Returns:
torch.nn.Module
Return the entire model for the task, which takes the direct items from DataLoader as the input. Usually, the components can be built by
build_upstream
,build_featurizer
,build_downstream
, and are concated together to get the final model. The upstream extracts multiple hidden states, the featuizer reduce them into a single hidden state, and the downstream takes the hidden states as the feature for the downstream-specific model.
- build_optimizer(build_optimizer: dict, parameters)[source]#
- Parameters:
build_optimizer (dict) –
same in
default_config
, refer to belowkey
description
name
(str) - the optimizer class name in
torch.optim
conf
(dict) - the arguments for initializing the optimizer class. e.g.
{"lr": 1.0e-4}
parameters (iterable) – the standard params accepted by
torch.optim.Optimizer
.
- Returns:
torch.optim.Optimizer
An optimizer following standard torch usage
- build_scheduler(build_scheduler: dict, optimizer)[source]#
- Parameters:
build_scheduler (dict) –
same in
default_config
key
description
name
(str) - the scheduler class name in
torch.optim.lr_scheduler
conf
(dict) - the arguments for initializing the scheduler class. e.g.
{"gamma": 0.01}
fortorch.optim.lr_scheduler.StepLR
optimizer – the standard torch optimizer accepted by Scheduler in
torch.optim.lr_scheduler
.
- Returns:
torch scheduler
A scheduler following standard torch usage
- build_task(build_task: dict, model, encoder, test_trials=None)[source]#
Build the task, which defines the logics for every train/valid/test forward step for the
model
, and the logics for how to reduce all the batch results from multiple train/valid/test steps into metricsBy default build
SpeakerVerification
- Parameters:
build_task (dict) – same in
default_config
, no argument supported for nowmodel (torch.nn.Module) – the model built by
build_model
encoder – the encoder built by
build_encoder
test_trials (List[Tuple[int, str, str]]) – each tuple in the list consists of
(label, enroll_utt_id, test_utt_id)
. label is either 0 or 1
- Returns:
Task
- build_upstream(build_upstream: dict)[source]#
By default build the upstream with
s3prl.nn.upstream.S3PRLUpstream
- Parameters:
build_upstream (dict) – same in
default_config
, arguments fors3prl.nn.upstream.S3PRLUpstream
- Returns:
s3prl.nn.interface.AbsUpstream
Return an upstream model, whose forward takes the waveform input and returns multiple hidden states as features.
- evaluate(evaluate: dict, mode: str, task, dataset, batch_sampler, collate_fn, eval_batch: int, dump_dir: str, device: str, num_workers: int)[source]#
The evaluate routine used by
train
(during validation phase) andrun
(during testing phase).- Parameters:
evaluate (dict) – same in
default_config
, no argument supported for now**others – only meaningful when you want to override this train method, which is not the common case. Hence we skip the documentation for now.
- classmethod get_class_from_name(name: str)[source]#
- Parameters:
name (str) – the
__name__
of the problem class- Returns:
Problem
- load_model(model_ckpt_dir: str)[source]#
Return the saved model.
- Parameters:
model_ckpt_dir (str) – Restore the model with
build_model
and the checkpoint saved in this directory.- Returns:
torch.nn.Module
- load_model_and_task(ckpts_dir: str, task_overrides: Optional[dict] = None)[source]#
This is a helper method to combine
load_model
andload_task
together to directly load the model and the task. This method assumes the model is saved underckpts_dir / 'model'
and the task is saved underckpts_dir / 'task'
- Returns:
tuple
model (
torch.nn.Module
)task (
s3prl.task.Task
)
- load_task(task_ckpt_dir: str, model: Module, task_overrides: Optional[dict] = None)[source]#
Return the saved task.
- Parameters:
task_ckpt_dir (str) – Restore the task with
build_task
and the checkpoint saved in this directory.model (torch.nn.Module) – the model for the task, since the model is separately saved and is required for
build_task
.task_overrides (dict) – overrides the saved initialization arguments, so can change the loaded task’s behavior. Like, change the decoding hyperparameters.
- Returns:
- run(target_dir: str, cache_dir: str, remove_all_cache: bool = False, start: int = 0, stop: Optional[int] = None, num_workers: int = 6, eval_batch: int = -1, device: str = 'cuda', world_size: int = 1, rank: int = 0, test_ckpt_dir: Optional[str] = None, test_ckpt_steps: Optional[List[int]] = None, prepare_data: Optional[dict] = None, build_encoder: Optional[dict] = None, build_dataset: Optional[dict] = None, build_batch_sampler: Optional[dict] = None, build_collate_fn: Optional[dict] = None, build_upstream: Optional[dict] = None, build_featurizer: Optional[dict] = None, build_downstream: Optional[dict] = None, build_model: Optional[dict] = None, build_task: Optional[dict] = None, build_optimizer: Optional[dict] = None, build_scheduler: Optional[dict] = None, save_model: Optional[dict] = None, save_task: Optional[dict] = None, train: Optional[dict] = None, evaluate: Optional[dict] = None)[source]#
stage
description
0
Parse the corpus and save the metadata file (waveform path, label…)
1
Build the encoder for encoding the speaker labels
2
Train the model
3
Evaluate the model on multiple test sets, multiple checkpoints will be evaluated for each test set (See
test_ckpt_steps
)4
Report the best result find on each test set
- Parameters:
target_dir (str) – The directory that stores the script result.
cache_dir (str) – The directory that caches the processed data. Default: /home/user/.cache/s3prl/data
remove_all_cache (bool) – Whether to remove all the cache stored under cache_dir. Default: False
start (int) – The starting stage of the problem script. Default: 0
stop (int) – The stoping stage of the problem script, set None to reach the final stage. Default: None
num_workers (int) – num_workers for all the torch DataLoder
eval_batch (int) – During evaluation (valid or test), limit the number of batch. This is helpful for the fast development to check everything won’t crash. If is -1, disable this feature and evaluate the entire epoch. Default: -1
device (str) – The device type for all torch-related operation: “cpu” or “cuda” Default: “cuda”
world_size (int) – How many processes are running this script simultaneously (in parallel). Usually this is just 1, however if you are runnig distributed training, this should be > 1. Default: 1
rank (int) – When distributed training, world_size > 1. Take
world_size == 8
for example, this means 8 processes (8 GPUs) are runing in parallel. The script needs to know which process among 8 processes it is. In this case,rank
can range from 0~7. All the 8 processes have the sameworld_size
but differentrank
(process id).test_ckpt_dir (str) – Specify the checkpoint path for testing. If not, use checkpoints specified by
test_ckpts_steps
.test_ckpt_steps (List[int]) – After training, multiple steps of checkpoints are saved. This option specifies which checkpoints (multiple) will be used for evaluation.
**kwds – The other arguments like
prepare_data
andbuild_model
are method specific-arguments for methods likeprepare_data
andbuild_model
, and will not be used in the corerun
logic. See the specific method documentation for their supported arguments and meaning
- save_model(save_model: dict, model_ckpt_dir: str, build_model_all_args: dict, model: Module)[source]#
Save the model state_dict and the model initialization arguments into the given directory. If you override this method, it is highly possible you also need to override
load_model
- Parameters:
save_model (dict) – same in
default_config
, so the user can save additional settings, like the configuration of the dataset by duplicating the dataset hypers inside thesave_model
field. You can rely on theomegaconf
package to simplify the duplication.model_ckpt_dir (str) – save the model into the this directory.
build_model_all_args (dict) – all the arguments of
build_model
. By saving this dictionary, you can easily reconstruct the same model by callingbuild_model
with the saved dictionary.model (torch.nn.Module) – the model to be saved.
- Returns:
None
- save_task(save_task: dict, task_ckpt_dir: str, build_task_all_args_except_model: dict, task: Task)[source]#
Save the task’s state,
task.get_state()
, and the initialization arguments into the given directory. If you override this method, it is highly possible you also need to overrideload_task
.- Parameters:
save_task (dict) – same in
default_config
, so the user can save additional settings, like the configuration of the dataset by duplicating the dataset hypers inside thesave_task
field. You can rely on theomegaconf
package to simplify the duplication.task_ckpt_dir (str) – save the task into this directory.
build_task_all_args_except_model (dict) – all the arguments of
build_task
except themodel
argument since the model should be sapartely saved bysave_model
. By saving this dictionary, you can easily reconstruct the same task by callingbuild_task
with the saved dictionary.task (Task) – the task to be saved.
- Returns:
None
- train(train: dict, train_dir: str, build_model_all_args: dict, build_task_all_args_except_model: dict, save_model: dict, save_task: dict, build_optimizer: dict, build_scheduler: dict, evaluate: dict, train_dataset, train_batch_sampler, train_collate_fn, valid_dataset, valid_batch_sampler, valid_collate_fn, num_workers: int, world_size: int, rank: int, eval_batch: int, device: str, global_config: Optional[dict] = None)[source]#
- Parameters:
train (dict) –
same in
default_config
key
description
total_steps
(int) - the total optimization steps
log_step
(int) - logging frequency. log every
log_step
stepeval_step
(int) - evaluation frequency. Evaluate every
eval_step
step. Note that you can control how many batch to evaluate to speed up the development by theeval_batch
argument inrun
save_step
(int) - save the checkpoint every
save_step
step.gradient_clipping
(float) - clip the gradient. important for RNNs.
gradient_accumulate
(int) - accumulate multiple steps’ gradient before updating network parameters to simulate large-batch optimization.
valid_metric
(str) - the metric to select the best valid checkpoint. Different Tasks have different supported valid_metrics. See
build_task
for the supported metrics.valid_higher_better
(bool) - some metrics are higher better, while some are lower better this will affect how to save the best validation checkpoint.
auto_resume
(bool) - if there are already the last checkpoint in
target_dir
(seerun
), whether to resume from it or delete it and start a new training session.resume_ckpt_dir
(str) - you can directly specify the checkpoint path to resume which is not necessary in
target_dir
(seerun
).seed
(int) - fix the seed before the training start
keep_num_ckpts
(int) - to prevent saving too many checkpoints, only save the
keep_num_ckpts
latest checkpoints and delete the old ones.use_scheduler
(bool) - whether to use the scheduler
**others – only meaningful when you want to override this train method, which is not the common case. Hence we skip the documentation for now.
prepare_voxceleb1_for_sv#
- s3prl.problem.asv.superb_asv.prepare_voxceleb1_for_sv(target_dir: str, cache_dir: str, get_path_only: str, dataset_root: str, force_download: bool = False)[source][source]#
Prepare VoxCeleb1 for speaker verification following
SuperbASV.prepare_data
format.- Parameters:
dataset_root (str) – The root path of Fluent Speech Command
force_download (bool) – always re-download the metadata for VoxCeleb1