|
|
|
name: &name BertIR |
|
trainer: |
|
devices: 1 |
|
num_nodes: 1 |
|
max_epochs: 2 |
|
max_steps: -1 |
|
accumulate_grad_batches: 1 |
|
precision: 16 |
|
accelerator: gpu |
|
strategy: ddp |
|
log_every_n_steps: 1 |
|
val_check_interval: 0.05 |
|
enable_checkpointing: False |
|
logger: false |
|
|
|
model: |
|
nemo_path: null |
|
|
|
language_model: |
|
pretrained_model_name: bert-base-uncased |
|
sim_score_dropout: 0.1 |
|
lm_checkpoint: null |
|
config: |
|
attention_probs_dropout_prob: 0.1 |
|
hidden_act: gelu |
|
hidden_dropout_prob: 0.1 |
|
hidden_size: 768 |
|
initializer_range: 0.02 |
|
intermediate_size: 3072 |
|
max_position_embeddings: 512 |
|
num_attention_heads: 12 |
|
num_hidden_layers: 12 |
|
type_vocab_size: 2 |
|
vocab_size: 30522 |
|
config_file: null |
|
|
|
tokenizer: |
|
tokenizer_name: ${model.language_model.pretrained_model_name} |
|
vocab_file: null |
|
tokenizer_model: null |
|
special_tokens: null |
|
|
|
train_ds: |
|
passages: null |
|
queries: null |
|
query_to_passages: null |
|
|
|
|
|
num_negatives: 10 |
|
batch_size: 6 |
|
psg_cache_format: npz |
|
shuffle: true |
|
num_samples: -1 |
|
num_workers: 1 |
|
drop_last: false |
|
pin_memory: false |
|
|
|
validation_ds: |
|
passages: null |
|
queries: null |
|
query_to_passages: null |
|
num_negatives: 10 |
|
batch_size: 6 |
|
psg_cache_format: pkl |
|
shuffle: false |
|
num_samples: -1 |
|
num_workers: 1 |
|
drop_last: false |
|
pin_memory: false |
|
|
|
optim: |
|
name: adam |
|
lr: 1e-5 |
|
betas: [0.9, 0.999] |
|
weight_decay: 0 |
|
|
|
sched: |
|
name: WarmupAnnealing |
|
warmup_steps: null |
|
warmup_ratio: 0.05 |
|
last_epoch: -1 |
|
|
|
|
|
monitor: val_loss |
|
reduce_on_plateau: false |
|
|
|
exp_manager: |
|
exp_dir: null |
|
name: *name |
|
create_tensorboard_logger: True |
|
create_checkpoint_callback: True |
|
|
|
hydra: |
|
run: |
|
dir: . |
|
job_logging: |
|
root: |
|
handlers: null |
|
|