Index
Symbols
VisualEncoder
Transformer model, training with 264
A
abstractive summaries
Adaptive Moment Estimation (Adam Optimizer) 119
Attention mechanism 123
Audio-Visual Speech Recognition (AVSR) 228
B
Bahdanau Attention 126
Bahdanau attention layer 197, 198, 199
Batch Normalization (BatchNorm) 245
used, for decoding penalties 218, 219, 220
used for improving text summarization 214, 216, 217
BERT-based transfer learning 123
encoder-decoder networks 123, 124
BERT fine-tuning approach
for SQuAD question answering 341, 342
bidirectional encoder representations from transformers (BERT) model 132, 133
about 131
custom layers, building 142, 143, 144, 145, 146, 147
normalization 133, 134, 135, 136, 137, 138, 139
sequences 135
tokenization 133, 134, 135, 136, 137, 138, 139
Bi-directional...