bert_config.json 485 B

123456789101112131415161718192021
  1. {
  2. "architectures": [
  3. "BertModel"
  4. ],
  5. "attention_probs_dropout_prob": 0.1,
  6. "hidden_act": "gelu",
  7. "hidden_dropout_prob": 0.1,
  8. "hidden_size": 768,
  9. "initializer_range": 0.02,
  10. "intermediate_size": 3072,
  11. "layer_norm_eps": 1e-12,
  12. "max_position_embeddings": 512,
  13. "model_type": "bert",
  14. "num_attention_heads": 12,
  15. "num_hidden_layers": 12,
  16. "pad_token_id": 0,
  17. "type_vocab_size": 2,
  18. "vocab_size": 30522,
  19. "encoder_width": 768,
  20. "add_cross_attention": true
  21. }