{ "metadata": { "total_size": 11338446336 }, "weight_map": { "bert.embeddings.word_embeddings.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.0.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.1.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.10.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.11.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.12.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.13.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.14.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.15.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.15.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.15.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.16.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.16.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.17.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.18.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.19.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.2.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.2.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.20.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.20.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.21.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.22.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.23.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.24.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.25.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.26.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.27.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.28.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.29.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.3.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.3.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.30.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.30.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.output.dense.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.output.dense.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.key.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.key.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.query.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.query.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.value.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.attention.self.value.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.ln.bias": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.ln.weight": "pytorch_model-00002-of-00003.bin", "bert.encoder.layer.31.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.31.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.31.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.31.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.output.dense.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.output.dense.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.key.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.key.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.query.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.query.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.value.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.attention.self.value.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.32.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.output.dense.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.output.dense.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.key.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.key.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.query.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.query.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.value.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.attention.self.value.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.33.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.output.dense.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.output.dense.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.key.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.key.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.query.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.query.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.value.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.attention.self.value.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.34.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.output.dense.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.output.dense.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.key.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.key.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.query.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.query.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.value.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.attention.self.value.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.ln.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.35.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00003.bin", "bert.encoder.layer.4.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.4.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.5.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.6.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.7.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.8.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.output.dense.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.output.dense.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.key.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.key.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.query.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.query.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.value.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.attention.self.value.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.ln.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.ln.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00003.bin", "bert.encoder.layer.9.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00003.bin", "bert.encoder.ln.bias": "pytorch_model-00003-of-00003.bin", "bert.encoder.ln.weight": "pytorch_model-00003-of-00003.bin", "output_embed.weight": "pytorch_model-00003-of-00003.bin" } }