Upload Phi4MMForCausalLM
Browse filesUpdated weights (Speech full fine-tuning, 4 epochs)
- config.json +2 -2
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +2 -2
- model-00002-of-00003.safetensors +2 -2
- model-00003-of-00003.safetensors +2 -2
- model.safetensors.index.json +63 -63
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "./
|
3 |
"architectures": [
|
4 |
"Phi4MMForCausalLM"
|
5 |
],
|
@@ -209,7 +209,7 @@
|
|
209 |
},
|
210 |
"tie_word_embeddings": true,
|
211 |
"torch_dtype": "bfloat16",
|
212 |
-
"transformers_version": "4.
|
213 |
"use_cache": true,
|
214 |
"vision_lora": {
|
215 |
"dp": 0.0,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "./output",
|
3 |
"architectures": [
|
4 |
"Phi4MMForCausalLM"
|
5 |
],
|
|
|
209 |
},
|
210 |
"tie_word_embeddings": true,
|
211 |
"torch_dtype": "bfloat16",
|
212 |
+
"transformers_version": "4.46.3",
|
213 |
"use_cache": true,
|
214 |
"vision_lora": {
|
215 |
"dp": 0.0,
|
generation_config.json
CHANGED
@@ -6,5 +6,5 @@
|
|
6 |
199999
|
7 |
],
|
8 |
"pad_token_id": 199999,
|
9 |
-
"transformers_version": "4.
|
10 |
}
|
|
|
6 |
199999
|
7 |
],
|
8 |
"pad_token_id": 199999,
|
9 |
+
"transformers_version": "4.46.3"
|
10 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b93562f8e2135e7d780834a9f2300699e9e71d56d64e986386e92906ae571ca9
|
3 |
+
size 4998420448
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c974d2962a9fa092d2219161c87234175e860328c5978591548f0886e929f33
|
3 |
+
size 4983891952
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d44c5ed024d7f64060d50ad89a41380def480e396d6ae71b774ee9f6125ec38a
|
3 |
+
size 1905111704
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
@@ -1741,50 +1741,50 @@
|
|
1741 |
"model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1742 |
"model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1743 |
"model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1744 |
-
"model.layers.25.input_layernorm.weight": "model-
|
1745 |
-
"model.layers.25.mlp.down_proj.base_layer.weight": "model-
|
1746 |
-
"model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-
|
1747 |
-
"model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-
|
1748 |
-
"model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-
|
1749 |
-
"model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-
|
1750 |
-
"model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-
|
1751 |
-
"model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1752 |
-
"model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1753 |
-
"model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-
|
1754 |
-
"model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1755 |
-
"model.layers.25.post_attention_layernorm.weight": "model-
|
1756 |
"model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1757 |
"model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
"model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1759 |
"model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1760 |
"model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1761 |
-
"model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-
|
1762 |
-
"model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1763 |
-
"model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1764 |
-
"model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1765 |
-
"model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1766 |
-
"model.layers.26.input_layernorm.weight": "model-
|
1767 |
-
"model.layers.26.mlp.down_proj.base_layer.weight": "model-
|
1768 |
-
"model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-
|
1769 |
-
"model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-
|
1770 |
-
"model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-
|
1771 |
-
"model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-
|
1772 |
-
"model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-
|
1773 |
-
"model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1774 |
-
"model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1775 |
-
"model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-
|
1776 |
-
"model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1777 |
-
"model.layers.26.post_attention_layernorm.weight": "model-
|
1778 |
-
"model.layers.26.self_attn.o_proj.base_layer.weight": "model-
|
1779 |
-
"model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-
|
1780 |
-
"model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-
|
1781 |
-
"model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-
|
1782 |
-
"model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-
|
1783 |
-
"model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-
|
1784 |
-
"model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1785 |
-
"model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1786 |
-
"model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1787 |
-
"model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1788 |
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1789 |
"model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1790 |
"model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
@@ -1797,16 +1797,16 @@
|
|
1797 |
"model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1798 |
"model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1799 |
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1800 |
-
"model.layers.27.self_attn.o_proj.base_layer.weight": "model-
|
1801 |
-
"model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-
|
1802 |
-
"model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-
|
1803 |
-
"model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-
|
1804 |
-
"model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-
|
1805 |
-
"model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-
|
1806 |
-
"model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1807 |
-
"model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1808 |
-
"model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1809 |
-
"model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1810 |
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1811 |
"model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1812 |
"model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
@@ -1989,22 +1989,22 @@
|
|
1989 |
"model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1990 |
"model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1991 |
"model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1992 |
-
"model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-
|
1993 |
-
"model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1994 |
-
"model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1995 |
"model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1996 |
-
"model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1997 |
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1998 |
"model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1999 |
-
"model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-
|
2000 |
-
"model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-
|
2001 |
-
"model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-
|
2002 |
-
"model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-
|
2003 |
-
"model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-
|
2004 |
-
"model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
2005 |
-
"model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
2006 |
-
"model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
2007 |
-
"model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
2008 |
"model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
2009 |
"model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
2010 |
"model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 11887118272
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
|
|
1741 |
"model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1742 |
"model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1743 |
"model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1744 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1745 |
+
"model.layers.25.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1746 |
+
"model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1747 |
+
"model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1748 |
+
"model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1749 |
+
"model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1750 |
+
"model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1751 |
+
"model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1752 |
+
"model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1753 |
+
"model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1754 |
+
"model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1755 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1756 |
"model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1757 |
"model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
"model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1759 |
"model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1760 |
"model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1761 |
+
"model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1762 |
+
"model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1763 |
+
"model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1764 |
+
"model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1765 |
+
"model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1766 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1767 |
+
"model.layers.26.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1768 |
+
"model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1769 |
+
"model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1770 |
+
"model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1771 |
+
"model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1772 |
+
"model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1773 |
+
"model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1774 |
+
"model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1775 |
+
"model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1776 |
+
"model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1777 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1778 |
+
"model.layers.26.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1779 |
+
"model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1780 |
+
"model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1781 |
+
"model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1782 |
+
"model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1783 |
+
"model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1784 |
+
"model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1785 |
+
"model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1786 |
+
"model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1787 |
+
"model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1788 |
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1789 |
"model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1790 |
"model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
|
|
1797 |
"model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1798 |
"model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1799 |
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1800 |
+
"model.layers.27.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1801 |
+
"model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1802 |
+
"model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1803 |
+
"model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1804 |
+
"model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1805 |
+
"model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1806 |
+
"model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1807 |
+
"model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-00003-of-00003.safetensors",
|
1808 |
+
"model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1809 |
+
"model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1810 |
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1811 |
"model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1812 |
"model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
|
|
1989 |
"model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1990 |
"model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1991 |
"model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1992 |
+
"model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1993 |
+
"model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1994 |
+
"model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1995 |
"model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1996 |
+
"model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1997 |
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1998 |
"model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1999 |
+
"model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
2000 |
+
"model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
2001 |
+
"model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
2002 |
+
"model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
2003 |
+
"model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
2004 |
+
"model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
2005 |
+
"model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
2006 |
+
"model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
2007 |
+
"model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
2008 |
"model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
2009 |
"model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
2010 |
"model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|