|
33 | 33 | logger = logging.getLogger(__name__)
|
34 | 34 |
|
35 | 35 | _DEFAULT_4BIT_CONFIGS = {
|
36 |
| - "databricks/dolly-v2-3b": {"bits": 4, "sym": False, "group_size": 32, "ratio": 0.5}, |
| 36 | + "databricks/dolly-v2-3b": {"bits": 4, "sym": False, "group_size": 128, "ratio": 0.8}, |
37 | 37 | "EleutherAI/gpt-j-6b": {"bits": 4, "sym": False, "group_size": 64},
|
38 | 38 | "facebook/opt-6.7b": {"bits": 4, "sym": False, "group_size": 64, "ratio": 0.8},
|
39 | 39 | "bigscience/bloomz-7b1": {"bits": 4, "sym": False, "group_size": 32, "ratio": 0.6},
|
40 | 40 | "togethercomputer/RedPajama-INCITE-7B-Instruct": {"bits": 4, "sym": False, "group_size": 128},
|
41 |
| - "HuggingFaceH4/zephyr-7b-beta": {"bits": 4, "sym": True, "group_size": 64, "ratio": 0.6}, |
42 |
| - "meta-llama/Llama-2-7b": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.6}, |
| 41 | + "HuggingFaceH4/zephyr-7b-beta": { |
| 42 | + "bits": 4, |
| 43 | + "sym": True, |
| 44 | + "group_size": 128, |
| 45 | + "ratio": 0.8, |
| 46 | + "dataset": "wikitext2", |
| 47 | + "awq": True, |
| 48 | + }, |
| 49 | + "meta-llama/Llama-2-7b": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.7}, |
43 | 50 | "meta-llama/Llama-2-7b-chat": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.8},
|
44 | 51 | "meta-llama/Llama-2-13b-chat": {"bits": 4, "sym": True, "group_size": 64, "ratio": 0.8},
|
45 | 52 | "stabilityai/stablelm-3b-4e1t": {"bits": 4, "sym": True, "group_size": 64, "ratio": 0.8},
|
46 | 53 | "stablelm-epoch-3b-preview": {"bits": 4, "sym": True, "group_size": 64, "ratio": 0.8},
|
47 |
| - "stable-zephyr-3b-dpo": {"bits": 4, "sym": False, "group_size": 64, "ratio": 0.8}, |
| 54 | + "stabilityai/stablelm-zephyr-3b": {"bits": 4, "sym": False, "group_size": 128, "ratio": 1.0}, |
48 | 55 | "pansophic/rocket-3B": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.8},
|
49 | 56 | "THUDM/chatglm2-6b": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.72},
|
50 | 57 | "Qwen/Qwen-7B-Chat": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.6},
|
51 | 58 | "openlm-research/open_llama_3b": {"bits": 4, "sym": True, "group_size": 64, "all_layers": True},
|
52 | 59 | "tiiuae/falcon-7b": {"bits": 4, "sym": True, "group_size": 64, "all_layers": True},
|
53 | 60 | "psmathur/orca_mini_3b": {"bits": 4, "sym": True, "group_size": 64, "all_layers": True},
|
54 | 61 | "mistralai/Mixtral-8x7B-v0.1": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.8},
|
| 62 | + "facebook/opt-2.7b": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.7}, |
| 63 | + "togethercomputer/RedPajama-INCITE-Chat-3B-v1": {"bits": 4, "sym": False, "group_size": 128, "ratio": 0.8}, |
| 64 | + "lmsys/vicuna-7b-v1.5": {"bits": 4, "sym": False, "group_size": 128, "ratio": 1.0}, |
| 65 | + "stabilityai/stablelm-tuned-alpha-3b": {"bits": 4, "sym": False, "group_size": 128, "ratio": 0.8}, |
| 66 | + "mistralai/Mistral-7B-v0.1": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.9}, |
| 67 | + "baichuan-inc/Baichuan2-7B-Chat": { |
| 68 | + "bits": 4, |
| 69 | + "sym": True, |
| 70 | + "group_size": 128, |
| 71 | + "ratio": 0.8, |
| 72 | + "dataset": "wikitext2", |
| 73 | + "awq": True, |
| 74 | + }, |
55 | 75 | }
|
56 | 76 |
|
57 | 77 |
|
|
0 commit comments