You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As we have moved to the transformer version 4.50.0, we have changed the
attention implementation and hence number of attention heads and hidden
size are no more attributes of this class.
We have added these parameters are optional arguments for the
replicate_kv_heads script. User can now pass them as arguments, if not
passed explicitly then it shall be picked from config.json file of the
model.
---------
Signed-off-by: Hem Agnihotri <[email protected]>
Copy file name to clipboardExpand all lines: scripts/replicate_kv_head/README.md
+3-1Lines changed: 3 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -30,4 +30,6 @@ Replace `<hf_token>` with your actual token.
30
30
### Arguments
31
31
- **--model_name**: Model card name to use (default: “meta-llama/Meta-Llama-3-8B-Instruct”).
32
32
- **--prompt**: Prompt to use for the model (default: “My name is”).
33
-
- **--repeat**: Factor to repeat key-value heads (default: 2).
33
+
- **--repeat**: Factor to repeat key-value heads (default: 2).
34
+
- **--num_attention_heads**: Number of attentin heads (default: None). This is optional param, if not given explicitly the will be read from config.json.
35
+
- **--hidden_size**: Hidden size (default: None). This is optional param, if not given explicitly the will be read from config.json.
0 commit comments