camenduru commited on
Commit
77984b1
·
1 Parent(s): a0061e7

thanks to microsoft/Llama-2-Onnx ❤

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +296 -35
  2. .gitignore +1 -0
  3. LICENSE +126 -0
  4. MODEL-CARD-META-LLAMA-2.md +97 -0
  5. ONNX/Constant_69_attr__value +0 -0
  6. ONNX/Constant_77_attr__value +0 -0
  7. ONNX/LlamaV2_7B_FT_float16.onnx +3 -0
  8. ONNX/embeddings.data +3 -0
  9. ONNX/onnx__MatMul_20670 +3 -0
  10. ONNX/onnx__MatMul_20671 +3 -0
  11. ONNX/onnx__MatMul_20672 +3 -0
  12. ONNX/onnx__MatMul_20692 +3 -0
  13. ONNX/onnx__MatMul_20693 +3 -0
  14. ONNX/onnx__MatMul_20694 +3 -0
  15. ONNX/onnx__MatMul_20695 +3 -0
  16. ONNX/onnx__MatMul_20696 +3 -0
  17. ONNX/onnx__MatMul_20697 +3 -0
  18. ONNX/onnx__MatMul_20698 +3 -0
  19. ONNX/onnx__MatMul_20718 +3 -0
  20. ONNX/onnx__MatMul_20719 +3 -0
  21. ONNX/onnx__MatMul_20720 +3 -0
  22. ONNX/onnx__MatMul_20721 +3 -0
  23. ONNX/onnx__MatMul_20722 +3 -0
  24. ONNX/onnx__MatMul_20723 +3 -0
  25. ONNX/onnx__MatMul_20724 +3 -0
  26. ONNX/onnx__MatMul_20744 +3 -0
  27. ONNX/onnx__MatMul_20745 +3 -0
  28. ONNX/onnx__MatMul_20746 +3 -0
  29. ONNX/onnx__MatMul_20747 +3 -0
  30. ONNX/onnx__MatMul_20748 +3 -0
  31. ONNX/onnx__MatMul_20749 +3 -0
  32. ONNX/onnx__MatMul_20750 +3 -0
  33. ONNX/onnx__MatMul_20770 +3 -0
  34. ONNX/onnx__MatMul_20771 +3 -0
  35. ONNX/onnx__MatMul_20772 +3 -0
  36. ONNX/onnx__MatMul_20773 +3 -0
  37. ONNX/onnx__MatMul_20774 +3 -0
  38. ONNX/onnx__MatMul_20775 +3 -0
  39. ONNX/onnx__MatMul_20776 +3 -0
  40. ONNX/onnx__MatMul_20796 +3 -0
  41. ONNX/onnx__MatMul_20797 +3 -0
  42. ONNX/onnx__MatMul_20798 +3 -0
  43. ONNX/onnx__MatMul_20799 +3 -0
  44. ONNX/onnx__MatMul_20800 +3 -0
  45. ONNX/onnx__MatMul_20801 +3 -0
  46. ONNX/onnx__MatMul_20802 +3 -0
  47. ONNX/onnx__MatMul_20822 +3 -0
  48. ONNX/onnx__MatMul_20823 +3 -0
  49. ONNX/onnx__MatMul_20824 +3 -0
  50. ONNX/onnx__MatMul_20825 +3 -0
.gitattributes CHANGED
@@ -1,35 +1,296 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ONNX/onnx__MatMul_20745 filter=lfs diff=lfs merge=lfs -text
2
+ ONNX/onnx__MatMul_20979 filter=lfs diff=lfs merge=lfs -text
3
+ ONNX/onnx__MatMul_21086 filter=lfs diff=lfs merge=lfs -text
4
+ ONNX/onnx__MatMul_21342 filter=lfs diff=lfs merge=lfs -text
5
+ ONNX/onnx__MatMul_20724 filter=lfs diff=lfs merge=lfs -text
6
+ ONNX/onnx__MatMul_21032 filter=lfs diff=lfs merge=lfs -text
7
+ ONNX/onnx__MatMul_21165 filter=lfs diff=lfs merge=lfs -text
8
+ ONNX/onnx__MatMul_21451 filter=lfs diff=lfs merge=lfs -text
9
+ ONNX/onnx__MatMul_20854 filter=lfs diff=lfs merge=lfs -text
10
+ ONNX/onnx__MatMul_20902 filter=lfs diff=lfs merge=lfs -text
11
+ ONNX/transformer.block_list.27.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
12
+ ONNX/transformer.block_list.5.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
13
+ ONNX/Constant_77_attr__value filter=lfs diff=lfs merge=lfs -text
14
+ ONNX/onnx__MatMul_20928 filter=lfs diff=lfs merge=lfs -text
15
+ ONNX/onnx__MatMul_20958 filter=lfs diff=lfs merge=lfs -text
16
+ ONNX/onnx__MatMul_21318 filter=lfs diff=lfs merge=lfs -text
17
+ ONNX/transformer.block_list.4.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
18
+ ONNX/transformer.block_list.8.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
19
+ ONNX/onnx__MatMul_20672 filter=lfs diff=lfs merge=lfs -text
20
+ ONNX/onnx__MatMul_20906 filter=lfs diff=lfs merge=lfs -text
21
+ ONNX/onnx__MatMul_21213 filter=lfs diff=lfs merge=lfs -text
22
+ ONNX/onnx__MatMul_20927 filter=lfs diff=lfs merge=lfs -text
23
+ ONNX/onnx__MatMul_21322 filter=lfs diff=lfs merge=lfs -text
24
+ ONNX/onnx__MatMul_21425 filter=lfs diff=lfs merge=lfs -text
25
+ ONNX/onnx__MatMul_21502 filter=lfs diff=lfs merge=lfs -text
26
+ ONNX/transformer.block_list.18.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
27
+ ONNX/transformer.block_list.24.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
28
+ ONNX/onnx__MatMul_20744 filter=lfs diff=lfs merge=lfs -text
29
+ ONNX/onnx__MatMul_21399 filter=lfs diff=lfs merge=lfs -text
30
+ ONNX/onnx__MatMul_21498 filter=lfs diff=lfs merge=lfs -text
31
+ ONNX/transformer.block_list.25.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
32
+ ONNX/transformer.block_list.26.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
33
+ ONNX/transformer.block_list.28.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
34
+ ONNX/transformer.block_list.2.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
35
+ ONNX/onnx__MatMul_20797 filter=lfs diff=lfs merge=lfs -text
36
+ ONNX/onnx__MatMul_20822 filter=lfs diff=lfs merge=lfs -text
37
+ ONNX/onnx__MatMul_20932 filter=lfs diff=lfs merge=lfs -text
38
+ ONNX/onnx__MatMul_21109 filter=lfs diff=lfs merge=lfs -text
39
+ ONNX/onnx__MatMul_21139 filter=lfs diff=lfs merge=lfs -text
40
+ ONNX/onnx__MatMul_21319 filter=lfs diff=lfs merge=lfs -text
41
+ ONNX/onnx__MatMul_21370 filter=lfs diff=lfs merge=lfs -text
42
+ ONNX/onnx__MatMul_20931 filter=lfs diff=lfs merge=lfs -text
43
+ ONNX/onnx__MatMul_21138 filter=lfs diff=lfs merge=lfs -text
44
+ ONNX/onnx__MatMul_21446 filter=lfs diff=lfs merge=lfs -text
45
+ ONNX/transformer.block_list.31.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
46
+ ONNX/onnx__MatMul_20720 filter=lfs diff=lfs merge=lfs -text
47
+ ONNX/onnx__MatMul_21034 filter=lfs diff=lfs merge=lfs -text
48
+ ONNX/onnx__MatMul_21112 filter=lfs diff=lfs merge=lfs -text
49
+ ONNX/onnx__MatMul_21396 filter=lfs diff=lfs merge=lfs -text
50
+ ONNX/transformer.block_list.3.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
51
+ ONNX/onnx__MatMul_20828 filter=lfs diff=lfs merge=lfs -text
52
+ ONNX/onnx__MatMul_20880 filter=lfs diff=lfs merge=lfs -text
53
+ ONNX/transformer.block_list.15.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
54
+ ONNX/onnx__MatMul_21264 filter=lfs diff=lfs merge=lfs -text
55
+ ONNX/transformer.block_list.21.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
56
+ ONNX/onnx__MatMul_20692 filter=lfs diff=lfs merge=lfs -text
57
+ ONNX/onnx__MatMul_20721 filter=lfs diff=lfs merge=lfs -text
58
+ ONNX/onnx__MatMul_21162 filter=lfs diff=lfs merge=lfs -text
59
+ ONNX/onnx__MatMul_21268 filter=lfs diff=lfs merge=lfs -text
60
+ ONNX/onnx__MatMul_21372 filter=lfs diff=lfs merge=lfs -text
61
+ ONNX/onnx__MatMul_21476 filter=lfs diff=lfs merge=lfs -text
62
+ ONNX/onnx__MatMul_21501 filter=lfs diff=lfs merge=lfs -text
63
+ ONNX/onnx__MatMul_21036 filter=lfs diff=lfs merge=lfs -text
64
+ ONNX/onnx__MatMul_21163 filter=lfs diff=lfs merge=lfs -text
65
+ ONNX/onnx__MatMul_21426 filter=lfs diff=lfs merge=lfs -text
66
+ ONNX/transformer.block_list.25.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
67
+ ONNX/transformer.block_list.6.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
68
+ ONNX/onnx__MatMul_20824 filter=lfs diff=lfs merge=lfs -text
69
+ ONNX/onnx__MatMul_21057 filter=lfs diff=lfs merge=lfs -text
70
+ ONNX/onnx__MatMul_21240 filter=lfs diff=lfs merge=lfs -text
71
+ ONNX/transformer.block_list.23.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
72
+ ONNX/transformer.block_list.7.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
73
+ ONNX/onnx__MatMul_20723 filter=lfs diff=lfs merge=lfs -text
74
+ ONNX/onnx__MatMul_20852 filter=lfs diff=lfs merge=lfs -text
75
+ ONNX/onnx__MatMul_21056 filter=lfs diff=lfs merge=lfs -text
76
+ ONNX/onnx__MatMul_21192 filter=lfs diff=lfs merge=lfs -text
77
+ ONNX/onnx__MatMul_21316 filter=lfs diff=lfs merge=lfs -text
78
+ ONNX/onnx__MatMul_21450 filter=lfs diff=lfs merge=lfs -text
79
+ ONNX/transformer.block_list.12.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
80
+ ONNX/onnx__MatMul_20952 filter=lfs diff=lfs merge=lfs -text
81
+ ONNX/onnx__MatMul_21010 filter=lfs diff=lfs merge=lfs -text
82
+ ONNX/onnx__MatMul_21031 filter=lfs diff=lfs merge=lfs -text
83
+ ONNX/onnx__MatMul_21160 filter=lfs diff=lfs merge=lfs -text
84
+ ONNX/onnx__MatMul_21265 filter=lfs diff=lfs merge=lfs -text
85
+ ONNX/onnx__MatMul_20876 filter=lfs diff=lfs merge=lfs -text
86
+ ONNX/onnx__MatMul_21082 filter=lfs diff=lfs merge=lfs -text
87
+ ONNX/onnx__MatMul_21110 filter=lfs diff=lfs merge=lfs -text
88
+ ONNX/onnx__MatMul_21135 filter=lfs diff=lfs merge=lfs -text
89
+ ONNX/onnx__MatMul_21187 filter=lfs diff=lfs merge=lfs -text
90
+ ONNX/onnx__MatMul_21320 filter=lfs diff=lfs merge=lfs -text
91
+ ONNX/onnx__MatMul_20799 filter=lfs diff=lfs merge=lfs -text
92
+ ONNX/onnx__MatMul_20954 filter=lfs diff=lfs merge=lfs -text
93
+ ONNX/onnx__MatMul_20957 filter=lfs diff=lfs merge=lfs -text
94
+ ONNX/onnx__MatMul_21217 filter=lfs diff=lfs merge=lfs -text
95
+ ONNX/transformer.block_list.29.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
96
+ ONNX/LlamaV2_7B_FT_float16.onnx filter=lfs diff=lfs merge=lfs -text
97
+ ONNX/onnx__MatMul_20982 filter=lfs diff=lfs merge=lfs -text
98
+ ONNX/onnx__MatMul_21166 filter=lfs diff=lfs merge=lfs -text
99
+ ONNX/onnx__MatMul_21238 filter=lfs diff=lfs merge=lfs -text
100
+ ONNX/onnx__MatMul_21243 filter=lfs diff=lfs merge=lfs -text
101
+ ONNX/onnx__MatMul_21267 filter=lfs diff=lfs merge=lfs -text
102
+ ONNX/onnx__MatMul_21218 filter=lfs diff=lfs merge=lfs -text
103
+ ONNX/onnx__MatMul_21373 filter=lfs diff=lfs merge=lfs -text
104
+ ONNX/onnx__MatMul_20749 filter=lfs diff=lfs merge=lfs -text
105
+ ONNX/onnx__MatMul_20980 filter=lfs diff=lfs merge=lfs -text
106
+ ONNX/onnx__MatMul_21293 filter=lfs diff=lfs merge=lfs -text
107
+ ONNX/onnx__MatMul_21452 filter=lfs diff=lfs merge=lfs -text
108
+ ONNX/transformer.block_list.13.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
109
+ ONNX/transformer.block_list.19.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
110
+ ONNX/transformer.block_list.5.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
111
+ ONNX/onnx__MatMul_20722 filter=lfs diff=lfs merge=lfs -text
112
+ ONNX/onnx__MatMul_20826 filter=lfs diff=lfs merge=lfs -text
113
+ ONNX/onnx__MatMul_21030 filter=lfs diff=lfs merge=lfs -text
114
+ ONNX/transformer.block_list.26.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
115
+ ONNX/onnx__MatMul_20694 filter=lfs diff=lfs merge=lfs -text
116
+ ONNX/onnx__MatMul_20802 filter=lfs diff=lfs merge=lfs -text
117
+ ONNX/onnx__MatMul_20930 filter=lfs diff=lfs merge=lfs -text
118
+ ONNX/onnx__MatMul_21058 filter=lfs diff=lfs merge=lfs -text
119
+ ONNX/onnx__MatMul_21059 filter=lfs diff=lfs merge=lfs -text
120
+ ONNX/onnx__MatMul_21369 filter=lfs diff=lfs merge=lfs -text
121
+ ONNX/transformer.block_list.1.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
122
+ ONNX/onnx__MatMul_21136 filter=lfs diff=lfs merge=lfs -text
123
+ ONNX/transformer.block_list.10.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
124
+ ONNX/transformer.block_list.24.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
125
+ ONNX/onnx__MatMul_21216 filter=lfs diff=lfs merge=lfs -text
126
+ ONNX/onnx__MatMul_21241 filter=lfs diff=lfs merge=lfs -text
127
+ ONNX/onnx__MatMul_21346 filter=lfs diff=lfs merge=lfs -text
128
+ ONNX/transformer.block_list.0.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
129
+ ONNX/onnx__MatMul_20772 filter=lfs diff=lfs merge=lfs -text
130
+ ONNX/onnx__MatMul_21087 filter=lfs diff=lfs merge=lfs -text
131
+ ONNX/onnx__MatMul_21137 filter=lfs diff=lfs merge=lfs -text
132
+ ONNX/onnx__MatMul_21290 filter=lfs diff=lfs merge=lfs -text
133
+ ONNX/onnx__MatMul_20776 filter=lfs diff=lfs merge=lfs -text
134
+ ONNX/onnx__MatMul_20823 filter=lfs diff=lfs merge=lfs -text
135
+ ONNX/onnx__MatMul_20875 filter=lfs diff=lfs merge=lfs -text
136
+ ONNX/onnx__MatMul_21191 filter=lfs diff=lfs merge=lfs -text
137
+ ONNX/onnx__MatMul_21347 filter=lfs diff=lfs merge=lfs -text
138
+ ONNX/transformer.block_list.28.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
139
+ ONNX/onnx__MatMul_21009 filter=lfs diff=lfs merge=lfs -text
140
+ ONNX/onnx__MatMul_21343 filter=lfs diff=lfs merge=lfs -text
141
+ ONNX/onnx__MatMul_21344 filter=lfs diff=lfs merge=lfs -text
142
+ ONNX/transformer.block_list.17.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
143
+ ONNX/transformer.block_list.30.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
144
+ ONNX/onnx__MatMul_20718 filter=lfs diff=lfs merge=lfs -text
145
+ ONNX/onnx__MatMul_20877 filter=lfs diff=lfs merge=lfs -text
146
+ ONNX/onnx__MatMul_21423 filter=lfs diff=lfs merge=lfs -text
147
+ ONNX/transformer.block_list.0.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
148
+ ONNX/onnx__MatMul_21004 filter=lfs diff=lfs merge=lfs -text
149
+ ONNX/onnx__MatMul_21005 filter=lfs diff=lfs merge=lfs -text
150
+ ONNX/onnx__MatMul_21424 filter=lfs diff=lfs merge=lfs -text
151
+ ONNX/transformer.block_list.12.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
152
+ ONNX/transformer.block_list.13.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
153
+ ONNX/transformer.block_list.19.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
154
+ ONNX/transformer.block_list.8.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
155
+ ONNX/onnx__MatMul_20770 filter=lfs diff=lfs merge=lfs -text
156
+ ONNX/onnx__MatMul_20801 filter=lfs diff=lfs merge=lfs -text
157
+ ONNX/onnx__MatMul_21108 filter=lfs diff=lfs merge=lfs -text
158
+ ONNX/onnx__MatMul_21113 filter=lfs diff=lfs merge=lfs -text
159
+ ONNX/onnx__MatMul_21321 filter=lfs diff=lfs merge=lfs -text
160
+ ONNX/onnx__MatMul_21348 filter=lfs diff=lfs merge=lfs -text
161
+ ONNX/onnx__MatMul_21475 filter=lfs diff=lfs merge=lfs -text
162
+ ONNX/onnx__MatMul_20771 filter=lfs diff=lfs merge=lfs -text
163
+ ONNX/onnx__MatMul_21007 filter=lfs diff=lfs merge=lfs -text
164
+ ONNX/transformer.block_list.10.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
165
+ ONNX/onnx__MatMul_20695 filter=lfs diff=lfs merge=lfs -text
166
+ ONNX/onnx__MatMul_20878 filter=lfs diff=lfs merge=lfs -text
167
+ ONNX/onnx__MatMul_20983 filter=lfs diff=lfs merge=lfs -text
168
+ ONNX/onnx__MatMul_21062 filter=lfs diff=lfs merge=lfs -text
169
+ ONNX/onnx__MatMul_20670 filter=lfs diff=lfs merge=lfs -text
170
+ ONNX/onnx__MatMul_20874 filter=lfs diff=lfs merge=lfs -text
171
+ ONNX/onnx__MatMul_21186 filter=lfs diff=lfs merge=lfs -text
172
+ ONNX/onnx__MatMul_20696 filter=lfs diff=lfs merge=lfs -text
173
+ ONNX/onnx__MatMul_20955 filter=lfs diff=lfs merge=lfs -text
174
+ ONNX/onnx__MatMul_21083 filter=lfs diff=lfs merge=lfs -text
175
+ ONNX/onnx__MatMul_20746 filter=lfs diff=lfs merge=lfs -text
176
+ ONNX/onnx__MatMul_20825 filter=lfs diff=lfs merge=lfs -text
177
+ ONNX/onnx__MatMul_21400 filter=lfs diff=lfs merge=lfs -text
178
+ ONNX/onnx__MatMul_21449 filter=lfs diff=lfs merge=lfs -text
179
+ ONNX/onnx__MatMul_21478 filter=lfs diff=lfs merge=lfs -text
180
+ ONNX/transformer.block_list.11.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
181
+ ONNX/onnx__MatMul_20698 filter=lfs diff=lfs merge=lfs -text
182
+ ONNX/onnx__MatMul_21006 filter=lfs diff=lfs merge=lfs -text
183
+ ONNX/onnx__MatMul_21188 filter=lfs diff=lfs merge=lfs -text
184
+ ONNX/transformer.block_list.17.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
185
+ ONNX/onnx__MatMul_20748 filter=lfs diff=lfs merge=lfs -text
186
+ ONNX/onnx__MatMul_20853 filter=lfs diff=lfs merge=lfs -text
187
+ ONNX/onnx__MatMul_20953 filter=lfs diff=lfs merge=lfs -text
188
+ ONNX/onnx__MatMul_21212 filter=lfs diff=lfs merge=lfs -text
189
+ ONNX/onnx__MatMul_20774 filter=lfs diff=lfs merge=lfs -text
190
+ ONNX/onnx__MatMul_20929 filter=lfs diff=lfs merge=lfs -text
191
+ ONNX/onnx__MatMul_20984 filter=lfs diff=lfs merge=lfs -text
192
+ ONNX/onnx__MatMul_21060 filter=lfs diff=lfs merge=lfs -text
193
+ ONNX/onnx__MatMul_21242 filter=lfs diff=lfs merge=lfs -text
194
+ ONNX/onnx__MatMul_21317 filter=lfs diff=lfs merge=lfs -text
195
+ ONNX/onnx__MatMul_21140 filter=lfs diff=lfs merge=lfs -text
196
+ ONNX/onnx__MatMul_21161 filter=lfs diff=lfs merge=lfs -text
197
+ ONNX/onnx__MatMul_21244 filter=lfs diff=lfs merge=lfs -text
198
+ ONNX/onnx__MatMul_21374 filter=lfs diff=lfs merge=lfs -text
199
+ ONNX/embeddings.data filter=lfs diff=lfs merge=lfs -text
200
+ ONNX/onnx__MatMul_20697 filter=lfs diff=lfs merge=lfs -text
201
+ ONNX/onnx__MatMul_20879 filter=lfs diff=lfs merge=lfs -text
202
+ ONNX/onnx__MatMul_21085 filter=lfs diff=lfs merge=lfs -text
203
+ ONNX/transformer.block_list.22.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
204
+ ONNX/onnx__MatMul_21088 filter=lfs diff=lfs merge=lfs -text
205
+ ONNX/onnx__MatMul_21398 filter=lfs diff=lfs merge=lfs -text
206
+ ONNX/onnx__MatMul_21422 filter=lfs diff=lfs merge=lfs -text
207
+ ONNX/Constant_69_attr__value filter=lfs diff=lfs merge=lfs -text
208
+ ONNX/onnx__MatMul_20800 filter=lfs diff=lfs merge=lfs -text
209
+ ONNX/onnx__MatMul_20904 filter=lfs diff=lfs merge=lfs -text
210
+ ONNX/onnx__MatMul_21215 filter=lfs diff=lfs merge=lfs -text
211
+ ONNX/transformer.block_list.9.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
212
+ ONNX/onnx__MatMul_20693 filter=lfs diff=lfs merge=lfs -text
213
+ ONNX/onnx__MatMul_20981 filter=lfs diff=lfs merge=lfs -text
214
+ ONNX/onnx__MatMul_21061 filter=lfs diff=lfs merge=lfs -text
215
+ ONNX/transformer.block_list.14.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
216
+ ONNX/transformer.block_list.22.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
217
+ ONNX/onnx__MatMul_20901 filter=lfs diff=lfs merge=lfs -text
218
+ ONNX/onnx__MatMul_21084 filter=lfs diff=lfs merge=lfs -text
219
+ ONNX/onnx__MatMul_21266 filter=lfs diff=lfs merge=lfs -text
220
+ ONNX/onnx__MatMul_20905 filter=lfs diff=lfs merge=lfs -text
221
+ ONNX/onnx__MatMul_21134 filter=lfs diff=lfs merge=lfs -text
222
+ ONNX/transformer.block_list.21.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
223
+ ONNX/onnx__MatMul_20850 filter=lfs diff=lfs merge=lfs -text
224
+ ONNX/onnx__MatMul_21114 filter=lfs diff=lfs merge=lfs -text
225
+ ONNX/onnx__MatMul_21164 filter=lfs diff=lfs merge=lfs -text
226
+ ONNX/onnx__MatMul_21395 filter=lfs diff=lfs merge=lfs -text
227
+ ONNX/onnx__MatMul_21472 filter=lfs diff=lfs merge=lfs -text
228
+ ONNX/transformer.block_list.31.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
229
+ ONNX/onnx__MatMul_21291 filter=lfs diff=lfs merge=lfs -text
230
+ ONNX/onnx__MatMul_21394 filter=lfs diff=lfs merge=lfs -text
231
+ ONNX/onnx__MatMul_21448 filter=lfs diff=lfs merge=lfs -text
232
+ ONNX/transformer.block_list.20.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
233
+ ONNX/transformer.block_list.2.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
234
+ ONNX/transformer.block_list.20.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
235
+ ONNX/transformer.block_list.23.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
236
+ ONNX/onnx__MatMul_20827 filter=lfs diff=lfs merge=lfs -text
237
+ ONNX/onnx__MatMul_21292 filter=lfs diff=lfs merge=lfs -text
238
+ ONNX/onnx__MatMul_20671 filter=lfs diff=lfs merge=lfs -text
239
+ ONNX/onnx__MatMul_20956 filter=lfs diff=lfs merge=lfs -text
240
+ ONNX/onnx__MatMul_21270 filter=lfs diff=lfs merge=lfs -text
241
+ ONNX/transformer.block_list.1.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
242
+ ONNX/transformer.block_list.3.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
243
+ ONNX/onnx__MatMul_20978 filter=lfs diff=lfs merge=lfs -text
244
+ ONNX/onnx__MatMul_21033 filter=lfs diff=lfs merge=lfs -text
245
+ ONNX/onnx__MatMul_21345 filter=lfs diff=lfs merge=lfs -text
246
+ ONNX/onnx__MatMul_21500 filter=lfs diff=lfs merge=lfs -text
247
+ ONNX/onnx__MatMul_21189 filter=lfs diff=lfs merge=lfs -text
248
+ ONNX/onnx__MatMul_21239 filter=lfs diff=lfs merge=lfs -text
249
+ ONNX/onnx__MatMul_21295 filter=lfs diff=lfs merge=lfs -text
250
+ ONNX/transformer.block_list.15.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
251
+ ONNX/onnx__MatMul_20775 filter=lfs diff=lfs merge=lfs -text
252
+ ONNX/onnx__MatMul_20798 filter=lfs diff=lfs merge=lfs -text
253
+ ONNX/onnx__MatMul_20851 filter=lfs diff=lfs merge=lfs -text
254
+ ONNX/onnx__MatMul_21420 filter=lfs diff=lfs merge=lfs -text
255
+ ONNX/transformer.block_list.30.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
256
+ ONNX/transformer.block_list.9.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
257
+ ONNX/transformer.block_list.16.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
258
+ ONNX/onnx__MatMul_20719 filter=lfs diff=lfs merge=lfs -text
259
+ ONNX/onnx__MatMul_20796 filter=lfs diff=lfs merge=lfs -text
260
+ ONNX/onnx__MatMul_20900 filter=lfs diff=lfs merge=lfs -text
261
+ ONNX/onnx__MatMul_20926 filter=lfs diff=lfs merge=lfs -text
262
+ ONNX/onnx__MatMul_21269 filter=lfs diff=lfs merge=lfs -text
263
+ ONNX/onnx__MatMul_21296 filter=lfs diff=lfs merge=lfs -text
264
+ ONNX/onnx__MatMul_21421 filter=lfs diff=lfs merge=lfs -text
265
+ ONNX/transformer.block_list.6.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
266
+ ONNX/onnx__MatMul_20848 filter=lfs diff=lfs merge=lfs -text
267
+ ONNX/onnx__MatMul_21190 filter=lfs diff=lfs merge=lfs -text
268
+ ONNX/onnx__MatMul_21371 filter=lfs diff=lfs merge=lfs -text
269
+ ONNX/onnx__MatMul_21477 filter=lfs diff=lfs merge=lfs -text
270
+ ONNX/transformer.block_list.27.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
271
+ ONNX/transformer.block_list.29.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
272
+ ONNX/transformer.block_list.4.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
273
+ ONNX/transformer.block_list.7.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
274
+ ONNX/onnx__MatMul_20747 filter=lfs diff=lfs merge=lfs -text
275
+ ONNX/onnx__MatMul_20849 filter=lfs diff=lfs merge=lfs -text
276
+ ONNX/onnx__MatMul_21035 filter=lfs diff=lfs merge=lfs -text
277
+ ONNX/onnx__MatMul_21474 filter=lfs diff=lfs merge=lfs -text
278
+ ONNX/onnx__MatMul_21499 filter=lfs diff=lfs merge=lfs -text
279
+ ONNX/onnx__MatMul_21294 filter=lfs diff=lfs merge=lfs -text
280
+ ONNX/onnx__MatMul_21473 filter=lfs diff=lfs merge=lfs -text
281
+ ONNX/transformer.block_list.11.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
282
+ ONNX/transformer.block_list.16.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
283
+ ONNX/transformer.layer_norm.weight filter=lfs diff=lfs merge=lfs -text
284
+ ONNX/onnx__MatMul_20750 filter=lfs diff=lfs merge=lfs -text
285
+ ONNX/onnx__MatMul_20773 filter=lfs diff=lfs merge=lfs -text
286
+ ONNX/onnx__MatMul_21008 filter=lfs diff=lfs merge=lfs -text
287
+ ONNX/onnx__MatMul_21111 filter=lfs diff=lfs merge=lfs -text
288
+ ONNX/onnx__MatMul_21397 filter=lfs diff=lfs merge=lfs -text
289
+ ONNX/onnx__MatMul_21447 filter=lfs diff=lfs merge=lfs -text
290
+ ONNX/transformer.block_list.14.proj_norm.weight filter=lfs diff=lfs merge=lfs -text
291
+ ONNX/onnx__MatMul_20903 filter=lfs diff=lfs merge=lfs -text
292
+ ONNX/onnx__MatMul_21214 filter=lfs diff=lfs merge=lfs -text
293
+ ONNX/onnx__MatMul_21368 filter=lfs diff=lfs merge=lfs -text
294
+ ONNX/transformer.block_list.18.attn_norm.weight filter=lfs diff=lfs merge=lfs -text
295
+ embeddings.pth filter=lfs diff=lfs merge=lfs -text
296
+ RESPONSIBLE-USE-GUIDE-META-LLAMA-2.pdf filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+
LICENSE ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LLAMA 2 COMMUNITY LICENSE AGREEMENT
2
+ Llama 2 Version Release Date: July 18, 2023
3
+
4
+ "Agreement" means the terms and conditions for use, reproduction, distribution and
5
+ modification of the Llama Materials set forth herein.
6
+
7
+ "Documentation" means the specifications, manuals and documentation
8
+ accompanying Llama 2 distributed by Meta at ai.meta.com/resources/models-and-
9
+ libraries/llama-downloads/.
10
+
11
+ "Licensee" or "you" means you, or your employer or any other person or entity (if
12
+ you are entering into this Agreement on such person or entity's behalf), of the age
13
+ required under applicable laws, rules or regulations to provide legal consent and that
14
+ has legal authority to bind your employer or such other person or entity if you are
15
+ entering in this Agreement on their behalf.
16
+
17
+ "Llama 2" means the foundational large language models and software and
18
+ algorithms, including machine-learning model code, trained model weights,
19
+ inference-enabling code, training-enabling code, fine-tuning enabling code and other
20
+ elements of the foregoing distributed by Meta at ai.meta.com/resources/models-and-
21
+ libraries/llama-downloads/.
22
+
23
+ "Llama Materials" means, collectively, Meta's proprietary Llama 2 and
24
+ Documentation (and any portion thereof) made available under this Agreement.
25
+
26
+ "Meta" or "we" means Meta Platforms Ireland Limited (if you are located in or, if you
27
+ are an entity, your principal place of business is in the EEA or Switzerland) and Meta
28
+ Platforms, Inc. (if you are located outside of the EEA or Switzerland).
29
+
30
+ By clicking "I Accept" below or by using or distributing any portion or element of the
31
+ Llama Materials, you agree to be bound by this Agreement.
32
+
33
+ 1. License Rights and Redistribution.
34
+
35
+ a. Grant of Rights. You are granted a non-exclusive, worldwide, non-
36
+ transferable and royalty-free limited license under Meta's intellectual property or
37
+ other rights owned by Meta embodied in the Llama Materials to use, reproduce,
38
+ distribute, copy, create derivative works of, and make modifications to the Llama
39
+ Materials.
40
+
41
+ b. Redistribution and Use.
42
+
43
+ i. If you distribute or make the Llama Materials, or any derivative works
44
+ thereof, available to a third party, you shall provide a copy of this Agreement to such
45
+ third party.
46
+ ii. If you receive Llama Materials, or any derivative works thereof, from
47
+ a Licensee as part of an integrated end user product, then Section 2 of this
48
+ Agreement will not apply to you.
49
+
50
+ iii. You must retain in all copies of the Llama Materials that you
51
+ distribute the following attribution notice within a "Notice" text file distributed as a
52
+ part of such copies: "Llama 2 is licensed under the LLAMA 2 Community License,
53
+ Copyright (c) Meta Platforms, Inc. All Rights Reserved."
54
+
55
+ iv. Your use of the Llama Materials must comply with applicable laws
56
+ and regulations (including trade compliance laws and regulations) and adhere to the
57
+ Acceptable Use Policy for the Llama Materials (available at
58
+ https://ai.meta.com/llama/use-policy), which is hereby incorporated by reference into
59
+ this Agreement.
60
+
61
+ v. You will not use the Llama Materials or any output or results of the
62
+ Llama Materials to improve any other large language model (excluding Llama 2 or
63
+ derivative works thereof).
64
+
65
+ 2. Additional Commercial Terms. If, on the Llama 2 version release date, the
66
+ monthly active users of the products or services made available by or for Licensee,
67
+ or Licensee's affiliates, is greater than 700 million monthly active users in the
68
+ preceding calendar month, you must request a license from Meta, which Meta may
69
+ grant to you in its sole discretion, and you are not authorized to exercise any of the
70
+ rights under this Agreement unless or until Meta otherwise expressly grants you
71
+ such rights.
72
+
73
+ 3. Disclaimer of Warranty. UNLESS REQUIRED BY APPLICABLE LAW, THE
74
+ LLAMA MATERIALS AND ANY OUTPUT AND RESULTS THEREFROM ARE
75
+ PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND,
76
+ EITHER EXPRESS OR IMPLIED, INCLUDING, WITHOUT LIMITATION, ANY
77
+ WARRANTIES OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR
78
+ FITNESS FOR A PARTICULAR PURPOSE. YOU ARE SOLELY RESPONSIBLE
79
+ FOR DETERMINING THE APPROPRIATENESS OF USING OR REDISTRIBUTING
80
+ THE LLAMA MATERIALS AND ASSUME ANY RISKS ASSOCIATED WITH YOUR
81
+ USE OF THE LLAMA MATERIALS AND ANY OUTPUT AND RESULTS.
82
+
83
+ 4. Limitation of Liability. IN NO EVENT WILL META OR ITS AFFILIATES BE
84
+ LIABLE UNDER ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, TORT,
85
+ NEGLIGENCE, PRODUCTS LIABILITY, OR OTHERWISE, ARISING OUT OF THIS
86
+ AGREEMENT, FOR ANY LOST PROFITS OR ANY INDIRECT, SPECIAL,
87
+ CONSEQUENTIAL, INCIDENTAL, EXEMPLARY OR PUNITIVE DAMAGES, EVEN
88
+ IF META OR ITS AFFILIATES HAVE BEEN ADVISED OF THE POSSIBILITY OF
89
+ ANY OF THE FOREGOING.
90
+
91
+ 5. Intellectual Property.
92
+
93
+ a. No trademark licenses are granted under this Agreement, and in
94
+ connection with the Llama Materials, neither Meta nor Licensee may use any name
95
+ or mark owned by or associated with the other or any of its affiliates, except as
96
+ required for reasonable and customary use in describing and redistributing the
97
+ Llama Materials.
98
+
99
+ b. Subject to Meta's ownership of Llama Materials and derivatives made by or
100
+ for Meta, with respect to any derivative works and modifications of the Llama
101
+ Materials that are made by you, as between you and Meta, you are and will be the
102
+ owner of such derivative works and modifications.
103
+
104
+ c. If you institute litigation or other proceedings against Meta or any entity
105
+ (including a cross-claim or counterclaim in a lawsuit) alleging that the Llama
106
+ Materials or Llama 2 outputs or results, or any portion of any of the foregoing,
107
+ constitutes infringement of intellectual property or other rights owned or licensable
108
+ by you, then any licenses granted to you under this Agreement shall terminate as of
109
+ the date such litigation or claim is filed or instituted. You will indemnify and hold
110
+ harmless Meta from and against any claim by any third party arising out of or related
111
+ to your use or distribution of the Llama Materials.
112
+
113
+ 6. Term and Termination. The term of this Agreement will commence upon your
114
+ acceptance of this Agreement or access to the Llama Materials and will continue in
115
+ full force and effect until terminated in accordance with the terms and conditions
116
+ herein. Meta may terminate this Agreement if you are in breach of any term or
117
+ condition of this Agreement. Upon termination of this Agreement, you shall delete
118
+ and cease use of the Llama Materials. Sections 3, 4 and 7 shall survive the
119
+ termination of this Agreement.
120
+
121
+ 7. Governing Law and Jurisdiction. This Agreement will be governed and
122
+ construed under the laws of the State of California without regard to choice of law
123
+ principles, and the UN Convention on Contracts for the International Sale of Goods
124
+ does not apply to this Agreement. The courts of California shall have exclusive
125
+ jurisdiction of any dispute arising out of this Agreement.
126
+
MODEL-CARD-META-LLAMA-2.md ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # **Llama 2 Model Details**
2
+ Note: Use of this model is governed by the Meta license.
3
+
4
+ Meta developed and publicly released the Llama 2 family of large language models (LLMs), a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. Our fine-tuned LLMs, called Llama-2-Chat, are optimized for dialogue use cases. Llama-2-Chat models outperform open-source chat models on most benchmarks we tested, and in our human evaluations for helpfulness and safety, are on par with some popular closed-source models like ChatGPT and PaLM. We provide a detailed description of our approach to fine-tuning and safety improvements of Llama-2-Chat in order to enable the community to build on our work and contribute to the responsible development of LLMs.
5
+
6
+ **Model Developers** Meta AI
7
+
8
+ **Variations** Llama 2 comes in a range of parameter sizes — 7B, 13B, and 70B — as well as pretrained and fine-tuned variations.
9
+
10
+ **Input** Models input text only.
11
+
12
+ **Output** Models generate text only.
13
+
14
+ **Model Architecture** Llama 2 is an auto-regressive language optimized transformer. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align to human preferences for helpfulness and safety.
15
+
16
+
17
+ ||Training Data|Params|Content Length|GQA|Tokens|LR|
18
+ |---|---|---|---|---|---|---|
19
+ |Llama 2|*A new mix of publicly available online data*|7B|4k|&#10007;|2.0T|3.0 x 10<sup>-4</sup>|
20
+ |Llama 2|*A new mix of publicly available online data*|13B|4k|&#10007;|2.0T|3.0 x 10<sup>-4</sup>|
21
+ |Llama 2|*A new mix of publicly available online data*|70B|4k|&#10004;|2.0T|1.5 x 10<sup>-4</sup>|
22
+
23
+ *Llama 2 family of models.* Token counts refer to pretraining data only. All models are trained with a global batch-size of 4M tokens. Bigger models - 70B -- use Grouped-Query Attention (GQA) for improved inference scalability.
24
+
25
+ **Model Dates** Llama 2 was trained between January 2023 and July 2023.
26
+
27
+ **Status** This is a static model trained on an offline dataset. Future versions of the tuned models will be released as we improve model safety with community feedback.
28
+
29
+ **License** A custom commercial license is available. Please see the Artifacts tab.
30
+
31
+ **Where to send questions or comments about the model** Instructions on how to provide feedback or comments on the model can be found in the model README, or by opening an issue in the GitHub repository.
32
+
33
+ # **Intended Use**
34
+ **Intended Use Cases** Llama 2 is intended for commercial and research use in English. Tuned models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks.
35
+
36
+ **Out-of-scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws).Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Llama 2.
37
+
38
+ # **Hardware and Software**
39
+ **Training Factors** We used custom training libraries, Meta's Research Super Cluster, and production clusters for pretraining. Fine-tuning, annotation, and evaluation were also performed on third-party cloud compute.
40
+
41
+ **Carbon Footprint** Pretraining utilized a cumulative 3.3M GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 539 tCO2eq, 100% of which were offset by Meta’s sustainability program.
42
+
43
+ ||Time (GPU hours)|Power Consumption (W)|Carbon Emitted(tCO<sub>2</sub>eq)|
44
+ |---|---|---|---|
45
+ |Llama 2 7B|184320|400|31.22|
46
+ |Llama 2 13B|368640|400|62.44|
47
+ |Llama 2 70B|1720320|400|291.42|
48
+ |Total|3311616||539.00|
49
+
50
+ **CO<sub>2</sub> emissions during pretraining.** Time: total GPU time required for training each model. Power Consumption: peak power capacity per GPU device for the GPUs used adjusted for power usage efficiency. 100% of the emissions are directly offset by Meta's sustainability program, and because we are openly releasing these models, the pretraining costs do not need to be incurred by others.
51
+
52
+ # **Training Data**
53
+ **Overview** Llama 2 was pretrained on 2 trillion tokens of data from publicly available sources. The fine-tuning data includes publicly available instruction datasets, as well as over one million new human-annotated examples. Neither the pretraining nor the fine-tuning datasets include Meta user data.
54
+
55
+ **Data Freshness** The pretraining data has a cutoff of September 2022, but some tuning data is more recent, up to July 2023.
56
+
57
+ # **Evaluation Results**
58
+
59
+ In this section, we report the results for the Llama 1 and Llama 2 models on standard academic benchmarks.For all the evaluations, we use our internal evaluations library.
60
+
61
+ |Model|Size|Code|Commonsense Reasoning|World Knowledge|Reading Comprehension|Math|MMLU|BBH|AGI Eval|
62
+ |---|---|---|---|---|---|---|---|---|---|
63
+ |Llama 1|7B|14.1|60.8|46.2|58.5|6.95|35.1|30.3|23.9|
64
+ |Llama 1|13B|18.9|66.1|52.6|62.3|10.9|46.9|37.0|33.9|
65
+ |Llama 1|33B|26.0|70.0|58.4|67.6|21.4|57.8|39.8|41.7|
66
+ |Llama 1|65B|30.7|70.7|60.5|68.6|30.8|63.4|43.5|47.6|
67
+ |Llama 2|7B|16.8|63.9|48.9|61.3|14.6|45.3|32.6|29.3|
68
+ |Llama 2|13B|24.5|66.9|55.4|65.8|28.7|54.8|39.4|39.1|
69
+ |Llama 2|70B|**37.5**|**71.9**|**63.6**|**69.4**|**35.2**|**68.9**|**51.2**|**54.2**|
70
+
71
+ **Overall performance on grouped academic benchmarks.** *Code:* We report the average pass@1 scores of our models on HumanEval and MBPP. *Commonsense Reasoning:* We report the average of PIQA, SIQA, HellaSwag, WinoGrande, ARC easy and challenge, OpenBookQA, and CommonsenseQA. We report 7-shot results for CommonSenseQA and 0-shot results for all other benchmarks. *World Knowledge:* We evaluate the 5-shot performance on NaturalQuestions and TriviaQA and report the average. *Reading Comprehension:* For reading comprehension, we report the 0-shot average on SQuAD, QuAC, and BoolQ. *MATH:* We report the average of the GSM8K (8 shot) and MATH (4 shot) benchmarks at top 1.
72
+
73
+ |||TruthfulQA|Toxigen|
74
+ |---|---|---|---|
75
+ |Llama 1|7B|27.42|23.00|
76
+ |Llama 1|13B|41.74|23.08|
77
+ |Llama 1|33B|44.19|22.57|
78
+ |Llama 1|65B|48.71|21.77|
79
+ |Llama 2|7B|33.29|**21.25**|
80
+ |Llama 2|13B|41.86|26.10|
81
+ |Llama 2|70B|**50.18**|24.60|
82
+
83
+ **Evaluation of pretrained LLMs on automatic safety benchmarks.** For TruthfulQA, we present the percentage of generations that are both truthful and informative (the higher the better). For ToxiGen, we present the percentage of toxic generations (the smaller the better).
84
+
85
+
86
+ |||TruthfulQA|Toxigen|
87
+ |---|---|---|---|
88
+ |Llama-2-Chat|7B|57.04|**0.00**|
89
+ |Llama-2-Chat|13B|62.18|**0.00**|
90
+ |Llama-2-Chat|70B|**64.14**|0.01|
91
+
92
+ **Evaluation of fine-tuned LLMs on different safety datasets.** Same metric definitions as above.
93
+
94
+ # **Ethical Considerations and Limitations**
95
+ Llama 2 is a new technology that carries risks with use. Testing conducted to date has not, and could not, cover all scenarios, including uses in languages other than English. For these reasons, as with all LLMs, Llama 2’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Llama 2, developers should perform safety testing and tuning tailored to their specific applications of the model.
96
+
97
+ Please see the Responsible Use Guide available at [https://ai.meta.com/llama/responsible-use-guide/](ai.meta.com/llama/responsible-use-guide)
ONNX/Constant_69_attr__value ADDED
Binary file (262 kB). View file
 
ONNX/Constant_77_attr__value ADDED
Binary file (262 kB). View file
 
ONNX/LlamaV2_7B_FT_float16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:049c978e0ceccd635887c5dd91421599890a83636134da33df5810d17ca5ddf8
3
+ size 4395211
ONNX/embeddings.data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:054c038df2610a9e5b610029e5930094eec5c822192a0f5d642dd80f14c351bd
3
+ size 262144000
ONNX/onnx__MatMul_20670 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64190d5eff94f3693875fb63652d3a52269aaa7a16674fc986a85765c0bc1880
3
+ size 33554432
ONNX/onnx__MatMul_20671 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e06ee03081ae2e1e3493075e83d64310bd8d25bd34114681052f24e20582401
3
+ size 33554432
ONNX/onnx__MatMul_20672 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fac087108df0910ff5b0527849b3f37b84302736b863488e62d548c1e072c8ab
3
+ size 33554432
ONNX/onnx__MatMul_20692 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c599e5c079152d13e57c6ec55cc80486f8b09eaf9af57f956638b7d581ff449
3
+ size 33554432
ONNX/onnx__MatMul_20693 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52685c44e8fc8b3c3570906dee77127c97f998051d7ba8745c14c3a6606d45dc
3
+ size 90177536
ONNX/onnx__MatMul_20694 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3a93bd2c382568584e52c538dc3ba86163d084845e84c9b53e44a1fb721bd44
3
+ size 90177536
ONNX/onnx__MatMul_20695 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7b8586f3159e5fe11e414d297a25a0e47f75aa3633e85336bc0132a237288aa
3
+ size 90177536
ONNX/onnx__MatMul_20696 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f65f9aea6b7a4b5710b67fa321b6387d5f68344d124ca2b4656b44a857d291e
3
+ size 33554432
ONNX/onnx__MatMul_20697 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2fbe2bc942e428759901db09df39f2aa4d98fd1853dc616711078d0bf935352
3
+ size 33554432
ONNX/onnx__MatMul_20698 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee3d36a4011f90f0d38664473eb155340776020cc1eda88b2f23ac6bfcca8b7b
3
+ size 33554432
ONNX/onnx__MatMul_20718 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5bc48c6fa063f7f758359b188b28ea4cdc33f8213ff52e4d2c301585d6ec87d
3
+ size 33554432
ONNX/onnx__MatMul_20719 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:668b974e2bc62e6bfcf6c5cdb5d0cbf4f75192e5ce23d79a0ddd03f8ca03f8af
3
+ size 90177536
ONNX/onnx__MatMul_20720 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e74d2ea98b0d4ffb9dc546b3e0e5c9abb2ed17d7cecee81587cc3743f02b1ef8
3
+ size 90177536
ONNX/onnx__MatMul_20721 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1e50ec83e75aea91248422926a90a43e7e65fd32352d270069f72392106c0d
3
+ size 90177536
ONNX/onnx__MatMul_20722 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce3257c553ece92d49578122805c8b7216752e125449633cf7aa1517e34f1514
3
+ size 33554432
ONNX/onnx__MatMul_20723 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78b46afd6677f4b14ff3a5cb94df46ed8257e806cdb2ff05a7621613c673423b
3
+ size 33554432
ONNX/onnx__MatMul_20724 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d887bf27ead6f74edeb3f59144ad572221c9a523ff1b00ccdcea9ee309c8b163
3
+ size 33554432
ONNX/onnx__MatMul_20744 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e3cda94d1766f1aeefc5414b7a8fd8a488b2a050e65d07f15d8c3ed336997d3
3
+ size 33554432
ONNX/onnx__MatMul_20745 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce35f107461923cb7a2202e9afcbd0c8ac134d004207775e37058c3ce16af61b
3
+ size 90177536
ONNX/onnx__MatMul_20746 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c88d286e5ddce59fb14a51e8d718b1881227d55fd612fc2a26c542a78a0199d4
3
+ size 90177536
ONNX/onnx__MatMul_20747 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e823cc13de98c77d31bac6a75412af911b52de00c8771fc590ab1f010c3350b
3
+ size 90177536
ONNX/onnx__MatMul_20748 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f915c17fe192b091565e2ea75b5e59ac5e580827decee3210788c521de7377c
3
+ size 33554432
ONNX/onnx__MatMul_20749 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b01b0a35673ee77852a652a0838c6c8d91b7bd1fb5373a328defc842599fd33
3
+ size 33554432
ONNX/onnx__MatMul_20750 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b9f91ff4c9a073e4b7cbbd28d629459f5fe9ad1579b452bbdce3700e5583023
3
+ size 33554432
ONNX/onnx__MatMul_20770 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e925bef2d5069e9d9462dca8d386a3dcba2d6c2e3d33130a9843bab756bc7cff
3
+ size 33554432
ONNX/onnx__MatMul_20771 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ecf8486634b5a9d2610c03db52c7218efe2003cc14af4efb9a4e4c62763102f
3
+ size 90177536
ONNX/onnx__MatMul_20772 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0b32f72be8b8c4f25ad6cecfe1697a32e58b1ea9bbecaa9efad9e9482dd2509
3
+ size 90177536
ONNX/onnx__MatMul_20773 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e68ef4a4fb921bdff349ddf8a1e163d281e683d98632d1b115e4de8d6383f40e
3
+ size 90177536
ONNX/onnx__MatMul_20774 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cf4553c17bd0de8b91d7d357da45ed667e42bb6eaf7837a684c726ddd243800
3
+ size 33554432
ONNX/onnx__MatMul_20775 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e719aa17dbe5c97ab1631cda38b3a9c82a627d770e80acd964ea408be082e242
3
+ size 33554432
ONNX/onnx__MatMul_20776 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa1b6593684f71302e72304e2d26c08b37471cc7f835d2522b69a89780fe36d0
3
+ size 33554432
ONNX/onnx__MatMul_20796 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab134c44360b3d320b72be7a179f0e5098b5cdb34519077328474690576dd93d
3
+ size 33554432
ONNX/onnx__MatMul_20797 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6af301790daebbfbe8dec897e8bffb221e36ea0ae183c4f3dfd0f307289a826
3
+ size 90177536
ONNX/onnx__MatMul_20798 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c906f2a2b1afa87a71fa1929aff06f31ca42de21e19333ef0eff1d3769178294
3
+ size 90177536
ONNX/onnx__MatMul_20799 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b4949eadd4631602af20d7ef62050021978300870ee745b7badf91d72786c42
3
+ size 90177536
ONNX/onnx__MatMul_20800 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38a513113e233c4d0e4bde6d17cf64b438799d2899d75460e50b850bdb9f9eb
3
+ size 33554432
ONNX/onnx__MatMul_20801 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6611b457af6848badc4153368a4fc874eeb03574ffe7c689ffd4822c576aa68e
3
+ size 33554432
ONNX/onnx__MatMul_20802 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e1a7c3b4fc05f2f05754f314e45068d0e8b721fdbd0bb7639b50839958bda20
3
+ size 33554432
ONNX/onnx__MatMul_20822 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85bf85cd9d61e5f1f2ce35afbb3bd16797411a9f21af3664b2ac380cbcd4970b
3
+ size 33554432
ONNX/onnx__MatMul_20823 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48734c02f84f97023b2ae2a59c5c8ed3dd9077bd64d5133631f4fe4fcfecd77b
3
+ size 90177536
ONNX/onnx__MatMul_20824 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17671d866f3a185a3f925bf862d7282a8377c7044b9f91e2d9ccdd8ecdbcb332
3
+ size 90177536
ONNX/onnx__MatMul_20825 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f50c89e0c1557a0ad94a0ede787cd6b19e3088a0f6ff8705380ff35ab2ab0455
3
+ size 90177536