Llama-3.1-Nemotron-70B-Instruct
/
model_weights
/model.decoder.layers.self_attention.linear_proj.weight
-
258 Bytes
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS
-
16.8 MB
LFS