diff --git a/gallery/index.yaml b/gallery/index.yaml index 2bbe068e..d3050ef3 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -475,6 +475,27 @@ - filename: Llama-3-Lumimaid-8B-v0.1.q4_k_m.gguf sha256: 23ac0289da0e096d5c00f6614dfd12c94dceecb02c313233516dec9225babbda uri: huggingface://NeverSleep/Llama-3-Lumimaid-8B-v0.1-GGUF/Llama-3-Lumimaid-8B-v0.1.q4_k_m.gguf +- <<: *llama3 + name: "llama-3-lumimaid-8b-v0.1-oas-iq-imatrix" + urls: + - https://huggingface.co/Lewdiculous/Llama-3-Lumimaid-8B-v0.1-OAS-GGUF-IQ-Imatrix + icon: https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/JUxfdTot7v7LTdIGYyzYM.png + license: cc-by-nc-4.0 + description: | + This model uses the Llama3 prompting format. + + Llama3 trained on our RP datasets, we tried to have a balance between the ERP and the RP, not too horny, but just enough. + + We also added some non-RP dataset, making the model less dumb overall. It should look like a 40%/60% ratio for Non-RP/RP+ERP data. + + "This model received the Orthogonal Activation Steering treatment, meaning it will rarely refuse any request." + overrides: + parameters: + model: Llama-3-Lumimaid-8B-v0.1-OAS-Q4_K_M-imat.gguf + files: + - filename: Llama-3-Lumimaid-8B-v0.1-OAS-Q4_K_M-imat.gguf + sha256: 1199440aa13c55f5f2cad1cb215535306f21e52a81de23f80a9e3586c8ac1c50 + uri: huggingface://Lewdiculous/Llama-3-Lumimaid-8B-v0.1-OAS-GGUF-IQ-Imatrix/Llama-3-Lumimaid-8B-v0.1-OAS-Q4_K_M-imat.gguf - <<: *llama3 name: "suzume-llama-3-8B-multilingual" urls: