Avg. Total Time
13.53s
Avg. TTFT
4.69s
Avg. Prefill TPS
625.54
Avg. Gen TPS
16.79
Context Size
32768
Quantization
r64
Engine
aphrodite
Creation Method
Unknown
Model Type
Gemma27B
Chat Template
Gemma 2
Reasoning
No
Vision
Yes
Parameters
27B
Added At
6/23/2025
license: gemma library_name: transformers pipeline_tag: image-text-to-text extra_gated_heading: Access Gemma on Hugging Face extra_gated_prompt: >- To access Gemma on Hugging Face, you’re required to review and agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging Face and click below. Requests are processed immediately. extra_gated_button_content: Acknowledge license base_model: google/gemma-3-27b-it tags:
This is an uncensored version of google/gemma-3-27b-it created with abliteration (see remove-refusals-with-transformers to know more about it).
This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
It was only the text part that was processed, not the image part.
The abliterated model will no longer say "I'm sorry, but I cannot fulfill your request to ..."
Ollama supports multimodal (Vision). gemma-3-abliterated defaults to f16, not Q4_K_M, and the effect of Q4_K_M is not very good, nor is it provided.
All new versions of gemma-3-abliterated have been released; please re-download and test.
You can use huihui_ai/gemma3-abliterated directly
ollama run huihui_ai/gemma3-abliterated:27b
You can use this model in your applications by loading it with Hugging Face's transformers library:
# pip install accelerate
from transformers import AutoProcessor, Gemma3ForConditionalGeneration
from PIL import Image
import requests
import torch
model_id = "huihui-ai/gemma-3-27b-it-abliterated"
model = Gemma3ForConditionalGeneration.from_pretrained(
model_id, device_map="auto"
).eval()
processor = AutoProcessor.from_pretrained(model_id)
messages = [
{
"role": "system",
"content": [{"type": "text", "text": "You are a helpful assistant."}]
},
{
"role": "user",
"content": [
{"type": "image", "image": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg"},
{"type": "text", "text": "Describe this image in detail."}
]
}
]
inputs = processor.apply_chat_template(
messages, add_generation_prompt=True, tokenize=True,
return_dict=True, return_tensors="pt"
).to(model.device, dtype=torch.bfloat16)
input_len = inputs["input_ids"].shape[-1]
with torch.inference_mode():
generation = model.generate(**inputs, max_new_tokens=100, do_sample=False)
generation = generation[0][input_len:]
decoded = processor.decode(generation, skip_special_tokens=True)
print(decoded)
# **Overall Impression:** The image is a close-up shot of a vibrant garden scene,
# focusing on a cluster of pink cosmos flowers and a busy bumblebee.
# It has a slightly soft, natural feel, likely captured in daylight.
If you like it, please click 'like' and follow us for more updates.
You can follow x.com/support_huihui to get the latest model information from huihui.ai.
bc1qqnkhuchxw0zqjh2ku3lu4hq45hc6gy84uk70ge