All Models

All Models

Gemma 4 E4B IT

Gemma 4 E4B IT

Text

Gemma 4 E4B IT

Small, efficient instruction model built for low-cost and fast inference.

  • Lightweight Model. ~4B-scale model designed for speed and minimal resource usage.

  • Good Enough Quality. Handles common tasks like chat, summaries, and simple reasoning without heavy compute.

  • Fast Inference. Low latency makes it suitable for real-time applications and edge deployments.

  • Apache 2.0 License. Fully open and safe for commercial use.

Why pick it for Norman AI?

Gemma 4 E4B IT is for high-throughput, low-cost workloads. Use it when you need something fast and cheap that still produces usable outputs.

messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-e4b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-e4b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-e4b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-e4b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)

·

©

2026