All Models

All Models

Gemma 4 31B IT

Gemma 4 31B IT

Text

Gemma 4 31B IT

Large instruction-tuned model built for strong general-purpose text and chat tasks.

  • High Quality Responses. 31B parameter model tuned for instruction following, chat, and structured outputs. Reliable across a wide range of everyday tasks.

  • Balanced Intelligence. Strong mix of reasoning, knowledge, and writing ability without needing agent-style prompting or special modes.

  • Production Ready. Stable behavior, predictable outputs, and easy to integrate into standard chat or API workflows.

  • Apache 2.0 License. Fully open and permissive - safe for commercial use without restrictions.

Why pick it for Norman AI?

Gemma 4 31B IT is a solid default LLM. Use it when you need consistent, high-quality responses for chat, assistants, or general automation without optimizing for heavy reasoning or agents.

messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-31b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-31b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-31b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)
messages = [
    {"role": "system", "content": "You are a helpful AI assistant."},
    {"role": "user", "content": "Can you provide ways to eat combinations of bananas and dragonfruits?"},
    {"role": "assistant",
     "content": "Sure! Here are some ways to eat bananas and dragonfruits together"},
    {"role": "user", "content": "What about solving an 2x + 3 = 7 equation?"},
]

response = await norman.invoke(
    {
        "model_name": "gemma-4-31b-it",
        "inputs": [
            {
                "display_title": "Prompt",
                "data": messages
            }
        ]
    }
)

·

©

2026