functiongemma-270m-mika

ๅŸบไบŽ lmstudio-community/functiongemma-270m-it-MLX-bf16 ไฟฎๆ”นๅ‘ๅธƒใ€‚

ๆžถๆž„ไฟกๆฏ

ๅ‚ๆ•ฐ ๅ€ผ
ๆžถๆž„ Gemma3ForCausalLM
ๅ‚ๆ•ฐ้‡ ~270M
Hidden Size 640
ๅฑ‚ๆ•ฐ 18 (sliding + full attentionๆททๅˆ)
่ฏ่กจๅคงๅฐ 262,144
ๆœ€ๅคงไธŠไธ‹ๆ–‡ 32,768 tokens
็ฒพๅบฆ bfloat16

ไฟฎๆ”นๅ†…ๅฎน

  • โœ… ไผ˜ๅŒ–ไบ†็”Ÿๆˆๅ‚ๆ•ฐ้…็ฝฎ๏ผˆtemperature/top_p/top_k๏ผ‰
  • โœ… ๆ–ฐๅขž generation_config.json
  • โœ… ไผ˜ๅŒ–ไบ†ๆ”ฏๆŒ Function Calling ็š„ Chat Template
  • โœ… ๆๅ‡ max_length ่‡ณ 8192

ๅฟซ้€Ÿๅผ€ๅง‹

MLX ๆ–นๅผ๏ผˆMac Apple Silicon ๆŽจ่๏ผ‰

from mlx_lm import load, generate

model, tokenizer = load("DarylFranxx/functiongemma-270m-mika")

# ๆ™ฎ้€šๅฏน่ฏ
response = generate(model, tokenizer, 
                    prompt="ไฝ ๅฅฝ๏ผŒ่ฏทไป‹็ปไธ€ไธ‹่‡ชๅทฑ", 
                    max_tokens=256)
print(response)

Function Calling ็คบไพ‹

from mlx_lm import load, generate

model, tokenizer = load("DarylFranxx/functiongemma-270m-mika")

tools = [
    {
        "name": "get_weather",
        "description": "่Žทๅ–ๆŒ‡ๅฎšๅŸŽๅธ‚็š„ๅคฉๆฐ”ไฟกๆฏ",
        "parameters": {
            "type": "object",
            "properties": {
                "city": {"type": "string", "description": "ๅŸŽๅธ‚ๅ็งฐ"},
                "date": {"type": "string", "description": "ๆ—ฅๆœŸ๏ผŒๆ ผๅผYYYY-MM-DD"}
            },
            "required": ["city"]
        }
    }
]

messages = [
    {"role": "user", "content": "ๅŒ—ไบฌไปŠๅคฉๅคฉๆฐ”ๆ€Žไนˆๆ ท๏ผŸ"}
]

# ไฝฟ็”จtokenizer็š„chat_templateๆ ผๅผๅŒ–
prompt = tokenizer.apply_chat_template(
    messages,
    tools=tools,
    tokenize=False,
    add_generation_prompt=True
)

response = generate(model, tokenizer, prompt=prompt, max_tokens=512)
print(response)

Transformers ๆ–นๅผ

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "DarylFranxx/functiongemma-270m-mika"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto"
)

messages = [{"role": "user", "content": "Hello!"}]
inputs = tokenizer.apply_chat_template(
    messages, return_tensors="pt", add_generation_prompt=True
).to(model.device)

outputs = model.generate(inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0][inputs.shape[-1]:], skip_special_tokens=True))

ๆณจๆ„ไบ‹้กน

  • ้œ€่ฆ transformers >= 4.57.3 ๆ‰ๆ”ฏๆŒ Gemma3ForCausalLM
  • MLXๆ ผๅผไป…้€‚็”จไบŽ Apple Silicon (M1/M2/M3/M4)
  • ้ตๅฎˆๅŽŸๅง‹ๆจกๅž‹ Apache 2.0 ่ฎธๅฏ่ฏ
Downloads last month
45
Safetensors
Model size
0.3B params
Tensor type
BF16
ยท
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DarylFranxx/functiongemma-270m-mika

Space using DarylFranxx/functiongemma-270m-mika 1