| |
|
| | --- |
| | license: llama3.2 |
| | base_model: meta-llama/Llama-3.2-3B-Instruct |
| | tags: |
| | - function-calling |
| | - llama3.2 |
| | - fine-tuned |
| | - lora |
| | language: |
| | - en |
| | --- |
| | |
| | # Llama 3.2 3B Function Calling Model |
| |
|
| | This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) for function calling tasks. |
| |
|
| | ## Model Details |
| |
|
| | - **Base Model**: Llama 3.2 3B Instruct |
| | - **Fine-tuning Method**: LoRA (Low-Rank Adaptation) |
| | - **Dataset**: Salesforce/xlam-function-calling-60k (1000 samples) |
| | - **Training**: 2 epochs with learning rate 2e-5 |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | import torch |
| | |
| | model = AutoModelForCausalLM.from_pretrained("TurkishCodeMan/llama3.2-3b-intruct-function-calling") |
| | tokenizer = AutoTokenizer.from_pretrained("TurkishCodeMan/llama3.2-3b-intruct-function-calling") |
| | |
| | prompt = '''<|system|> |
| | Available functions: |
| | - get_weather: Gets current weather for a location |
| | |
| | GPT 4 Correct user: |
| | <|user|> |
| | What's the weather in Tokyo? |
| | GPT 4 correct assistant:''' |
| | |
| | inputs = tokenizer(prompt, return_tensors="pt") |
| | outputs = model.generate(**inputs, max_new_tokens=64, do_sample=False) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | ## Training Details |
| |
|
| | - **Learning Rate**: 2e-5 |
| | - **Batch Size**: 2 (per device) |
| | - **Gradient Accumulation**: 8 steps |
| | - **LoRA Rank**: 8 |
| | - **LoRA Alpha**: 16 |
| | - **Target Modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj |
| | |
| | ## Performance |
| | |
| | The model demonstrates excellent function calling capabilities: |
| | - Correct function selection |
| | - Proper argument formatting |
| | - Professional response structure |
| | |