utkarshshukla2912 commited on
Commit
f840a48
·
verified ·
1 Parent(s): 60b9e9b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md CHANGED
@@ -25,10 +25,41 @@ You can plug it into your calling or voice AI stack to automatically extract:
25
 
26
  It’s built to handle real-world Hindi, English, and mixed Hinglish calls, including noisy transcripts.
27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
  - **Developed by:** RinggAI
30
  - **License:** apache-2.0
31
  - **Finetuned from model :** unsloth/Qwen2.5-1.5B-Instruct
 
 
 
 
 
 
32
 
33
 
34
 
 
25
 
26
  It’s built to handle real-world Hindi, English, and mixed Hinglish calls, including noisy transcripts.
27
 
28
+ Finetuning Parameters:
29
+ ```
30
+ rank = 64
31
+ lora_alpha = rank*2,
32
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
33
+ "gate_proj", "up_proj", "down_proj",],
34
+ SFTConfig(
35
+ dataset_text_field = "prompt",
36
+ per_device_train_batch_size = 32,
37
+ gradient_accumulation_steps = 1, # Use GA to mimic batch size!
38
+ warmup_steps = 5,
39
+ num_train_epochs = 2,
40
+ learning_rate = 2e-4,
41
+ logging_steps = 50,
42
+ optim = "adamw_8bit",
43
+ weight_decay = 0.001,
44
+ lr_scheduler_type = "linear",
45
+ seed = SEED,
46
+ report_to = "wandb",
47
+ eval_strategy="steps",
48
+ eval_steps=200,
49
+ )
50
+ The model was finetuned on ~100,000 curated transcripts across different domanins and language preferences
51
+ ```
52
+ ![Training Overview](metrics.png)
53
 
54
  - **Developed by:** RinggAI
55
  - **License:** apache-2.0
56
  - **Finetuned from model :** unsloth/Qwen2.5-1.5B-Instruct
57
+ - Parameter decision where made using
58
+ **Schulman, J., & Thinking Machines Lab. (2025).**
59
+ *LoRA Without Regret.*
60
+ Thinking Machines Lab: Connectionism.
61
+ DOI: 10.64434/tml.20250929
62
+ Link: https://thinkingmachines.ai/blog/lora/
63
 
64
 
65