Tonic commited on
Commit
81f39f1
·
1 Parent(s): b4f1cb3

adds parameters to medical config

Browse files
config/train_gpt_oss_custom.py CHANGED
@@ -203,6 +203,7 @@ class GPTOSSEnhancedCustomConfig:
203
  dataloader_num_workers: int = 4 # Number of data loading workers
204
  dataloader_pin_memory: bool = True # Pin memory for faster GPU transfer
205
  dataloader_prefetch_factor: int = 2 # Prefetch factor for data loading
 
206
 
207
  # Memory Management
208
  max_memory_per_gpu: Optional[str] = None # e.g., "80GB", "40GB"
 
203
  dataloader_num_workers: int = 4 # Number of data loading workers
204
  dataloader_pin_memory: bool = True # Pin memory for faster GPU transfer
205
  dataloader_prefetch_factor: int = 2 # Prefetch factor for data loading
206
+ dataset_num_proc: Optional[int] = None # Parallel CPU processes for datasets map/filter ops
207
 
208
  # Memory Management
209
  max_memory_per_gpu: Optional[str] = None # e.g., "80GB", "40GB"
config/train_gpt_oss_medical_o1_sft.py CHANGED
@@ -78,7 +78,6 @@ config = GPTOSSEnhancedCustomConfig(
78
  dataloader_num_workers=4,
79
  dataloader_pin_memory=True,
80
  dataloader_prefetch_factor=2,
81
- dataset_num_proc=4,
82
  group_by_length=True,
83
  remove_unused_columns=True,
84
 
@@ -87,8 +86,8 @@ config = GPTOSSEnhancedCustomConfig(
87
  # ============================================================================
88
  use_lora=True,
89
  lora_config={
90
- "r": 8,
91
- "lora_alpha": 16,
92
  "lora_dropout": 0.05,
93
  "target_modules": "all-linear",
94
  "target_parameters": [
 
78
  dataloader_num_workers=4,
79
  dataloader_pin_memory=True,
80
  dataloader_prefetch_factor=2,
 
81
  group_by_length=True,
82
  remove_unused_columns=True,
83
 
 
86
  # ============================================================================
87
  use_lora=True,
88
  lora_config={
89
+ "r": 16,
90
+ "lora_alpha": 32,
91
  "lora_dropout": 0.05,
92
  "target_modules": "all-linear",
93
  "target_parameters": [