@@ -23,8 +23,7 @@ def test_build_training_callbacks_mfu_callback():
23
23
config .training .include_performance_metrics = True
24
24
config .data .train .pack = True
25
25
config .model .model_max_length = 128
26
- model = torch .nn .Module ()
27
- model .dtype = torch .bfloat16 # type: ignore
26
+ model = torch .nn .Sequential (torch .nn .Linear (10 , 10 ))
28
27
with patch ("torch.cuda.is_available" , return_value = True ):
29
28
with patch ("torch.cuda.get_device_name" , return_value = "NVIDIA A100-PCIE-40GB" ):
30
29
result = build_training_callbacks (config , model , None )
@@ -38,7 +37,7 @@ def test_build_training_callbacks_mfu_callback():
38
37
def test_build_training_callbacks_no_cuda (mock_logger_warning ):
39
38
config = TrainingConfig ()
40
39
config .training .include_performance_metrics = True
41
- model = torch .nn .Module ( )
40
+ model = torch .nn .Sequential ( torch . nn . Linear ( 10 , 10 ) )
42
41
with patch ("torch.cuda.is_available" , return_value = False ):
43
42
result = build_training_callbacks (config , model , None )
44
43
assert len (result ) == 2
@@ -54,7 +53,7 @@ def test_build_training_callbacks_peft(mock_logger_warning):
54
53
config = TrainingConfig ()
55
54
config .training .include_performance_metrics = True
56
55
config .training .use_peft = True
57
- model = torch .nn .Module ( )
56
+ model = torch .nn .Sequential ( torch . nn . Linear ( 10 , 10 ) )
58
57
with patch ("torch.cuda.is_available" , return_value = True ):
59
58
result = build_training_callbacks (config , model , None )
60
59
assert len (result ) == 2
@@ -69,7 +68,7 @@ def test_build_training_callbacks_peft(mock_logger_warning):
69
68
def test_build_training_callbacks_no_pack (mock_logger_warning ):
70
69
config = TrainingConfig ()
71
70
config .training .include_performance_metrics = True
72
- model = torch .nn .Module ( )
71
+ model = torch .nn .Sequential ( torch . nn . Linear ( 10 , 10 ) )
73
72
with patch ("torch.cuda.is_available" , return_value = True ):
74
73
result = build_training_callbacks (config , model , None )
75
74
assert len (result ) == 2
@@ -85,8 +84,7 @@ def test_build_training_callbacks_unknown_device_name(mock_logger_warning):
85
84
config = TrainingConfig ()
86
85
config .training .include_performance_metrics = True
87
86
config .data .train .pack = True
88
- model = torch .nn .Module ()
89
- model .dtype = torch .bfloat16 # type: ignore
87
+ model = torch .nn .Sequential (torch .nn .Linear (10 , 10 ))
90
88
with patch ("torch.cuda.is_available" , return_value = True ):
91
89
with patch ("torch.cuda.get_device_name" , return_value = "Foo" ):
92
90
result = build_training_callbacks (config , model , None )
@@ -103,8 +101,7 @@ def test_build_training_callbacks_no_model_max_length(mock_logger_warning):
103
101
config = TrainingConfig ()
104
102
config .training .include_performance_metrics = True
105
103
config .data .train .pack = True
106
- model = torch .nn .Module ()
107
- model .dtype = torch .bfloat16 # type: ignore
104
+ model = torch .nn .Sequential (torch .nn .Linear (10 , 10 ))
108
105
with patch ("torch.cuda.is_available" , return_value = True ):
109
106
with patch ("torch.cuda.get_device_name" , return_value = "NVIDIA A100-PCIE-40GB" ):
110
107
result = build_training_callbacks (config , model , None )
0 commit comments