You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I am running Oscar with torch 2.1 but am facing this error. Could you please take a look and suggest a solution?
Traceback (most recent call last):
File "./Oscar/oscar/run_vqa.py", line 1232, in
main()
File "./Oscar/oscar/run_vqa.py", line 1155, in main
global_step, tr_loss = train(args, train_dataset, eval_dataset, model, tokenizer)
File "./Oscar/oscar/run_vqa.py", line 477, in train
optimizer = AdamW(optimizer_grouped_parameters, lr=args.learning_rate, eps=args.adam_epsilon)
File "./Oscar/./transformers/pytorch_transformers/optimization.py", line 128, in init
super(AdamW, self).init(params, defaults)
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/optim/optimizer.py", line 266, in init
self.add_param_group(cast(dict, param_group))
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_compile.py", line 22, in inner
import torch._dynamo
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/init.py", line 2, in
from . import allowed_functions, convert_frame, eval_frame, resume_execution
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/convert_frame.py", line 46, in
from .output_graph import OutputGraph
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/output_graph.py", line 35, in
from . import config, logging as torchdynamo_logging, variables
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/init.py", line 53, in
from .torch import TorchVariable
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/torch.py", line 131, in
transformers.configuration_utils.PretrainedConfig.eq = (
AttributeError: module 'transformers' has no attribute 'configuration_utils'
The text was updated successfully, but these errors were encountered:
Hi,
I am running Oscar with torch 2.1 but am facing this error. Could you please take a look and suggest a solution?
Traceback (most recent call last):
File "./Oscar/oscar/run_vqa.py", line 1232, in
main()
File "./Oscar/oscar/run_vqa.py", line 1155, in main
global_step, tr_loss = train(args, train_dataset, eval_dataset, model, tokenizer)
File "./Oscar/oscar/run_vqa.py", line 477, in train
optimizer = AdamW(optimizer_grouped_parameters, lr=args.learning_rate, eps=args.adam_epsilon)
File "./Oscar/./transformers/pytorch_transformers/optimization.py", line 128, in init
super(AdamW, self).init(params, defaults)
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/optim/optimizer.py", line 266, in init
self.add_param_group(cast(dict, param_group))
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_compile.py", line 22, in inner
import torch._dynamo
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/init.py", line 2, in
from . import allowed_functions, convert_frame, eval_frame, resume_execution
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/convert_frame.py", line 46, in
from .output_graph import OutputGraph
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/output_graph.py", line 35, in
from . import config, logging as torchdynamo_logging, variables
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/init.py", line 53, in
from .torch import TorchVariable
File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/torch.py", line 131, in
transformers.configuration_utils.PretrainedConfig.eq = (
AttributeError: module 'transformers' has no attribute 'configuration_utils'
The text was updated successfully, but these errors were encountered: