Problem with deleting supposed .pyabsa folder #415
Unanswered
luisferpazos
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
**Hello Community, I’m Luis Fernando Pazos. Im a begginer Python developer with experience in RPA field in low code tools. Now Im trying to learn more about Python, machine learning, RAG, etc. Happy to have found this space to hopefully solve some of my problems while trying to develop some projects.
For now I’m trying to work with PyABSA for Aspect Based Sentiment Analysis. For now im just going through the examples folder to learn about the library and its functions, etc.
This is the first code Im trying to execute**
and im getting this error:
Traceback (most recent call last): File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\pyabsa\tasks\AspectPolarityClassification\prediction\sentiment_classifier.py”, line 83, in init self.model = APCEnsembler( ^^^^^^^^^^^^^ File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\pyabsa\tasks\AspectPolarityClassification\instructor\ensembler.py”, line 79, in init [ File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\pyabsa\tasks\AspectPolarityClassification\instructor\ensembler.py”, line 80, in str(self.config.args[k]) File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\transformers\tokenization_utils_base.py”, line 1511, in repr f" special_tokens={self.special_tokens_map}, clean_up_tokenization_spaces={self.clean_up_tokenization_spaces}," ^^^^^^^^^^^^^^^^^^^^^^^ File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\transformers\tokenization_utils_base.py”, line 1111, in getattr raise AttributeError(f"{self.class .name } has no attribute {key}") AttributeError: DebertaV2TokenizerFast has no attribute special_tokens_map. Did you mean: ‘get_special_tokens_mask’? During handling of the above exception, another exception occurred: Traceback (most recent call last): File “C:\Users\pazos\Documents\vs\demo\PYABSA\examples-v2\aspect_polarity_classification\inference.py”, line 15, in sent_classifier = APC.SentimentClassifier(“multilingual”, max_seq_len=512) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File “C:\Users\pazos\Documents\vs\demo\pyabsa_env\Lib\site-packages\pyabsa\tasks\AspectPolarityClassification\prediction\sentiment_classifier.py”, line 104, in init raise RuntimeError( RuntimeError: Fail to load the model from multilingual! Please make sure the version of checkpoint and PyABSA are compatible. Try to remove he checkpoint and download again Exception: DebertaV2TokenizerFast has no attribute special_tokens_map
**I asked to Claude and said that the error could be solve by downgrading the Transaformers, which I did to 4.29.0.
And also clear the PyABSA cache which it says that it shoul be on some folder called .pyabsa.
I have search for this folder in my system but haven´t found it.
Any suggestion?**
Beta Was this translation helpful? Give feedback.
All reactions