You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First, I want to preface that I'm not an expert with BERTopic's internals, so I apologize if I'm misunderstanding something fundamental.
I'm running into an issue where I have limited VRAM and want to use BERTopic with an LLM for topic labeling. My workflow involves:
Pre-computing embeddings using a large embedding model
Saving these embeddings to disk
Loading them back for topic modeling
The challenge I'm facing is that when I try to initialize BERTopic with embedding_model=None while providing pre-computed embeddings, I get an error. This means I have to keep the large embedding model in memory even though I've already generated the embeddings.
Here's a minimal example of my current workaround using a dummy embedder:
The challenge I'm facing is that when I try to initialize BERTopic with embedding_model=None while providing pre-computed embeddings, I get an error. This means I have to keep the large embedding model in memory even though I've already generated the embeddings.
Before talking about solutions, it's important that we have a clear picture of this error that you mention first. The thing is, you can use embedding_model=None with pre-computed embeddings. The only thing that is stopping you is what is in representation_models that might make use of word-level embeddings (such as KeyBERTInspired).
First, I want to preface that I'm not an expert with BERTopic's internals, so I apologize if I'm misunderstanding something fundamental.
I'm running into an issue where I have limited VRAM and want to use BERTopic with an LLM for topic labeling. My workflow involves:
The challenge I'm facing is that when I try to initialize BERTopic with embedding_model=None while providing pre-computed embeddings, I get an error. This means I have to keep the large embedding model in memory even though I've already generated the embeddings.
Here's a minimal example of my current workaround using a dummy embedder:
Question
The text was updated successfully, but these errors were encountered: