-
|
Running private gpt with recommended setup ("ui llms-ollama embeddings-ollama vector-stores-qdrant") on WSL (Ubuntu, Windows 11, 32 gb RAM, i7, Nvidia GeForce RTX 4060 ). LLM Chat (no context from files) works well. When trying to upload a small (1Kb) text file it stucks either on 0% while generating embeddings. |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments
-
|
Had the same issue with ollama, it just wasn't generating the embeddings on WSL. Worked fine when I switched to llama.cpp though. |
Beta Was this translation helpful? Give feedback.
-
|
@d1g33k if I try this for llama.cpp: I am running on ubuntu using a docker container |
Beta Was this translation helpful? Give feedback.
-
|
If you are using |
Beta Was this translation helpful? Give feedback.
-
|
In my case, I don't use embedding:
mode: gemini
# Should be matching the value above in most cases
ingest_mode:
# simple
pipeline # the fastest mode according to docs
count_workers:
# 4
# 8
16 # depends on your machine
embed_dim: 768
gemini:
api_key: ${GOOGLE_API_KEY:}
model:
models/gemini-1.5-pro
# models/gemini-1.5-flash
embedding_model: models/text-embedding-004To solve the slow ingestion for this file, I split each row of the tabular data into a single-row table file. Then I ingest all those files at once. And the performance of file ingestion for this whole tabular data really speeds up. Another thing I think we should investigate is the Just hope these notes could help someone out there using this great project. |
Beta Was this translation helpful? Give feedback.
Had the same issue with ollama, it just wasn't generating the embeddings on WSL. Worked fine when I switched to llama.cpp though.