Uh oh!
There was an error while loading.Please reload this page.
- Notifications
You must be signed in to change notification settings - Fork1.9k
-
Hello,
|
BetaWas this translation helpful?Give feedback.
All reactions
❤️ 2
Replies: 6 comments 11 replies
-
By "similar tools", I mean things like:
|
BetaWas this translation helpful?Give feedback.
All reactions
❤️ 1
-
ok, now the question is clearer. |
BetaWas this translation helpful?Give feedback.
All reactions
-
I understand the goal of the project. |
BetaWas this translation helpful?Give feedback.
All reactions
-
yeah, it could be a great idea, thank you! |
BetaWas this translation helpful?Give feedback.
All reactions
-
hi, we did this, let me know if you have more ideaslink |
BetaWas this translation helpful?Give feedback.
All reactions
-
not impressive at all with the first example: |
BetaWas this translation helpful?Give feedback.
All reactions
-
Your ouput is better (starts at the top, continues sequentially) but it still does not include all news. I used Ollama in docker. You should try to make embeddings work with this model, it will be much faster:https://ollama.com/library/nomic-embed-text |
BetaWas this translation helpful?Give feedback.
All reactions
-
Hei there, thank you for the feedback! Regarding the I tried using the model you suggested for the embeddings and it is actually pretty fast! Thanks :) |
BetaWas this translation helpful?Give feedback.
All reactions
-
@ftoppi Try with this |
BetaWas this translation helpful?Give feedback.
All reactions
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
-
using the embedding you are talking about we improve of the 50% the speed, (from 6 to 3 minutes). |
BetaWas this translation helpful?Give feedback.
All reactions
-
You're welcome. The embeddings dict in the config does not seem to work properly. It tries to connect to localhost. |
BetaWas this translation helpful?Give feedback.
All reactions
-
The bigger the model the worse the performance for embedding and CPU instead of GPU would of course be the worst out of all four possible positions of small big CPU GPU. TheMTEBis the way to go when you are choosing performant models for your hardware. Most everyone should be able to make sure the NVIDIA SDK is loaded in the docker and I think most of the projects have auto install if you don't have it. Ola running everything in GPU it's going to be the best. Most of the embedding models are small enough where that should be your best bet. Ollama does have quite several nice embedding models in their list. The Nomic embedding worked quite well for me (they always do good stuff) However I found theMixed Bread model was slightly more performant. This is helpful when you are upserting hundreds of documents into your vector database for your rag pipeline. I'm not 100% sure of the buffer size for the scraping delay on this project however there was a posting on Reddit that was going around so it piqued my interest :) |
BetaWas this translation helpful?Give feedback.
All reactions
-
will this be helpful for websites which has enabled anti scraping techniques ? |
BetaWas this translation helpful?Give feedback.