Replies: 2 comments
-
Hey @gooseillo don't use inference - but only "preprocessing". Have a look at how we build our lightweight Docker image |
Beta Was this translation helpful? Give feedback.
0 replies
-
Perhaps it would be useful to see how we make Docker images at https://github.com/deepset-ai/haystack/tree/main/docker and pay attention to cpu-remote-inference in our bake file |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
I'm trying to deploy a haystack RAG pipeline to EBS or EC2 on one of the lower config machines (2CPU - 2GB RAM and 2CPU 4GB RAM). While building the container, I used pip install 'farm-haystack[inference]' ## installs torch, sentence-transformers, sentencepiece, and huggingface-hub.
I was wondering, since I only intend to do inference on the CPU, why does it have to install torch and other heavier libraries. Is there a lighter install?
Otherwise, is there a cheaper deployment pipeline for a RAG solution?
Beta Was this translation helpful? Give feedback.
All reactions