GPU for LLM Inferencing Guide
A guide on what GPU and in which setup, to use for LLM Inference.
GPU for LLM Inferencing Guide Read More »
A guide on what GPU and in which setup, to use for LLM Inference.
GPU for LLM Inferencing Guide Read More »
This document presents a reference architecture for a simple Retrieval Augmented Generation solution based on a vector Db using OVHcloud managed services. In this use case we have a large number of pdf/markdown documents that are ingested as a single batch to create a knowledge base and a simple text chat interface for a user
Reference Architecture : Retrieval Augmented Generation (RAG) Read More »