baumhaus.digital/Miscellanous/Presentations/AE54/Helmeto 2024/mid-sized LLMs
We use mid-sized (< 8 billion parameters) large language models derived from Llama 3.1 8B and Mistral 7b base models.
On top of these models, we subsequently train specific adapters by means of Low Rank Adaptation (LoRA) methodology.
Additionally, Retrieval Augmented Generation (RAG) is also deployed in order to increase response accuracy.
0 Axones