Self-hosted large language models
Running LLMs locally is technically demanding. It requires specialized hardware, deep system integration knowledge, and expertise in model architecture, which most SMEs lack. At the same time, using cloud-based LLMs raises serious privacy and confidentiality concerns—especially when handling sensitive data or operating under strict regulations like GDPR. Off-the-shelf APIs are also expensive and poorly suited for domain-specific use cases, leaving SMEs without practical, compliant, or cost-effective AI solutions.
The project addresses these challenges by locally hosting an open-source LLM (starting with LLama and DeepSeek) on fortiss hardware, allowing full control over data and infrastructure. This setup enables experimentation with different models, finetuning strategies, and deployment architectures tailored to SME needs.
01.03.2025 - 30.06.2025