Skip to content
EARNST.

AI - Local Hardware

AI Development - Local Hardware | EARNST

AI on your own hardware. Local LLMs, RAG systems, document analysis. On-prem GPU servers, no cloud dependency.

What we build

AI systems that run on your own hardware -- dedicated GPU servers in your data center, not someone else's cloud. This includes document analysis (extracting information from PDFs, contracts, invoices), RAG systems (question answering based on your internal documents), text classification (routing, categorization, sentiment analysis), and custom workflows (automating repetitive tasks that require language understanding).

We work with open-source LLMs (Llama, Mistral, Phi) running on local hardware, vector databases for semantic search, and custom fine-tuning when necessary. The key difference from typical AI products: your data never leaves your building. No OpenAI API calls. No cloud dependency. No surprise pricing changes. Just models running on hardware you own, processing data you control, with performance you can predict.

Who needs this?

Businesses with data that can't be sent to external APIs for legal, competitive, or compliance reasons. Common scenarios: law firms analyzing contracts, healthcare organizations processing patient data, financial services handling sensitive documents, manufacturers with proprietary product information, HR departments processing applications and employee data.

If you're already using ChatGPT or similar services but feel uncomfortable about data privacy, or if you've been told "we'd use AI but GDPR won't allow it," this is the solution. Also relevant for companies processing high volumes where API costs would exceed the cost of running your own infrastructure.

How EARNST approaches it

We start by questioning whether AI is the right solution at all. Many problems can be solved with simpler, more reliable methods (rules, search, traditional ML). If AI genuinely helps, we prototype with the simplest model that could work, test it on real data, and measure whether accuracy meets business requirements. Only then do we move to production infrastructure.

Infrastructure runs on your own local servers -- dedicated GPU machines (NVIDIA A100, H100, or RTX 4090/L40S for smaller models). We use Docker for deployment, implement proper monitoring (response times, accuracy, resource usage), and set up fallback mechanisms for when models fail. We document limitations clearly. AI models make mistakes, we design systems that account for this reality rather than pretending everything works perfectly.

Project scope

A feasibility analysis (can AI solve this problem?) takes 1 to 2 weeks and includes prototype testing on sample data. Full implementation timelines vary significantly: a document extraction system (parsing invoices, extracting key fields) takes 4 to 6 weeks. A RAG system (internal knowledge base with question answering) takes 6 to 8 weeks. Complex custom workflows (multi-step automation with human oversight) can take 10 to 14 weeks.

Infrastructure costs depend on volume and model size. Smaller models run on standard servers (€100 to €300/month). Larger models or high volume require GPU instances (€500 to €2,000/month). We optimize for cost efficiency: using the smallest model that achieves required accuracy, batching requests, caching results.

Typical Results

0

Data transfers to third-party providers

100%

GDPR compliance through local infrastructure

On-Prem

Full control over models and data

What you get

Feasibility Analysis

Technical assessment of whether AI can solve your specific problem.

Model Selection & Training

Choosing and fine-tuning models for your use case and data.

Infrastructure Setup

Deploying models on your own servers with proper resource management.

API Development

Building APIs for integrating AI capabilities into your applications.

GDPR Documentation

Compliance documentation for data processing and privacy impact.

“6 Jahre erfolgreiche Zusammenarbeit - Digitalisierung und Automatisierung auf höchstem Niveau.”

Christoph Fraundorfer

GF, MY ESEL

Frequently Asked Questions

Do I need my own GPU servers?

Not necessarily. For smaller models (7B-13B parameters), cloud GPU instances in EU data centers are sufficient. For larger models or intensive continuous operation, dedicated hardware pays off. We help with TCO calculations.

How good are open-source models compared to ChatGPT?

For specialized tasks (document analysis, code assistance, domain-specific chatbots), fine-tuned open-source models achieve comparable or better results. For general conversation, commercial models have broader coverage.

How much does GDPR-compliant AI cost?

Feasibility analysis from 3,000 EUR. PoC 8,000 to 20,000 EUR. Production system 30,000 to 80,000 EUR depending on complexity. Plus ongoing infrastructure costs. Fixed price after feasibility analysis.

Can you migrate existing ChatGPT solutions?

Yes. If you currently use OpenAI APIs and need to switch for privacy reasons, we migrate to local open-source models. Usually with equivalent or better performance for your specific use case.

How long does a typical AI project take?

Feasibility analysis 1 week, PoC 2 to 4 weeks, production system 8 to 16 weeks. We always recommend a feasibility analysis first before investing in a full project.

Ready to discuss?

Tell us about your project. We will get back to you within 24 hours.