'use client' import { motion } from 'framer-motion' import { Language } from '@/lib/types' import GradientText from '../ui/GradientText' import FadeInView from '../ui/FadeInView' import { Brain, Cpu, Zap, ArrowRight } from 'lucide-react' interface AnnexAIStackSlideProps { lang: Language } export default function AnnexAIStackSlide({ lang }: AnnexAIStackSlideProps) { const roadmapSteps = [ { year: '2026', model: 'Qwen 32B', label: lang === 'de' ? 'Foundation' : 'Foundation', }, { year: '2027', model: 'Multi-Model 70B', label: lang === 'de' ? 'Skalierung' : 'Scaling', }, { year: '2028', model: 'Fine-Tuned 100B+', label: 'Enterprise', }, { year: '2030', model: '1000B Agent Network', label: lang === 'de' ? 'Volle Autonomie' : 'Full Autonomy', }, ] return (

{lang === 'de' ? 'KI-Stack & LLM-Architektur' : 'AI Stack & LLM Architecture'}

{lang === 'de' ? 'Von 32B zu 1000B Parametern' : 'From 32B to 1000B parameters'}

{/* LLM Evolution Roadmap */}
{roadmapSteps.map((step, index) => (
{step.year}
{step.model}
{step.label}
{index < roadmapSteps.length - 1 && ( )}
))}
{/* Bottom Section: 2-column grid */}
{/* Left Card - Multi-Model Router */}

{lang === 'de' ? 'Multi-Model Router' : 'Multi-Model Router'}

  • {lang === 'de' ? 'Automatische Modellauswahl basierend auf Aufgabe' : 'Automatic model selection based on task'}
  • {lang === 'de' ? 'Einfache Anfragen → kleineres Modell (schnell, günstig)' : 'Simple queries → smaller model (fast, cheap)'}
  • {lang === 'de' ? 'Komplexe Analysen → größeres Modell (präzise)' : 'Complex analysis → larger model (accurate)'}
  • {lang === 'de' ? 'Kostenoptimierung: 60% günstiger als immer großes Modell' : 'Cost optimization: 60% cheaper than always using large model'}
{/* Right Card - Inference Optimization */}

{lang === 'de' ? 'Inference-Optimierung' : 'Inference Optimization'}

  • {lang === 'de' ? 'Apple Neural Engine Beschleunigung' : 'Apple Neural Engine acceleration'}
  • {lang === 'de' ? 'Quantisierung (INT4/INT8) für schnellere Inferenz' : 'Quantization (INT4/INT8) for faster inference'}
  • Context window: 32k tokens
  • {lang === 'de' ? 'Durchschn. Antwortzeit: <2s für einfache Anfragen' : 'Avg response time: <2s for simple queries'}
  • {lang === 'de' ? 'GPU vs Apple Silicon: Apple 3x besseres Preis-/Leistungsverhältnis für Inferenz' : 'GPU vs Apple Silicon: Apple 3x better price/performance for inference'}
) }