This is what efficient AI looks like: Gemma 3n just dropped - a natively multimodal model that runs entirely on your device. No cloud. No API calls.
π§ Text, image, audio, and video - handled locally. β‘οΈOnly needs 2B in GPU memory to run π€― First sub-10B model to hit 1300+ Elo β Plug-and-play with Hugging Face, MLX, llama.cpp, and more.
Plus: Multilingual out of the box (140+ languages), fine-tune in a free Colab notebook.
π§ We just implemented Andrej Karpathy's "third paradigm" for LLM learning!
System Prompt Learning (SPL) enables LLMs to automatically learn problem-solving strategies from experience, rather than relying on static prompts.
π How it works: Your LLM builds a database of effective strategies, selects the best ones for each problem, and refines them over time based on success rates.
The best part? All strategies are human-readable and the system gets progressively better at problem types you use frequently.
β¨ Key benefits: π Cumulative learning over time π Transparent, inspectable strategies π Works with any OpenAI-compatible API β‘ Simple integration: just add "spl-" prefix to your model
Built as an open-source plugin in optillm. After 500 queries, our system developed 129 strategies and refined 97 of them!
This feels like a genuine step toward AI that learns from experience while staying completely interpretable.