Published inTowards AISome Insights About Phi-4: Microsoft’s New Small Foundation Model that Punches Above its WeightI recently started an AI-focused educational newsletter, that already has over 170,000 subscribers. TheSequence is a no-BS (meaning no…4d ago4d ago
Pretrain Your Own AI Models with Fast-LLMCreated by ServiceNow, the framework provides the key building blocks for pretraining AI models.Dec 12Dec 12
Published inTowards AIInside Anthropic’s Model Context Protocol (MCP)to Connect AI Assistants to DataThe protocol tries to standarize one of the most important elements of agentic applications.Dec 4Dec 4
Published inTowards AIInside Tülu 3: Allen AI’s New Post-Training FrameworkThe framework includes capabilities for data, models, post-training and evaluation under the same architecture.Nov 25Nov 25
Published inTowards AIInside FrontierMath: An Unprecedented Benchmark for Assessing Advanced Mathematical Reasoning in AIThe benchmark introduces evaluations that take AI mathematical reasoning to a new level.Nov 191Nov 191
Published inTowards AIMeet Magentic-One: Microsoft’s New Multi-Agent Framework for Solving Complex TasksThe framework is built on the AutoGen framework.Nov 121Nov 121
How Did Google Build NotebookLM’s Cool Podcast Generation Features?The technique combines several models into a comprehensive audio generation approach.Nov 7Nov 7
Published inTowards AIAnthropic New Research Shows that AI Models Can Sabotage Human EvaluationsThe new research proposes a framework for assessing a model’s ability to subvert human evaluations.Oct 28Oct 28
Inside Meta AI’s New Method to Build LLMs that Think Before they SpeakThought Preference Optimization could be the new foundation for “Thinking LLMs”.Oct 22Oct 22
Published inTowards AIInside OpenAI’s MLE-Bench: A New Benchmark for Evaluating Machine Learning Engineering Capabilities…The new benchmark evaluates AI agents in areas such as pretraining, evaluation and others.Oct 15Oct 15