Jesus RodriguezinTowards AIInside Infini Attention: Google DeepMind’s Technique Powering Gemini 2M Token WindowThe method combines compressive memory and attention mechanisms in a single structure.5d ago5d ago
Jesus RodriguezinTowards AIInside AlphaFold 3: A Technical View Into the New Version of Google DeepMind’s BioScience ModelA highly improved architecture drastically expands the capabilities of AlphaFold.May 13May 13
Jesus RodriguezinTowards AIPredicting Multiple Tokens at the Same Time: Inside Meta AI’s Technique for Faster and More Optimal…The mehod addresses the limitations of the classic next token prediction method.May 61May 61
Jesus RodriguezinTowards AISome Technical Notes About Phi-3: Microsoft’s Marquee Small Language ModelThe model ius able to outperform much larger alternatives and now run locally on mobile devices.Apr 29Apr 29
Jesus RodriguezinTowards AISome Technical Notes About Llama 3New tokenizer, optimized pretraining and some other details about Meta AI’s new model.Apr 22Apr 22
Jesus RodriguezinTowards AIInside Ferret-UI: Apple’s Multimodal LLM for Mobile Screen UnderstandingThe new research can have an impact in task automation in mobile apps.Apr 15Apr 15
Jesus RodriguezinTowards AIInside Jamba: Mamba, Transformers, and MoEs Together to Power a New Form of LLMsThe new architecture was pioneered by AI21 Labs and brought the best of several architecture paradigms in a single model.Apr 8Apr 8
Jesus RodriguezinTowards AIInside DBRX: Databricks’ Impressive Open Source LLMThe model shows a strong perfromance across many industry benchmarks.Apr 12Apr 12
Jesus RodriguezinTowards AIInside RAFT: UC Berkeley’s Method to Improve RAG for Domain Specific ScenariosThe method addresses some of the biggest limitations of RAG approaches.Mar 261Mar 261
Jesus RodriguezinTowards AIInside SIMA: Google DeepMind’s New Agent that Can Follow Language Instructions to Interact with Any…The research represents an important milestone in developing generalist agents for 3D interactions.Mar 182Mar 182