Ignacio G L.
San Francisco, California, United States
4K followers
500+ connections
View mutual connections with Ignacio G
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
View mutual connections with Ignacio G
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Experience
View Ignacio G’s full profile
-
See who you know in common
-
Get introduced
-
Contact Ignacio G directly
Other similar profiles
Explore more posts
-
Kameshwara Pavan Kumar Mantha
🚀 Self Attention vs. Causal Attention (Masked Attention): An Intuitive Explanation When i talk to many people they dont understand the foundational difference between different attention mechanisms which is the driving power of all generativeAI models and the below is my simple contribution to the community. In the world of Transformers and Generative AI models, two attention mechanisms dominate—Self Attention and Causal Attention. Let's simplify these with a clear example! 🔍 🤔 Self Attention: Imagine you're reading a sentence, "Learning Generative AI is extremely interesting." While understanding the word "extremely", you simultaneously glance at every other word (including the ones after) to grasp context fully. In self-attention, every word "attends" to every other word in the entire sentence, providing rich context from both sides. 🌐✨ ⏳ Causal Attention (Masked Attention): Now, imagine you're writing or predicting the next word of the same sentence, but at the word "extremely", you haven't seen the word "interesting" yet. Here, you can only rely on previously seen words—no future peeking! Causal attention restricts each word from seeing any future tokens, which aligns closely with how language models predict words sequentially. 🔒➡️ In Short: Self Attention: All words see each other—full context. 🌍 Causal Attention: Words only see previous words—sequential context. 🔗 Hope this clears things up! 💡😊 Note: numbers in the diagram are just to understand the concept. #AI #GenAI #LLM #AIAgents #RAG #AgenticWorkflows
45
2 Comments -
Nikhil Bhaskaran
lya Sutskever has confirmed that LLMs are not scaling with massive gains anymore Enough if Compute continues to scale, but data is not keeping up. New or synthetic data doesnt significantly improved performance. In next wave models can actually get smaller and do one job right. 1)Agentic AI system is the future 2)LLMs can be replaced with SLMs to create optimal systems 3)LLM/SLM not being absolutely deterministic in outputs wont matter as long as system design is robust.
46
3 Comments -
Amar Goel
Apple just published some great research on the limited reasoning capabilities of LLMs. Some key takeaways: - The paper reveals a sharp drop in performance when problems exceed a certain complexity threshold — showing that even so-called reasoning models fail on tougher tasks - What looks like reasoning is often just pattern-matching learned from training data—not true logical problem-solving. Some researchers like Gary Marcus and Subbarao Kambhampati have been claiming this for a long time. - So LLMs can appear smart, until you give them problems they truly haven't seen before. Then they collapse. Link below to the Apple research. My take: This research feels pretty spot-on. I think we are pretty far from AGI and AI that invents stuff like crazy. But for AI to be quite useful, or even very useful, we don't need AGI. We are at "useful" today and the continuously improving models and falling price of intelligence will enable lots of new applications. https://lnkd.in/gmPzUtfW
31
4 Comments -
Dom Heinrich
Before you all go nuts on DeepSeek AI and ride the hype...consider this, please: It is true that with DeepSeek you don’t need a #supercomputer to build incredible #AI. Their model, built on a fraction of the usual cost, is challenging the industry’s obsession with “bigger and faster.” 🧠 However, it is a great example that innovation isn’t always about brute force—it’s about thinking smarter with what you have. For the industry, it’s a chance to rethink what’s possible. Big Take-Aways: 🥇 DeepSeek achieves top-tier performance with a fraction of the cost and hardware, making AI more accessible for businesses with limited resources. 🐎 With 128K tokens, it processes significantly larger datasets, enhancing its utility for tasks requiring vast context. 🔑 DeepSeek excels in efficiency and accessibility but doesn’t yet match OpenAI’s versatility and creative adaptability, which remain essential for many industries. 🤝 Its approach complements high-power models like GPT-4, opening the door for hybrid solutions that combine efficiency with unmatched raw capabilities. So, let’s not sound the alarm for OpenAI, Google, Anthropic, NVIDIA and others (for now). Imagine combining DeepSeek's efficiency-driven approach with the raw power of existing models—suddenly, you have the best of both worlds. This highlights something we always teach and speak about at Creative AI Academy and Pratt Institute: Thinking Design replaces Design Thinking. In an AI-driven world, the solutions we create rely on the depth of our ideas, not just the tools we use. Change like this isn’t a threat — it’s an opportunity to evolve. #innovation #artificialintelligence #genai #thinkingdesign #aidesign #deepseek #openai #chatgpt
104
14 Comments
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content