Sign in to view Avijit’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Avijit’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Bengaluru, Karnataka, India
Sign in to view Avijit’s full profile
Avijit can introduce you to 10+ people at Sigmoid
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
15K followers
500+ connections
Sign in to view Avijit’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Avijit
Avijit can introduce you to 10+ people at Sigmoid
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Avijit
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Avijit’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Articles by Avijit
-
Decoding Transformers: The Heart of Large Language Models
Decoding Transformers: The Heart of Large Language Models
In the realm of artificial intelligence, Large Language Models (LLMs), are revolutionizing the landscape of natural…
56
-
Build Domain-Specific LLMs Using Retrieval Augmented GenerationOct 31, 2023
Build Domain-Specific LLMs Using Retrieval Augmented Generation
Organizations are in a race to adopt Large Language Models. Let’s dive into how you can build domain-specific LLMs…
22
1 Comment
Activity
15K followers
-
Avijit Swain shared this𝗜 𝗱𝗶𝗱𝗻’𝘁 𝗿𝗲𝗮𝗹𝗶𝘇𝗲 𝗶𝘁 𝗯𝗮𝗰𝗸 𝘁𝗵𝗲𝗻, 𝗯𝘂𝘁 𝘁𝗵𝗲 𝗺𝗼𝘀𝘁 𝘃𝗮𝗹𝘂𝗮𝗯𝗹𝗲 𝘁𝗵𝗶𝗻𝗴 𝗜 𝗹𝗲𝗮𝗿𝗻𝗲𝗱 𝘄𝗮𝘀𝗻’𝘁 𝗰𝗼𝗱𝗶𝗻𝗴. 𝙄𝙩 𝙬𝙖𝙨 𝙝𝙤𝙬 𝙩𝙤 𝙩𝙝𝙞𝙣𝙠! Learning at Scaler didn’t just teach me how to write code. It taught me how to: • break down problems • question assumptions • think deeply before jumping to solutions And that changed everything. Because today, I build GenAI and Agentic AI systems — where the challenge isn’t writing code, but designing systems that can reason, adapt, and operate in real-world complexity. That same foundation is what now allows me to: • mentor and teach people in AI/ML • simplify complex systems into intuitive ideas • help others move from using tools → to understanding systems Grateful to the instructors who shaped this mindset early on — Subhodeep Dey, Mudit Goel, Anant Mittal, Mohit Uniyal, Anshuman Singh, Abhimanyu Saxena You didn’t just teach concepts. You taught a way of thinking that still compounds every day. And that’s why I’m genuinely excited about Scaler Academy 3.0. It feels like a natural evolution — 𝗳𝗿𝗼𝗺 𝘁𝗲𝗮𝗰𝗵𝗶𝗻𝗴 𝗽𝗲𝗼𝗽𝗹𝗲 𝘄𝗵𝗮𝘁 𝘁𝗼 𝗹𝗲𝗮𝗿𝗻 → 𝘁𝗼 𝗽𝗿𝗲𝗽𝗮𝗿𝗶𝗻𝗴 𝘁𝗵𝗲𝗺 𝗳𝗼𝗿 𝗵𝗼𝘄 𝘁𝗵𝗲 𝗶𝗻𝗱𝘂𝘀𝘁𝗿𝘆 𝗮𝗰𝘁𝘂𝗮𝗹𝗹𝘆 𝘄𝗼𝗿𝗸𝘀 𝘁𝗼𝗱𝗮𝘆. What stands out to me: • AI is no longer a separate track — it’s becoming part of everything you build • The focus is shifting from solving problems → to designing end-to-end systems • Learning is moving closer to real-world workflows, not just theoretical exercises • There’s a stronger emphasis on execution, reasoning, and practical application If you’re starting your journey, or even reinventing yourself in this space — this is something worth checking out. 𝗕𝗲𝗰𝗮𝘂𝘀𝗲 𝗶𝗻 𝘁𝗵𝗲 𝗲𝗻𝗱, 𝘁𝗼𝗼𝗹𝘀 𝘄𝗶𝗹𝗹 𝗸𝗲𝗲𝗽 𝗰𝗵𝗮𝗻𝗴𝗶𝗻𝗴. 𝗕𝘂𝘁 𝘁𝗵𝗲 𝘄𝗮𝘆 𝘆𝗼𝘂 𝘁𝗵𝗶𝗻𝗸? 𝗧𝗵𝗮𝘁’𝘀 𝘄𝗵𝗮𝘁 𝘀𝘁𝗮𝘆𝘀. Link to the full video here - https://lnkd.in/gtts3mQD #Scaler #FundamentalFirst #AIForwardBuilt For The Next Decade,This Is The New Scaler. Fundamental First and AI forward.Built For The Next Decade,This Is The New Scaler. Fundamental First and AI forward.
-
Avijit Swain posted this𝗢𝗽𝗲𝗻 𝘁𝗼 𝗢𝗽𝗽𝗼𝗿𝘁𝘂𝗻𝗶𝘁𝗶𝗲𝘀 | 𝗚𝗲𝗻𝗔𝗜 & 𝗔𝗴𝗲𝗻𝘁𝗶𝗰 𝗔𝗜 After building and deploying production-grade GenAI and agentic systems, I’m exploring opportunities where I can contribute at scale. Most AI systems don’t fail at the model — they fail in everything around it: context, retrieval, orchestration, evaluation, and reliability. My work has been focused on solving exactly that. 𝗜 𝘁𝗵𝗶𝗻𝗸 𝗹𝗶𝗸𝗲 𝗮 𝗽𝗿𝗼𝗱𝘂𝗰𝘁 𝗺𝗮𝗻𝗮𝗴𝗲𝗿, 𝗯𝘂𝗶𝗹𝗱 𝗹𝗶𝗸𝗲 𝗮𝗻 𝗔𝗜 𝗮𝗿𝗰𝗵𝗶𝘁𝗲𝗰𝘁, 𝗮𝗻𝗱 𝘁𝗲𝘀𝘁 𝗹𝗶𝗸𝗲 𝗮 𝗤𝗖 — ensuring systems are not just impressive in demos, but robust in production. 💼 Roles I’m open to: • Artificial Intelligence Engineer • Generative AI Engineer • Senior Data Scientist • Lead Data Scientist 📍 Preferred Location: Bengaluru 🌍 Open to: On-site | Hybrid | Remote Over the past few years, I’ve worked on: • Designing multi-agent systems with planning, reasoning, and controlled tool orchestration • Building hybrid retrieval stacks (vector, graph, structured) for real-world query complexity • Developing memory architectures for long-term context, personalization, and follow-ups • Productionizing LLM applications with strong evaluation, guardrails, and failure handling I’m particularly interested in roles where I can: • Own and design end-to-end AI systems • Solve high-impact business problems using GenAI • Drive architecture, scalability, and system reliability If you’re working on serious AI systems — or hiring for them — I’d be happy to connect. #OpenToWork #GenAI #AgenticAI #AIEngineering #LLM #DataScience #Bengaluru
-
Avijit Swain shared thisThis is how learning needs to evolve. Not crash courses on AI. Not surface-level exposure. But a complete rethink — where fundamentals and AI go hand in hand. The #FundamentalFirst, #AIForward philosophy is exactly what the industry needs right now — especially with the growing gap between AI awareness and real capability. Great to see Scaler leaning into this transformation. If you’re serious about staying relevant in an AI-first world, this is worth exploring. #Scaler #FundamentalFirst #AIForwardAvijit Swain shared thisThis is the new Scaler. Fundamental first. AI forward. Built for the next decade. Over the past few weeks the teams at Scaler have been hard at work reimagining all our programs, every curriculum, every module rebuilt for the AI first world. This isn't just a refresh with an add on module, it is a rebuild where every topic has been relooked at and rebuilt with focus on strong fundamentals and AI. Why this matters: - 1.6 million global AI roles remain unfilled, highlighting a massive talent gap. - India’s AI talent demand is projected to exceed 1.25 million by 2027. - Workers with AI skills command a 56% wage premium. - Our research reveals a “confidence–capability gap,” with many professionals feeling AI-ready but lacking deep, practical expertise. Scaler’s #Fundamentalfirst, #AIforward approach ensures learners don’t just become AI-aware they become AI-native, equipped to build, evaluate, and lead in an AI-driven world. Scaler is proud to become India’s first fully AI-native tech career platform, with every program reimagined across curriculum, pedagogy, projects, and career preparation. This transformation was co-designed with 1,200+ hiring partners to ensure alignment with real industry needs. This is more than a curriculum update. It’s a commitment to helping professionals lead the AI decade. #Scaler #BuiltForTheNextDecade Anshuman Singh Abhimanyu Saxena Amar Srivastava shivank agrawal Rahul Karthikeyan Nitin Solanki
-
Avijit Swain shared this𝗠𝗼𝘀𝘁 𝗚𝗲𝗻𝗔𝗜 𝘀𝘆𝘀𝘁𝗲𝗺𝘀 𝗱𝗼𝗻’𝘁 𝗳𝗮𝗶𝗹 𝗮𝘁 𝘁𝗵𝗲 𝗺𝗼𝗱𝗲𝗹. 𝗧𝗵𝗲𝘆 𝗳𝗮𝗶𝗹 𝗮𝗿𝗼𝘂𝗻𝗱 𝗶𝘁. Most teams are still stuck here: ⏺️ “Let’s improve the prompt.” ⏺️ “Add more context.” ⏺️ “Maybe reword this instruction.” I’ve done it too. It works… until you try to take it to production. Because in real systems: • queries are messy • context is incomplete • users ask follow-ups • data keeps changing • workflows are multi-step and user-specific And suddenly your “perfect prompt” isn’t so perfect anymore. 𝗧𝗵𝗶𝘀 𝗶𝘀 𝗻𝗼𝘁 𝗮 𝗽𝗿𝗼𝗺𝗽𝘁 𝗽𝗿𝗼𝗯𝗹𝗲𝗺. 𝗧𝗵𝗶𝘀 𝗶𝘀 𝗮 𝘀𝘆𝘀𝘁𝗲𝗺 𝗱𝗲𝘀𝗶𝗴𝗻 𝗽𝗿𝗼𝗯𝗹𝗲𝗺. Here’s what actually matters in production: ▪️Retrieval → Can you fetch the right context? ▪️Orchestration → Can you decide what to do next? ▪️Tools → Can the model verify instead of guess? ▪️Memory → Can the system learn over time? ▪️Validation → Can you trust the output? The LLM is just one component in this loop. Most systems are still built like this: ❌ 𝙐𝙨𝙚𝙧 → 𝙋𝙧𝙤𝙢𝙥𝙩 → 𝙇𝙇𝙈 → 𝘼𝙣𝙨𝙬𝙚𝙧 Clean. Simple .… and fundamentally limited. What actually works in production: ✅ 𝗨𝘀𝗲𝗿 → 𝗜𝗻𝘁𝗲𝗻𝘁 → 𝗣𝗹𝗮𝗻 → 𝗥𝗲𝘁𝗿𝗶𝗲𝘃𝗲 → 𝗔𝗰𝘁 → 𝗩𝗮𝗹𝗶𝗱𝗮𝘁𝗲 → 𝗔𝗻𝘀𝘄𝗲𝗿 We’re moving from: 𝗣𝗿𝗼𝗺𝗽𝘁 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 → 𝗖𝗼𝗻𝘁𝗲𝘅𝘁 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 → 𝗦𝘆𝘀𝘁𝗲𝗺 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 What does that mean? ✅ Context is assembled, not given. The system decides what to fetch, when, and why. ✅ Reasoning is iterative, not one-shot. Agents think, act, observe, and refine. ✅ Knowledge is layered, not static. Structured data, unstructured docs, graphs, APIs—all working together. ✅ Memory is foundational, not optional. Without it, your system forgets everything every time. This is why most “good demos” don’t survive production. Because demos assume: • static context • clean inputs • single-step reasoning Production has none of that. And this is where most teams struggle. They spend weeks optimizing prompts… …but ignore: • retrieval quality • tool integration • failure handling • grounding & validation • memory lifecycle So the system looks impressive in a demo … and falls apart in production. The real leverage today is not: “Can you write a better prompt?” It’s: “𝗖𝗮𝗻 𝘆𝗼𝘂 𝗯𝘂𝗶𝗹𝗱 𝗮 𝘀𝘆𝘀𝘁𝗲𝗺 𝘁𝗵𝗮𝘁 𝗸𝗻𝗼𝘄𝘀 𝘄𝗵𝗮𝘁 𝗶𝘁 𝗱𝗼𝗲𝘀𝗻’𝘁 𝗸𝗻𝗼𝘄... 𝗮𝗻𝗱 𝗴𝗼𝗲𝘀 𝗼𝘂𝘁 𝘁𝗼 𝗳𝗶𝗻𝗱 𝗶𝘁?” That’s the difference between: ❌ A chatbot ✅ An intelligent system We don’t just need prompt engineers anymore. We need people who can: • think like a product manager • design like an AI architect • evaluate like a QA system The teams that win won’t have the best prompts. They’ll have the best systems around the model. #AI #AgenticAI #LLMs #RAG #AIArchitecture #SoftwareEngineering #GenerativeAI #FutureOfWork
-
Avijit Swain shared this𝗥𝗲𝗽𝗼𝘀𝘁𝗶𝗻𝗴 𝘁𝗵𝗶𝘀 𝘄𝗶𝘁𝗵 𝗮 𝗾𝘂𝗶𝗰𝗸 𝗻𝗼𝘁𝗲 𝗼𝗻 𝗼𝘂𝗿 𝘂𝗽𝗰𝗼𝗺𝗶𝗻𝗴 𝘀𝗲𝘀𝘀𝗶𝗼𝗻👇 RAG is not the silver bullet people think it is—and blindly using it for retrieval is where most systems start to fall apart. For many people today, RAG has become synonymous with retrieval—almost like how Maggi is synonymous with noodles. And that’s exactly the problem. 𝗠𝗼𝘀𝘁 𝗥𝗔𝗚 𝗶𝗺𝗽𝗹𝗲𝗺𝗲𝗻𝘁𝗮𝘁𝗶𝗼𝗻𝘀 𝗹𝗼𝗼𝗸 𝗴𝗼𝗼𝗱 𝗶𝗻 𝗱𝗲𝗺𝗼𝘀, 𝗯𝘂𝘁 𝗯𝗿𝗲𝗮𝗸 𝘁𝗵𝗲 𝗺𝗼𝗺𝗲𝗻𝘁 𝘆𝗼𝘂 𝘁𝗿𝘆 𝘁𝗼 𝘀𝗰𝗮𝗹𝗲 𝘁𝗵𝗲𝗺 𝘁𝗼 𝗿𝗲𝗮𝗹-𝘄𝗼𝗿𝗹𝗱 𝘂𝘀𝗲 𝗰𝗮𝘀𝗲𝘀. In this session, I’ll be busting some common myths around RAG, why most implementations don’t actually work in production, and how to think about building reliable retrieval systems. This is easily the most important session of the entire course. If you’re serious about building production-grade agents, this is one you shouldn’t miss. If you haven’t registered yet, feel free to DM me—it’s completely free of cost. See you there.
-
Avijit Swain shared thisReally appreciate this Dr. Nabanita Sinha, thanks a lot for sharing your experience. Glad to know the session resonated with you—that’s exactly the intent behind structuring it this way. The goal is not just to explain concepts, but to help people think in terms of systems and real-world applications. Completely agree with your point—there’s a visible gap today between knowing terminology and actually understanding how GenAI systems work in practice. And that’s what we’re trying to bridge through these sessions. Looking forward to having you in the upcoming sessions—this is just the beginning.Avijit Swain shared thisThis Saturday felt a little different. I attended a session on #Agentic AI—not for a certificate, just out of curiosity. Recently I connected with Avijit Swain and he invited me to join his Agentic AI training session which is completely free. Really appreciate the effort he put into organizing it and structuring the content. What I #enjoyed most was how he simplified complex concepts and answered questions with practical examples. It was also great to see how curious the participants were to learn. Having taken several #interviews for GenAI roles, I often see a #gap in candidates between truly understanding concepts and simply remembering technical terms. Writing prompts for copilots is not GenAI engineering. Real learning comes from #discussions, listening to others’ #experiences, and understanding how things work in real life—that’s what makes it truly effective. AI Strategy & Insights : https://lnkd.in/djd4Sqqe
-
Avijit Swain shared this𝗔𝗜 𝘄𝗶𝗹𝗹 𝗰𝗵𝗮𝗻𝗴𝗲 𝗽𝗿𝗼𝗱𝘂𝗰𝘁𝗶𝘃𝗶𝘁𝘆, 𝗬𝗲𝘀. 𝗕𝘂𝘁 𝗺𝗼𝗿𝗲 𝗶𝗺𝗽𝗼𝗿𝘁𝗮𝗻𝘁𝗹𝘆, 𝗶𝘁 𝘄𝗶𝗹𝗹 𝗰𝗵𝗮𝗻𝗴𝗲 𝘁𝗵𝗲 𝗻𝗮𝘁𝘂𝗿𝗲 𝗼𝗳 𝗲𝘅𝗽𝗲𝗿𝘁𝗶𝘀𝗲 𝗶𝘁𝘀𝗲𝗹𝗳! That is the shift I believe many people are still underestimating. A lot of the current conversation around AI is still centered on speed: faster coding, faster writing, faster analysis, faster execution. But I do not think speed is the deepest transformation. 𝗧𝗵𝗲 𝗱𝗲𝗲𝗽𝗲𝗿 𝘀𝗵𝗶𝗳𝘁 𝗶𝘀 𝘁𝗵𝗮𝘁 𝗔𝗜 𝗶𝘀 𝘀𝘁𝗮𝗿𝘁𝗶𝗻𝗴 𝘁𝗼 𝗰𝗵𝗮𝗻𝗴𝗲 𝘄𝗵𝗲𝗿𝗲 𝗵𝘂𝗺𝗮𝗻 𝘃𝗮𝗹𝘂𝗲 𝘀𝗶𝘁𝘀 𝗶𝗻 𝘁𝗵𝗲 𝘀𝘆𝘀𝘁𝗲𝗺. For years, expertise was often expressed through execution: knowing the process, performing the steps, producing the output. Now, as AI becomes increasingly capable of handling parts of execution, the center of gravity begins to move. Human value starts shifting upward: ▪️ from doing to defining, ▪️ from executing to judging, ▪️ from producing to architecting. In that world, expertise will be less about only knowing how to do the task yourself, and more about knowing: ⏺️ what outcome should exist ⏺️ what constraints matter ⏺️ what good looks like ⏺️ what can be delegated ⏺️ what must remain human-controlled ⏺️ where validation is required before trust can be granted This is why I believe the future of AI will not be won by organizations that simply automate the most. 𝗜𝘁 𝘄𝗶𝗹𝗹 𝗯𝗲 𝘄𝗼𝗻 𝗯𝘆 𝗼𝗿𝗴𝗮𝗻𝗶𝘇𝗮𝘁𝗶𝗼𝗻𝘀 𝘁𝗵𝗮𝘁 𝘂𝗻𝗱𝗲𝗿𝘀𝘁𝗮𝗻𝗱 𝘁𝗵𝗲 𝗯𝗼𝘂𝗻𝗱𝗮𝗿𝘆 𝗯𝗲𝘁𝘄𝗲𝗲𝗻 𝗺𝗮𝗰𝗵𝗶𝗻𝗲 𝗲𝘅𝗲𝗰𝘂𝘁𝗶𝗼𝗻 𝗮𝗻𝗱 𝗵𝘂𝗺𝗮𝗻 𝗷𝘂𝗱𝗴𝗺𝗲𝗻𝘁 𝗯𝗲𝘁𝘁𝗲𝗿 𝘁𝗵𝗮𝗻 𝗲𝘃𝗲𝗿𝘆𝗼𝗻𝗲 𝗲𝗹𝘀𝗲. Because AI can generate. AI can reason. AI can execute. But it still operates inside frames that humans define. And that makes the real differentiator something much deeper than prompting or task automation. It becomes: 𝗷𝘂𝗱𝗴𝗺𝗲𝗻𝘁, 𝘀𝘆𝘀𝘁𝗲𝗺 𝗱𝗲𝘀𝗶𝗴𝗻, 𝗴𝗼𝘃𝗲𝗿𝗻𝗮𝗻𝗰𝗲, 𝗮𝗻𝗱 𝘁𝗿𝘂𝘀𝘁. To me, that is the real meaning of AI maturity. Not just deploying copilots. Not just reducing manual effort. But redesigning work so that human intelligence and machine intelligence are used in the right proportions. The next generation of leaders, engineers, and domain experts will not create value only by executing tasks faster. 𝗧𝗵𝗲𝘆 𝘄𝗶𝗹𝗹 𝗰𝗿𝗲𝗮𝘁𝗲 𝘃𝗮𝗹𝘂𝗲 𝗯𝘆 𝗯𝗲𝗶𝗻𝗴 𝗮𝗯𝗹𝗲 𝘁𝗼 𝗱𝗶𝗿𝗲𝗰𝘁 𝗶𝗻𝘁𝗲𝗹𝗹𝗶𝗴𝗲𝗻𝗰𝗲 𝘄𝗲𝗹𝗹: setting intent, designing guardrails, building verification loops, and deciding where autonomy should end. That is why I believe AI will do more than improve productivity. 𝗜𝘁 𝘄𝗶𝗹𝗹 𝗿𝗲𝗱𝗲𝗳𝗶𝗻𝗲 𝗲𝘅𝗽𝗲𝗿𝘁𝗶𝘀𝗲 𝗶𝘁𝘀𝗲𝗹𝗳. And the people who stay most relevant in that future will not just be the best executors. They will be the ones who can combine context, judgment, architecture, and accountability into systems that make AI useful, reliable, and trustworthy at scale. #AI #AgenticAI #AIEngineering #GenerativeAI #DigitalTransformation
-
Avijit Swain shared this𝗦𝘁𝗼𝗽 𝗯𝘂𝗶𝗹𝗱𝗶𝗻𝗴 𝗔𝗜 𝗮𝗴𝗲𝗻𝘁𝘀 𝘄𝗶𝘁𝗵 𝗴𝗼𝗹𝗱𝗳𝗶𝘀𝗵 𝗺𝗲𝗺𝗼𝗿𝘆. 🐟 Most agents today feel intelligent… for five minutes! Then they reset, forget context, and start from zero again. 𝗧𝗵𝗮𝘁’𝘀 𝗻𝗼𝘁 𝗮𝗻 𝗟𝗟𝗠 𝗽𝗿𝗼𝗯𝗹𝗲𝗺. 𝗧𝗵𝗮𝘁’𝘀 𝗮 𝘀𝘆𝘀𝘁𝗲𝗺 𝗱𝗲𝘀𝗶𝗴𝗻 𝗽𝗿𝗼𝗯𝗹𝗲𝗺! Recently explored Agent "Memory: Building Memory-Aware Agents" by DeepLearning.AI — and it strongly reinforced something I’ve been seeing in real-world systems: 👉 𝗠𝗲𝗺𝗼𝗿𝘆 𝗶𝘀 𝗻𝗼𝘁 𝗮 𝗳𝗲𝗮𝘁𝘂𝗿𝗲. 𝗜𝘁’𝘀 𝗶𝗻𝗳𝗿𝗮𝘀𝘁𝗿𝘂𝗰𝘁𝘂𝗿𝗲. In fact, I had written about this earlier while working on production GenAI systems: 🔗 https://lnkd.in/g2yaq-a3 The core idea remains the same: 👉 Treating memory as storage is where most systems start failing. A few ideas that really stood out (and align with production learnings): • 𝗦𝘁𝗮𝘁𝗲𝗳𝘂𝗹 > 𝗦𝘁𝗮𝘁𝗲𝗹𝗲𝘀𝘀 If your agent resets every session, you’re not building an agent — you’re building a demo. • 𝗥𝗲𝘁𝗿𝗶𝗲𝘃𝗮𝗹 𝗶𝘀 𝘁𝗵𝗲 𝗿𝗲𝗮𝗹 𝗶𝗻𝘁𝗲𝗹𝗹𝗶𝗴𝗲𝗻𝗰𝗲 𝗹𝗮𝘆𝗲𝗿 It’s not about storing everything, but retrieving the right context at the right time. • 𝗠𝗲𝗺𝗼𝗿𝘆 != 𝗖𝗵𝗮𝘁 𝗛𝗶𝘀𝘁𝗼𝗿𝘆 Logs are not memory. Real memory is structured, selective, and evolves. • 𝗠𝗲𝗺𝗼𝗿𝘆 𝗶𝘀 𝗮 𝗹𝗶𝗳𝗲𝗰𝘆𝗰𝗹𝗲, 𝗻𝗼𝘁 𝗮 𝗱𝗮𝘁𝗮𝗯𝗮𝘀𝗲 Construct → Structure → Store → Retrieve → Update → Decay What this reinforces for me: We are moving from “𝗟𝗟𝗠 + 𝘁𝗼𝗼𝗹𝘀” → “𝗟𝗟𝗠 + 𝗺𝗲𝗺𝗼𝗿𝘆 + 𝗿𝗲𝗮𝘀𝗼𝗻𝗶𝗻𝗴 𝗹𝗼𝗼𝗽𝘀” → “𝗗𝗲𝗲𝗽 𝗔𝗴𝗲𝗻𝘁𝘀” And that shift is what will separate toy agents from production-grade systems! Deep Agents are not just tool-using systems — they are memory-aware, context-evolving, continuously learning systems. If you're building GenAI systems beyond demos, memory design is not optional — it's foundational in 2026. #GenAI #AgenticAI #DeepAgents #AIAgents #LLMSystem #AIArchitecture #MemorySystems #DeepLearningAIDesigning Memory Architectures for Production-Grade GenAI SystemsDesigning Memory Architectures for Production-Grade GenAI Systems
-
Avijit Swain shared thisThanks for the overwhelming response so far. We’ve crossed 100+ learners in the cohort, which clearly reflects the growing interest in understanding how real-world GenAI and Agentic AI systems are actually built and deployed. Excited to see so many people leaning into this space and investing in learning what truly matters beyond the hype.
-
Avijit Swain reacted on thisAvijit Swain reacted on thisExcited to share that I have officially joined PwC as an Associate in Advisory and Consulting! 🎉 Looking forward to learning and growing with the best! #PwC #NewBeginnings #Advisory #Consulting #CareerMilestone #Grateful
-
Avijit Swain reacted on thisAvijit Swain reacted on thisEver wondered how tools like ChatGPT actually “understand” what we say? I recently explored this concept in a simple way, and here’s what stood out 1️⃣ LLMs don’t process text — they process numbers Before anything happens, your input goes through 2 steps: 👉 Tokenization (breaking text into pieces) 👉 Embeddings (converting into numbers) No numbers = no understanding 2️⃣ Why convert text into numbers? Because everything inside an AI model is math - Learning = optimization - Prediction = probability - Training = numerical adjustments 👉 No math without numbers 👉 No AI without numbers 💡 One more thing: Numbers also allow models to measure similarity between words — something text alone cannot do 3️⃣ What is a Token? A token is a small unit of text Example: “Playing football” Can become: - Words → Playing | football - Subwords → Play | ing | football - Characters → P | l | a | y … Types of tokens: - Word tokens - Subword tokens (most common in LLMs) - Character tokens ❓ Why not just use characters? Because: - Too many tokens → slower processing - Harder to capture meaning - Less efficient learning 👉 Subwords = best balance between meaning + efficiency 4️⃣ Do LLMs work with tokens? → BIG NO They work with EMBEDDINGS Tokens are just an intermediate step 👉 Tokens → converted into vectors (numbers) This is called Embedding Example: “King” → [0.21, -0.45, 0.88, …] A word becomes a list of numbers 5️⃣ Why embeddings are powerful Because they capture relationships Example: - Germany ↔ Berlin - France ↔ Paris In embedding space: 👉 Germany is “close” to France 👉 Berlin is “close” to Paris Even cooler: King - Man + Woman ≈ Queen 💡 This is something tokens alone can NEVER do 6️⃣ How real embeddings work (simple view) 1. Words appearing in similar contexts → get similar vectors 2. Distance between vectors = similarity of meaning 3. Model learns these relationships during training 👉 It’s like creating a map of language in numbers Final thought LLMs don’t understand words They understand patterns in numbers Once you see this: Text → Tokens → Embeddings → Prediction AI stops feeling like magic and starts making sense Still learning this deeply Thanks Avijit Swain for igniting curiosity in this field and the classes which makes things easy to understand from other sources too. #AI #MachineLearning #GenAI #DeepLearning #Learn
Experience & Education
-
Simplilearn
****** * ********** * ***** * *********
-
********* ******
****** * ********** * ***** * *********
-
******
******
-
*****
****** ** ******* * ** ******** ******* *************** * ** *** *** undefined
-
-
******* ******** ****** ******
******** *********
-
View Avijit’s full experience
See their title, tenure and more.
Already on LinkedIn? Sign in
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Licenses & Certifications
Recommendations received
9 people have recommended Avijit
Join now to viewView Avijit’s full profile
-
See who you know in common
-
Get introduced
-
Contact Avijit directly
Explore more posts
-
Shashank Garewal
Freelance • 722 followers
🚀 Building a Time Series Forecasting Project - EDA Analysis Complete Continuing the structured time series analysis and forecasting project using real market data (MSFT stock prices). After completing Notebook 01 (Data Inspection), this notebook dives into deeper time series diagnostics to understand the statistical structure before moving into modeling. Previous Post: https://lnkd.in/dxgEatzB This step establishes the statistical groundwork for feature engineering and model development. 𝐖𝐡𝐚𝐭 𝐭𝐡𝐢𝐬 𝐩𝐡𝐚𝐬𝐞 𝐟𝐨𝐜𝐮𝐬𝐞𝐝 𝐨𝐧: • Horizon selection based on market and data regimes • Multi-horizon trend analysis (weekly/monthly smoothing) • Decomposition (classic & STL) with critical interpretation • Stationarity diagnostics of log returns • ACF/PACF analysis of log-returns to assess linear memory • Volatility clustering and rolling risk behaviour • Calendar effect testing in mean and variance 𝐊𝐞𝐲 𝐓𝐚𝐤𝐞𝐚𝐰𝐚𝐲𝐬: • Price levels are non-stationary and trend-dominated. • Log-returns provide a more stable modeling foundation. • Volatility is time-varying (heteroskedastic), not constant. • Log-returns show weak linear autocorrelation — suggesting limited short-term memory. • No statistically significant calendar effect in mean returns 𝐍𝐞𝐱𝐭 𝐮𝐩 → Notebook 03 intent to construct feature representations that explicitly capture temporal dependence and volatility structure for downstream modeling. Tools: Python | Pandas | NumPy | Matplotlib | Seaborn | Statsmodels | SciPy | yFinance Full repository: https://lnkd.in/d7idKJVT Notebook attached as PDF for reference. + Follow Shashank Garewal for the more such data science content and insights. #TimeSeries #DataScience #Python #Forecasting #QuantFinance #MachineLearning #Analytics #StockForecasting #Microsoft
7
1 Comment -
Prominent academy
5K followers
🔹 Fractal – Databricks Interview Experience (Data Engineering Round) ------------------------------- ❌ Most candidates fail Fractal’s Databricks round — not because they can’t write code, but because they don’t understand the platform deeply enough. Fractal’s Databricks interview is all about platform mastery: clusters, jobs, Delta, governance, performance, orchestration, and real production debugging — not PySpark alone. Here are fresh, never-used-before, Databricks-only scenario questions: 2025 1️⃣ Your Databricks Job uses a job cluster with autoscaling. For the first 5 minutes, executors keep scaling up and down rapidly. What Databricks-specific misconfiguration could cause “autoscaling thrashing”? 2️⃣ A Databricks notebook runs successfully when executed manually, but fails only when triggered as a Databricks Job. What workspace-level and job-cluster-level differences will you investigate? 3️⃣ Your Delta table is ACID compliant, but you notice that DELETE operations are taking unusually long in Databricks. What internal Delta Lake + Databricks mechanisms cause slow delete performance? 4️⃣ Databricks Repos merge conflict resolution still shows old notebook versions even after resolving the conflict in Git. What Databricks-specific caching or sync behavior explains this? 5️⃣ A reviewer in Fractal asks: “How do you prove your Databricks job has end-to-end data lineage?” What tools and workspace settings in Databricks provide lineage beyond Unity Catalog? 6️⃣ A Photon cluster fails with the error: Operation requires native execution but fallback is disabled. What Databricks setting or code path can trigger this? 7️⃣ Your Databricks streaming job slows down over days even though micro-batch sizes remain stable. What Databricks workspace-level causes can lead to state store bloat? 8️⃣ A Delta Live Table (DLT) pipeline stops with the message: “Event log exceeded maximum retention limit”. What DLT-specific configuration is responsible? 9️⃣ A SQL Warehouse dashboard refresh takes 8 minutes but the underlying SQL query runs in 15 seconds. What Databricks Warehouse features can cause UI-level render delays? 🔟 Your cluster library installation fails only during job execution but succeeds interactively. What Databricks job isolation behaviors cause library install differences? 1️⃣1️⃣ You enabled Delta Universal Format (UniForm) in a table. Suddenly your S3/GCS storage costs increase. Which Databricks UniForm behavior leads to higher storage usage? 1️⃣2️⃣ Your Unity Catalog external location validation fails even though cloud permissions are correct. What workspace-level settings can still cause UC validation to break? --------------- 💬 fractal focuses on real production reliability — not just PySpark syntax. They test your ability to design resilient, compliant, and automated Lakehouse pipelines. 📞 DM or WhatsApp +91 93594 45862 to access: ✅ Real Databricks Question Bank ✅ Mock Interviews ✅ End-to-End Project Practice ✅ Pay After Placement Options
2
-
Udit Soni
ICICI Bank • 6K followers
🎬 Sentiment Analysis: Then vs. Now 👉 Input: “The movie was not bad at all.” 👉 Goal: Predict Sentiment → Positive 🔹 How Traditional NLP Worked 1️⃣ Feature Extraction → Used Bag of Words / TF-IDF. ⚠️ Issue: Words treated independently. “bad” = negative, even in “not bad.” 2️⃣ Machine Learning Models → Logistic Regression, SVM on handcrafted features. ✅ Simple, but only as smart as the features we designed. 3️⃣ Deep Learning (RNNs, LSTMs) → Added sequential context. ⚠️ Limitation: Struggled with long sentences, sarcasm, nuance. Result: Often failed to capture meaning in real context. 🔹 How LLMs Changed the Game 1️⃣ Tokenization → Break text into subwords (e.g., “unbelievable” → “un”, “believe”, “able”). 2️⃣ Embeddings → Map tokens into vectors that capture semantic relationships. 3️⃣ Transformers → Process the entire sequence in parallel using self-attention. 4️⃣ Attention Mechanism → Learns which words matter. ✔️ Example: understands “not” flips the meaning of “bad.” 5️⃣ Scaling Up → Trained on massive datasets with billions of parameters. LLMs (GPT, LLaMA, PaLM) become general-purpose foundation models. ✅ Result: LLM correctly predicts sentiment = Positive because it captures context + meaning. 📌 Takeaway Traditional NLP = brittle, task-specific. LLMs = contextual, scalable, general-purpose. They’re now the backbone of summarization, translation, coding, Q&A, and more. 💡 Terms like tokenization, embeddings, attention, transformers, fine-tuning, zero/few-shot will be unpacked in upcoming posts. Stay tuned! Follow Udit Soni for more! #GenAI #LLM #NLP #ArtificialIntelligence
14
-
Abhishek Mungoli
InMobi • 50K followers
COVID-19 brought abrupt changes in consumer behavior and impacted the accuracy of forecasting models. Similarly, UPI (Unified Payment Interface) abrupted consumer's spending methods, bringing many of them online. Concept drift happens when the statistical properties of the target variable, which the model is trying to predict, itself change over time. This causes problems because the predictions become less accurate and become unreliable. Here's a short video for Data Science enthusiasts I have created on the topic of Data drift and model monitoring followed by a practical exercise at the end. Like and subscribe for more such interesting concepts. Also, like and share over here for maximum reach. : ) Video Link: youtu.be/tQjRQWfYQ10 YT channel Link: youtube.com/@datatrek #datatrek #datascience #machinelearning #statistics #deeplearning #ai
3
-
Arzoo Sharma
ArthaPoint • 30K followers
Sharing my handwritten notes on Functional Forms in Regression Models ✍️📒 Choosing the right functional form is critical in econometrics – it can completely change how we interpret relationships between variables. In these notes, I’ve explained: * Why functional form matters * Linear vs log-linear vs double-log forms * Semi-log models and their interpretation * How to test and select the right form * Worked-out examples for clarity I’m attaching the notes below for easy access. These should help you quickly revise and understand how functional forms shape economic analysis. Let me know if you’d like me to continue sharing handwritten notes on other econometrics topics. Join the Arthapoint community here https://lnkd.in/g3Q3D96h
73
-
Chesta Dhingra
JLL MENA • 1K followers
Over the years working with large datasets, I’ve realized something simple: 👉 Most performance issues don’t start with queries, they start with design. That’s why I’m starting a new series on SQL Optimization & Scalability — sharing practical lessons on schema design, indexing, partitioning, and more. Here’s 💡 Part 1: Faster SQL Starts with Smarter Schema Design & Indexing. While working in today’s evolved data industry, I’ve realized one thing very clearly: 👉 A well-designed schema solves 90–95% of business problems before we even start writing complex queries. Why? Because a robust schema and efficient database design can make or break how quickly we can turn raw data into insights — whether for day-to-day reporting, powering ML models, or feeding dashboards. At first, schema design feels simple when you’re only handling a few thousand rows. But as data grows exponentially, Data Definition Language (DDL) decisions suddenly become critical. The right schema design = faster query retrieval. Efficient DDL = smoother Data Manipulation Language (DML) operations at scale. Together, they set the foundation for every downstream use case — from analytics to AI. This series will explore how schema design, indexing, partitioning, and other optimizations unlock query speed and scalability. 💡 My takeaway so far: Schema is not just a database step — it’s a business impact multiplier. Indexing: The Shortcut That Makes SQL Fly When dealing with millions of rows, scanning data sequentially (O(N)) quickly becomes a bottleneck. That’s where indexing comes in — reducing lookups to O(logN) by letting SQL jump directly to the relevant data. Indexes build sorted, searchable structures (like B-Trees, Bitmaps, Columnstores), acting as smart catalogs that avoid expensive full table scans. 🔹 Clustered Index Defines the physical row order in a sorted manner. 👉 Think of it like a diary arranged strictly by date. If you want to find March entries, you flip straight to March — without reading Jan and Feb. CREATE CLUSTERED INDEX cix_fact_transact_date on dbo.fact_transactions(transaction_date). And QUERY SELECT * from dbo.fact_transactions WHERE transaction_date between BETWEEN '2024-01-01' AND '2024-01-31'; 🔹 Non-Clustered Index Creates shortcuts for specific query patterns (great for filters like city + date). 👉 It works like a table of contents — pointing directly to the page you need. CREATE NON-CLUSTERED INDEX nix_fact_city_date on dbo. fact_transactions(city_id,transaction_date) INCLUDE (price); Query SELECT transaction_date, price FROM dbo.fact_transactions WHERE city_id = 1001 AND transaction_date >= '2024-01-01' 💡 Takeaway: 🔹 Indexes aren’t just “performance tweaks.” They’re query accelerators that transform how databases scale. 🔹 Use clustered indexes for natural order. 🔹 Use non-clustered indexes for targeted lookups. #SQL #QueryOptimization #Database #DataScience #DataEngineering
8
-
Adeel Hamid
Minesite.AI • 2K followers
🎓 Day 11: Mastering Table Calculations in Tableau 🧠 What Are Table Calculations? Table Calculations are computations applied after the data has been aggregated and displayed in the view. They operate on the result of the visualization, not the raw data. 📌 Examples: 1. % of total sales 2. Running totals 3. Year-over-year growth 4. Rank of products within category 5. Moving averages ⚙️ Why Use Table Calculations? Show % of total per sub-category, Percent of Total visit www.siradeel.blogspot.com for full details about these sections 🛠️ Basic Table Calculations (No Formula Required) Tableau gives you built-in table calcs through right-click: Try This: Open a new sheet Drag Sub-Category to Rows Drag Sales to Columns Now: Right-click on SUM(Sales) in the view Choose Quick Table Calculation Select: ✅ Percent of Total ✅ Running Total ✅ Rank 🔍 You’ll see how numbers change instantly — that’s a table calc! 📐 Advanced Table Calculations (Using Functions) Let’s create our own calculation now. ✍️ Example 1: Calculate Percent of Total Manually Create a new calculated field: tableau CopyEdit SUM([Sales]) / TOTAL(SUM([Sales])) This divides each row’s sales by the total across the table. Tableau will ask: “Compute using?” ➤ Choose the direction (Table Across, Down, Pane, etc.) 🎯 Understanding Compute Using visit www.siradeel.blogspot.com for full details about these sections 🎓 Tip: Always check "Edit Table Calculation" to set this properly! 🧪 Common Table Calc Functions visit www.siradeel.blogspot.com for full details about these sections 🧪 Hands-on Example: Year-over-Year Growth (YOY) Goal: Compare this year’s sales to last year’s Steps: Drag Order Date to Columns → Choose Year Drag Sales to Rows Create a new calculated field: (SUM([Sales]) - LOOKUP(SUM([Sales]), -1)) / LOOKUP(SUM([Sales]), -1) Format it as percentage 🔍 This shows YOY growth using LOOKUP() to reference the previous year! 🧪 Challenge Task (Try Now!) Goal: Show top 5 sub-categories by running sales total Create a bar chart: Sub-Category vs Sales Sort it in descending order Create a new field: RUNNING_SUM(SUM([Sales])) Add a filter to only show top 5 using INDEX() <= 5 This teaches you to combine ranking + window calculations. 🔄 Table Calculation Best Practices ✅ Use "Edit Table Calculation" to control computation direction ✅ Use tooltips to explain what each number means ✅ Combine with filters for interactive dashboards ✅ Be cautious when changing table structure — table calcs can break
3
1 Comment -
Geomeife Elizabeth
3MTT Nigeria • 46 followers
🌟 Day 7 — The Process Matters I love CRISP-DM because of its smooth flow in data analysis. It involves six steps: 1️⃣ Business Understanding 2️⃣ Data Understanding 3️⃣ Data Preparation 4️⃣ Modeling 5️⃣ Evaluation 6️⃣ Deployment Most analysts stop at data preparation, bringing insights that guide decision-making — but that’s just the beginning for data scientists! 🚀 📖 “The end of a matter is better than its beginning.” (Ecclesiastes 7:8) Every stage counts — both in data and in life. 💪 How true is this? 🤔 #15DaysChallengeWithDataLiving #15DaysOfConsistencyWithDataLiving #CRISPDM #DataAnalytics
7
-
Mohammed Arsalan
T-Systems ICT India Pvt. Ltd. • 22K followers
Sarvam-M: A 24B Parameter Multilingual AI Model is finetuned Mistral-Small with Indic language desi nuances covering 11 major languages from india 🇮🇳 Key Features: • Dual-mode interface: Quick "non-think" responses + detailed "think" mode for complex reasoning 🧠 • Strong performance on math (GSM-8K) and coding (SWE-Bench) benchmarks 📊 • Supports both native Indic scripts and romanized text ✍️ • Built for real-world multilingual conversational agents 💬 I think it's a good start , instead of training model from scratch Sarvam focused on building indic dataset and model like in video shows good translingual conversational capability . Try it out via the Colab link - https://lnkd.in/gzcW4hru Playground - https://lnkd.in/gXijZrkY
27
1 Comment -
Anjali Bhargava
Fireblaze Technologies… • 1K followers
📊 Turning Data into Growth Stories | Sales Performance Dashboard 🚀 This snapshot proves why analytics is the backbone of smart business decisions 👇 🔹 $2.33M Total Sales with strong YoY growth 🔹 $292.3K Profit → profitability moving in the right direction 🔹 Returns under control at just 5.81% 🔹 Consistent upward sales trend vs previous year 🔹 Top 5 states driving the majority of revenue 🔹 Technology & Office Supplies leading category performance 📈 What stands out? It’s not just higher sales—it’s better margins, controlled returns, and regional focus. 💡 Key takeaway for leaders & analysts: ✔ Track YoY performance, not just totals ✔ Focus on profitable categories, not only volume ✔ Optimize regions that consistently outperform ✔ Use data to reduce returns and increase trust Data doesn’t just report the past — it shapes the future. If you’re passionate about data-driven growth, this is your kind of dashboard. #DataAnalytics #BusinessIntelligence #SalesAnalytics #PowerBI #DashboardDesign #DataDriven #GrowthMindset #AnalyticsTrends #LinkedInAnalytics
32
-
Upsynz.
63 followers
Polynomial Regression helps us understand non-linear relationships where straight lines fail 📈 By transforming features into polynomial terms, we can model real-world trends more accurately. Perfect for salary prediction, demand forecasting, and growth analysis 🚀 Save this post to revise Machine Learning basics! #PolynomialRegression #MachineLearning #DataScience #SupervisedLearning #RegressionAnalysis #MLBasics #AI #PythonForDataScience #DataAnalytics #LearnMachineLearning #Statistics #DataScienceStudent
1
-
Scouterly
225 followers
Open data science roles = stalled ML roadmaps. In India, the average data scientist role stays open for 120+ days — especially in BFSI, retail, and healthcare firms. That delay doesn’t just affect hiring metrics. It impacts: 🧠 ML model deployment timelines 🔄 Business insights delivery 💰 Project ROI One of our clients in Bengaluru lost a full quarter waiting for the “perfect” data scientist. The irony? They had to restart the model training anyway because priorities shifted. At Scouterly, we help you solve the core issue: ✅ Better role scoping ✅ Faster shortlisting from pre-vetted pools ✅ Guidance on realistic market expectations You don’t need 200 resumes. You need 2 right fits. Fast. Because in ML, every extra week of delay makes your data — and your assumptions — stale.
-
Beyz
285 followers
Ace Your Next Interview: Real Data Scientist Interview Questions at Databricks [check out interviewquestionbank.com] Landing a Data Scientist role at Databricks is highly competitive. Preparation is key to showcasing your skills and experience. Python Proficiency 🐍: * Tell me about your experience with different Python libraries like Pandas, NumPy, and Scikit-learn. What are your favorites and why? * Explain a time you had to debug a complex Python script. What was your approach? Data Wrangling & Analysis 📊: * Describe your process for cleaning and preparing a messy dataset. What tools and techniques do you typically use? * Walk me through your approach to analyzing a large dataset with millions of rows. How would you handle memory limitations? Machine Learning Modeling 🤖: * Explain the difference between supervised and unsupervised learning. Give examples of algorithms for each. * Describe a time you built a machine learning model. What challenges did you encounter and how did you overcome them? Ready to conquer your Databricks interview? Head over to InterviewQuestionBank.com to explore hundreds of real interview questions for Data Scientist roles and more! #interview #AI #Databricks #DataScientist #jobsearch #DataScience #machinelearning Let's get you hired!
-
Prateek Majumder
Reliance Industries Limited • 5K followers
Just came across this gem again — Forecasting: Principles and Practice, The Pythonic Way. https://otexts.com/fpppy/ It was first introduced to us by Prof. Anant Agarwal during my Postgraduation for the Time Series Analysis subject. Now that it's available in Python too, it's a powerful resource for anyone working on forecasting and time series analysis. Highly recommend for data folks! 🔍📈 #TimeSeries #Forecasting #Python #DataScience
21
1 Comment -
Swetha C S
LTIMindtree • 5K followers
Column-based vs Row-based file formats — a simple example While learning about column-based file formats like Parquet, I finally understood why they are so efficient for analytics. Sharing the understanding with a very simple example 👇 Imagine this data: Name, City, Salary A, BLR, 50K B, BLR, 60K C, BLR, 55K Now the task is to calculate average salary. Row-based format (CSV): Data is stored like this: A, BLR, 50K B, BLR, 60K C, BLR, 55K Even though we need only Salary, the system still reads: Name City Salary ❌ Extra data read ❌ Slower analytics Column-based format (Parquet): Data is stored like this: Name → A, B, C City → BLR, BLR, BLR Salary → 50K, 60K, 55K Now, to calculate average salary: Only the Salary column is read ✅ Less data read ✅ Faster processing ✅ Better performance Bonus insight: Repeated values (like BLR) are compressed much better in column-based formats, reducing storage and improving speed. Key takeaway: Row-based formats → Best for transactions Column-based formats → Best for analytics & reporting This clarity really helped me today—sharing in case it helps someone else too 😊 Codebasics Dhaval Patel Hemanand Vadivel NAVEEN S #DataAnalytics #Parquet #DataEngineering
9
-
Natlov Technologies Pvt Ltd
58K followers
📊 Unlocking Insights with Tableau: Turning Raw Data into Smart Decisions In the era of big data, making sense of information is crucial—and Tableau makes that easy. At Natlov Technologies Pvt Ltd, we empower teams to turn complex datasets into clear, actionable insights. 🔍 Why Tableau? ✅ Drag-and-drop dashboards ✅ Real-time data visualization ✅ Connects with Excel, SQL, Redshift, Snowflake & more ✅ Interactive, shareable reports ✅ Drives faster, smarter business decisions 💡 Whether it’s customer behavior, sales trends, or operations—Tableau helps decision-makers see the story in their data. 🚀 Pro Tip: Supercharge Tableau with Delta Lake, Redshift, or Azure Data Factory to build powerful BI pipelines. 📣 Are you using Tableau to power your insights? Share your favorite use case below! 👇 #Tableau #DataVisualization #Analytics #BusinessIntelligence #NatlovTechnologies #DashboardDesign #DataDriven #BI #DataAnalytics #SmartDecisions
1
-
Abhishek Mishra
Taxila Business School • 1K followers
https://lnkd.in/gY7v8Sjj From Data to Decisions: A Story-Driven Sales Analysis with Tableau Public Today, I’m excited to share a Tableau Public Sales Story I created — a narrative-driven analysis that turns raw sales data into actionable insights. In today’s business world, a chart alone isn’t enough — context and narrative are key to making data meaningful. That’s where Tableau Stories shine: they combine interactive visualizations with captions to guide your audience through your thought process — just like a presentation, but richer and more interactive. Here’s what I focused on: 🔹 Understanding Overall Sales Performance: The opening view provides high-level insight into total sales across categories — a foundation for deeper analysis. 🔹 Analyzing Trends Over Time: By exploring sales and profit trends over multiple periods, we uncover patterns that may relate to seasonality or strategy changes. 🔹 Breaking It Down by Category & Region: Segmenting sales by product categories and regions helps reveal which areas are driving growth and which need attention. 🔹 Highlighting Key Opportunities & Challenges: Through the story’s progression, we surface where profit margins are narrowing despite strong sales and where specific sub-segments may be underperforming. 💡 What makes a story different from a dashboard? It’s not just “what the numbers are” — it’s how they evolve, compare, and connect. Each story point builds on the previous one to enhance understanding. #DataVisualization #Tableau #Analytics #BusinessIntelligence #DataStorytelling
20
3 Comments -
Dr. Pankaj Sharma
Simplilearn • 3K followers
🚀 Day 24 – Weaviate RAG over PDFs 🚀 Today, I explored building a Retrieval-Augmented Generation (RAG) pipeline using Weaviate and LangChain! Key highlights: Loaded and chunked PDF documents for semantic search. Used OpenAI embeddings (text-embedding-3-small) to create vectors for each document chunk. Stored and queried vectors in Weaviate, enabling fast, scalable retrieval. Built a RAG chain with ChatOpenAI (gpt-3.5-turbo), allowing context-aware answers from documents. Tested queries like summarizing documents and extracting key facts. This approach makes it easy to turn unstructured PDFs into a knowledge base you can query naturally. 💻 GitHub repo with code: https://lnkd.in/guSRDkFM #GenerativeAI #LangChain #Weaviate #RAG #OpenAI #AI #NLP #Python #MachineLearning #KnowledgeManagement
24
-
ArthionAI
24 followers
𝗔𝗿𝗼𝘂𝗻𝗱 𝟭𝟱 𝗱𝗮𝘆𝘀 𝗮𝗴𝗼, our team RAHUL KUMAR 𝗮𝗻𝗱 Soumyaranjan P. 𝘀𝘁𝗮𝗿𝘁𝗲𝗱 𝘄𝗶𝘁𝗵 𝗮 𝘀𝗶𝗺𝗽𝗹𝗲 𝗶𝗱𝗲𝗮: 𝗳𝗶𝗻𝗮𝗻𝗰𝗲 𝗶𝘀 𝘁𝗼𝗼 𝗳𝗿𝗮𝗴𝗺𝗲𝗻𝘁𝗲𝗱. Market data in one place. Strategies somewhere else. Analysis scattered across tools. So we asked a blunt question: Why can’t all of this live in one AI-powered workspace? That’s how Arthion was born. We’re building Arthion to simplify finance—market analysis, market-making strategies, real-time updates, and AI-driven insights, all under one roof. In this video, we’re showcasing some of the core capabilities we’ve already shipped: A full-screen terminal experience Time-series graphs with infinite zoom High-performance visualizations for deep market exploration A live demo of our AI agent performing real-time market analysis For example, we asked our agent to find the top 5 ETFs by trading volume, and here’s what it returned—with full timestamps: 𝗦𝗢𝗫𝗦 (𝗣𝗿𝗼𝗦𝗵𝗮𝗿𝗲𝘀 𝗨𝗹𝘁𝗿𝗮𝗦𝗵𝗼𝗿𝘁 𝗦𝗲𝗺𝗶𝗰𝗼𝗻𝗱𝘂𝗰𝘁𝗼𝗿𝘀) 𝗣𝗿𝗶𝗰𝗲: $𝟮.𝟰𝟱 (+𝟮.𝟵𝟰%, +$𝟬.𝟬𝟳) 𝗩𝗼𝗹𝘂𝗺𝗲: 𝟮𝟲𝟵,𝟭𝟲𝟵,𝟮𝟵𝟱 𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽: 𝟮𝟬𝟮𝟲-𝟬𝟭-𝟬𝟴𝗧𝟬𝟴:𝟬𝟰:𝟱𝟳𝗭 𝗭𝗦𝗟 (𝗣𝗿𝗼𝗦𝗵𝗮𝗿𝗲𝘀 𝗨𝗹𝘁𝗿𝗮𝗦𝗵𝗼𝗿𝘁 𝗦𝗶𝗹𝘃𝗲𝗿) 𝗣𝗿𝗶𝗰𝗲: $𝟰.𝟮𝟬 (+𝟳.𝟰𝟮%, +$𝟬.𝟮𝟵) 𝗩𝗼𝗹𝘂𝗺𝗲: 𝟭𝟮𝟯,𝟬𝟱𝟯,𝟰𝟭𝟲 𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽: 𝟮𝟬𝟮𝟲-𝟬𝟭-𝟬𝟴𝗧𝟬𝟴:𝟬𝟰:𝟱𝟴𝗭 𝗧𝗭𝗔 (𝗗𝗶𝗿𝗲𝘅𝗶𝗼𝗻 𝗗𝗮𝗶𝗹𝘆 𝗦𝗺𝗮𝗹𝗹 𝗖𝗮𝗽 𝗕𝗲𝗮𝗿 𝟯𝘅) 𝗣𝗿𝗶𝗰𝗲: $𝟲.𝟳𝟮 (+𝟬.𝟳𝟱%, +$𝟬.𝟬𝟱) 𝗩𝗼𝗹𝘂𝗺𝗲: 𝟭𝟭𝟮,𝟵𝟲𝟳,𝟯𝟭𝟮 𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽: 𝟮𝟬𝟮𝟲-𝟬𝟭-𝟬𝟴𝗧𝟬𝟴:𝟬𝟱:𝟬𝟬𝗭 𝗦𝗟𝗩 (𝗶𝗦𝗵𝗮𝗿𝗲𝘀 𝗦𝗶𝗹𝘃𝗲𝗿 𝗧𝗿𝘂𝘀𝘁) 𝗣𝗿𝗶𝗰𝗲: $𝟳𝟬.𝟵𝟲 (-𝟯.𝟳𝟯%, -$𝟮.𝟳𝟱) 𝗩𝗼𝗹𝘂𝗺𝗲: 𝟵𝟮,𝟭𝟰𝟭,𝟲𝟴𝟱 𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽: 𝟮𝟬𝟮𝟲-𝟬𝟭-𝟬𝟴𝗧𝟬𝟴:𝟬𝟱:𝟬𝟭𝗭 𝗦𝗢𝗫𝗟 (𝗗𝗶𝗿𝗲𝘅𝗶𝗼𝗻 𝗗𝗮𝗶𝗹𝘆 𝗦𝗲𝗺𝗶𝗰𝗼𝗻𝗱𝘂𝗰𝘁𝗼𝗿𝘀 𝗕𝘂𝗹𝗹 𝟯𝘅) 𝗣𝗿𝗶𝗰𝗲: $𝟱𝟮.𝟮𝟵 (-𝟯.𝟭𝟴%, -$𝟭.𝟳𝟮) 𝗩𝗼𝗹𝘂𝗺𝗲: 𝟲𝟰,𝟵𝟴𝟯,𝟳𝟮𝟬 𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽: 𝟮𝟬𝟮𝟲-𝟬𝟭-𝟬𝟴𝗧𝟬𝟴:𝟬𝟱:𝟬𝟯𝗭 This is just the beginning. We’re actively improving data accuracy, expanding integrations, and shipping new features at speed. The goal is simple: make institutional-grade market intelligence accessible, fast, and intuitive. If this resonates, 𝗷𝗼𝗶𝗻 𝘁𝗵𝗲 𝘄𝗮𝗶𝘁𝗶𝗻𝗴 𝗹𝗶𝘀𝘁 𝗮𝘁 👉 https://arthionai.app and be among the first to try Arthion. We’re just getting started.
4
1 Comment
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore More