Sign in to view Siddharth’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
San Francisco, California, United States
Sign in to view Siddharth’s full profile
Siddharth can introduce you to 10+ people at Lyft
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
3K followers
500+ connections
Sign in to view Siddharth’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Siddharth
Siddharth can introduce you to 10+ people at Lyft
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Siddharth
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Siddharth’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Activity
3K followers
-
Siddharth Patil shared thisV v cool Des!Siddharth Patil shared thisNew year, new podcast from Lyft! Our own Desmond Chan is featured in the latest episode of the Mavens of Data (Maven Analytics) podcast, hosted by Kristen Kehrer! Catch this interactive dialogue as Desmond shares his insights on the realities of building data products at Lyft, where spatial and temporal data are core to the business. Listen if you want to learn more about: 1) What makes spatial and temporal data uniquely hard at scale 2) How Lyft’s Driver team is structured and what they’re focused on 3) How Desmond thinks about building product sense on his team If you’ve ever wondered what working in a large tech organization actually feels like, or how data teams operate when the product lives in the real world, this episode is for you! Listen on Spotify: https://lnkd.in/ej7tVy4E Or catch the live action on Youtube: https://lnkd.in/eh2X8NfD #Lyft #LyftTech #RideSharing #Analytics #DataScienceInside Data at Scale: What It’s Really Like Working on the Driver Team at Lyft | Mavens of DataInside Data at Scale: What It’s Really Like Working on the Driver Team at Lyft | Mavens of Data
-
Siddharth Patil reposted thisSiddharth Patil reposted thisExcited to share our latest blog post on the challenges of matching riders and drivers in ride-sharing—and the factors that can lead to suboptimal outcomes in real-time decisions. https://lnkd.in/eqsaR4MZ
-
Siddharth Patil reposted thisSiddharth Patil reposted thisFor the people who wants to join Lyft or starting their journey, sharing an inspiring read from the Lyft Engineering blog: “Intern Experience at Lyft”, authored by Data Scientists Morteza Taiebat and Han Gong, both of whom began their Lyft careers as interns and returned full-time. https://lnkd.in/ggNeAC7H Have any question feel free to ping!!
-
Siddharth Patil reposted thisSiddharth Patil reposted thisHappy to share that I've published my first blog post on the Lyft Engineering Blog: My Starter Project on the Lyft Rider Data Science Team The blog outlines my experience using Causal Inference to estimate long-term effects on rider behaviour. https://lnkd.in/gt_4FRTs #Lyft #DataScience #CausalInference #MachineLearningMy Starter Project on the Lyft Rider Data Science TeamMy Starter Project on the Lyft Rider Data Science Team
-
Siddharth Patil reposted thisSiddharth Patil reposted thisNew podcast drop from Lyft! Our own Rakesh Kumar is featured in the latest episode of the MLOps community podcast, hosted by Demetrios Brinkmann! This time, the spotlight is on real-time feature generation at scale, a critical component behind Lyft’s real-time forecasting systems. Real-time feature engineering is mission critical for Lyft’s dynamic marketplace. Rakesh shares how his team built a system capable of processing tens of millions of events per minute with low latency and high reliability. Listen if you want to learn more about: ✅ The evolution of Lyft’s real-time ML infrastructure - from a naïve pipeline to high-throughput systems ✅ Design trade-offs and lessons learned in building low-latency, production-grade data pipelines ✅ How MLOps enables monitoring, deployment, and lifecycle management of real-time features ✅ Managing complexity: data contracts, config management, online/offline workflows, and more ✅ Hot shards, latency tricks, and feature bundling strategies to keep costs in check Listen if you're someone working on scaling streaming systems, building ML platforms, or working on feature engineering in fast-changing environments. Listen on Spotify: https://shorturl.at/UhQxR Watch the full episode: https://shorturl.at/E3uQx This pairs perfectly with Josh Xi's episode ( https://shorturl.at/c4nXx) on model architecture trade-offs for real-time forecasting — catch that if you haven’t already. #Lyft #MLOps #RealTimeML #FeatureEngineering #Forecasting #MachineLearning #LyftTech
-
Siddharth Patil shared thisSo good! Lots to learn! 💙Siddharth Patil shared thisLyft has a new podcast fresh off the press! Join my colleague Josh Xi for a faceoff between #DNN and Time Series modeling on MLOps community. Josh and host Demetrios Brinkmann discussed #realtimeforecasting use cases in #Lyft Rideshare. Tune in to learn how data trends impact forecasts, and get expert tips on choosing the right model for your business needs. This episode is packed with technical insights and practical advice for anyone working with machine learning and forecasting at scale! Listen here on Spotify: https://shorturl.at/UhQxR Or watch the video here: https://shorturl.at/k35lY We will have a complementary MLOps episode on Lyft's real-time feature generation in a couple of months, so stay tuned for that too!Real-Time Forecasting Faceoff: Time Series vs. DNNs - Video | MLOps CommunityReal-Time Forecasting Faceoff: Time Series vs. DNNs - Video | MLOps Community
-
Siddharth Patil shared thisPlease give it a listen! I loved it!Siddharth Patil shared thisJoin my colleague Rachita Naik as she shares her thoughts on the DataCamp podcast. The podcasts deep-dives into #machinelearning use cases in Lyft Rideshare and ETA reliability prediction. In addition, Rachita also shares some challenges she sees in deploying ML models, and offers tips on how to stay updated with #ML/#AI trends. It's packed full of insights! Listen here! https://lnkd.in/e62dYaeUMachine Learning for Ride Sharing at Lyft, with Rachita Naik, ML Engineer at LyftMachine Learning for Ride Sharing at Lyft, with Rachita Naik, ML Engineer at Lyft
-
Siddharth Patil shared thisThis is worth a listen, from our folks at Lyft! Cheers.Siddharth Patil shared thisCheck out this week's episode of #AdventuresinMachineLearning with Deeksha Goyal and Michael Sun #ML: Unraveling the Complexities of Model Deployment in Dynamic Marketplaces https://lnkd.in/gbWYdk_B
-
Siddharth Patil shared thisI am beyond excited for this!Siddharth Patil shared thisInterested in learning from data science experts across the tech industry? Join us for Lyft's Data Science Conference in San Francisco on March 14, featuring a series of talks from Tao Y. at Amazon, Anahita Hassanzadeh at Lyft, and Kasia Rachuta at Square, plus networking opportunities. RSVP today: https://lnkd.in/d2dGtjv2
-
Siddharth Patil reacted on thisSiddharth Patil reacted on thisToday we’re introducing Fabi.ai 2.0 - Your AI analyst for ALL your data. When we talk to teams we noticed they were running into the same problems we had: Getting any sort of insight to help drive the business requires wrangling data across applications, copy-pasting SQL queries and pivot tables and wrestling with expensive, clunky legacy BI. AND you need a PhD in data engineering. In the age of AI there’s a better way. A MUCH better way. Fabi connects to all your data sources (Postgres, Snowflake, Stripe, HubSpot…), learns the context about your business and empowers everyone in your organization to perform data analysis at all levels to turn your insights into dashboards or automated workflows in minutes. Now, with Fabi, insights that can actually help you make better product, go-to-market and strategic decisions are one prompt away. We’re entering the golden age of data analysis and we’re excited to be making Fabi 2.0 generally available. Come try it out for yourself and experience the magic!
-
Siddharth Patil reacted on thisSiddharth Patil reacted on thisMost dashboards I’ve built ended up in the same place: the dashboard graveyard. As someone who’s bounced between data analysis and product roles, I’ve seen how powerful traditional BI tools (Metabase, Power BI, Looker, etc.) can be. But I’ve also hit the same walls over and over again: – High license + infra costs as you scale – Security/compliance overhead and admin work – Constant maintenance just to keep dashboards “not broken.” – Limited AI that still requires a data team in the loop – Insights trapped in dashboards instead of showing up where people actually work I’ve tried out Fabi.ai recently and what finally clicked for me is the new concept of moving away from “just more dashboards” toward a total AI-native analytics workflow. Instead of only visualizations, I can now: – Use natural language to generate queries in SQL and other advanced analysis methods in Python like K-Means, XGBoost ( would love cross session/project memory in the future!) Technical folks can easily inspect and edit this code. – Generate Python dashboards when I need analytical flexibility for ad-hoc analysis – Create workflows that push my reports and alerts back into Slack, email, or Google Sheets ( would love bidirectional sync with project management tools etc in the future!) – Easy connections to warehouses + tools so you don’t live in CSV hell It's not just about upgrading BI tools anymore, but moving to AI-native analytics that puts the right answer in front of the right person at the right time, and gives everyone (technical or not) their time back. If you’ve hit the limits of Metabase or other legacy BI tools in your own data and toolstack or PM work, Fabi is worth a shot! #dataanalytics #productmanagement #businessintelligence #AIBI
-
Siddharth Patil reacted on thisSiddharth Patil reacted on thisIf you're reading this, chances are you do not need a semantic layer... Why? 98% of data teams lack foundational models. And, if you lack proper data modeling of business processes, you have no business building a semantic layer. If you want to build one after, fine. Until then, modeling your data at the right grain, building additive measures, and having conversations with business stakeholders on how these processes work will get you to where you need to be. This alone will allow you to start extracting value from AI-enabled BI like Fabi.ai and other tools that 10x your insights without additional overhead. No new, expensive system needed.
-
Siddharth Patil reacted on thisToday is a big one. This morning, we launched BADAS — Beyond ADAS — Nexar’s real-world foundation model that predicts collisions before they happen. For Nexar Inc., this is a defining moment. And for me personally, it’s a reminder of why we started this journey: to make every mile safer, smarter, and more human. Traditional ADAS systems are good at reacting. BADAS is built to anticipate. It’s being trained not in lab simulations or curated test tracks, but on 10+ billion miles of actual human driving — millions of close calls, sharp turns, and moments where instinct and intelligence meet. The result? Fewer false alerts. Earlier predictions —> 3–5 seconds before impact. Models that finally understand what’s really dangerous to the driver, not what just looks dangerous to a computer. That’s the technical story. But the real story is what this means for the road ahead. The entire safety ecosystem — from carmakers to regulators to AV startups — is standing at the same crossroads: who decides when an AI is safe enough to drive? Our answer: the road itself. BADAS isn’t just a product; it’s proof that real-world data must be the benchmark for every model that claims to protect lives. It’s the foundation for how the world will test, validate, and trust the next generation of autonomy — and Nexar is proud to set that standard. We believe every company that truly cares about safety will use real-world data as part of their stack. Not because we say so, but because physics and humanity demand it. That’s why I’m so energized today. This launch isn’t the finish line — it’s the ignition point for a movement. A future where: Safety isn’t a feature — it’s the foundation. The best companies compete on responsibility, not just performance. Transparency and validation are built into the process, not bolted on at the end. At Nexar, we’re here for the builders, regulators, and innovators who believe in safety at scale. If that’s you — let’s talk. The world’s roads are ready to teach us more than any simulation ever could. #RealWorldAI #BADAS #ADAS #SafetyAtScale #MobilityIntelligence #VisionZero #AutonomousVehicles #EdgeAISiddharth Patil reacted on thisOur real-world data just beat the SoTA incident prediction models. Meet BADAS 1.0! 🔥 For decades, safety was measured by how fast we react. Today, BADAS (Beyond ADAS) changes the equation. Trained on 10B+ real-world miles and 60M+ annotated driving events from 350K+ connected vehicles, BADAS learns from what truly happens on the road, not what’s imagined in a lab. This is Beyond ADAS — a new class of Real-World AI delivering unprecedented prediction accuracy and behavioral foresight. It’s where mobility intelligence meets the unpredictability of real life. 🚀 Discover what BADAS can do for you: https://www.nexar-ai.com/ #BADAS #AI #ADAS #AutonomousVehicles #RoadSafety
Experience & Education
-
Lyft
*** ** ***********
-
*******
** ******** ** **** ******** *********
-
*******
**** ** **** ******** ********* ********
-
********** ** *********** ********
*** ******** ******* undefined
-
-
****** ********* ** *********** ******
********* undefined
View Siddharth’s full experience
See their title, tenure and more.
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Recommendations received
13 people have recommended Siddharth
Join now to viewView Siddharth’s full profile
-
See who you know in common
-
Get introduced
-
Contact Siddharth directly
Other similar profiles
Explore more posts
-
Fernando Migone, PhD
Welocalize • 6K followers
My team is on a roll!! In our latest research, we dug into one of the hardest challenges in AI today: causal reasoning — and exposed major blind spots in how state-of-the-art models handle complex, real-world causation. 💡 Most evaluations force models to choose one answer. But real-life causes aren’t that simple. So we tested 8 frontier LLMs using multi-select (polytomous) tasks across 576 scenarios. The results? Both surprising and sobering. 🔍 Key insights: - Models adopt a conservative strategy when identifying causes but over-select confounders—revealing asymmetries in reasoning across task types. - Chain-of-thought prompting? 🚫 Not helping here. - How you score these evaluations can change which model looks best: Our results show that rankings shift significantly depending on the scoring metric used. - We uncovered behavioral signatures by model family: some are cautious and precise, others more liberal and prone to over-selection. ⚠️ The implications for AI in healthcare, law, and policy are huge. Missing causal factors isn’t just suboptimal — it’s dangerous. 🧠 This research doesn't just offer a new benchmark. It offers a new lens for understanding model behavior under uncertainty and instruction complexity. 🔗 Full methodology, scoring analysis, and model comparisons https://lnkd.in/gvTEDhej Major kudos to the team: David Harper, Konstantinos Karageorgos & Abigail Thornton, PhD How are you measuring causal reasoning in your AI systems? #AI #LLMs #CausalReasoning #AIEvaluation #ResponsibleAI #WeloData #MachineLearning #AIRisk #ModelSelection #AIResearch
34
1 Comment -
Galen Marchetti
AstroBee • 5K followers
AstroBee generates ontologies...but what is an ontology, and if they're important, why don't you already have one? I've written a short piece on Kimball star schemas, ontologies, and how each relate to Bill Inmon's original approach to data warehousing in the early 90s. Dig in to find an overview of a fascinating technological debate and the re-emergence of an approach to data modeling that's been sleeping for 20 years.
18
3 Comments -
Walter Maguire
2K followers
Need to apply the power of graph to your business? Tired of needing an army of data scientists to do it? Rocketgraph solves this. But seeing is believing. So I've recorded a set of two-minute tutorials to highlight the platform and how it creates a user-experience that will enable your business analysts to use graph...without being data scientists. This video kicks the series off with our POV on the state of graph tools and sets the context for the rest of the tutorials. #rocketgraph #AI #graph https://lnkd.in/gapEpnC8
10
-
Noah Fenn
LiveRamp • 4K followers
Just read a compelling article from my colleague Noel McMichael that nails an AI problem I've been thinking about a lot lately. As we move toward deploying autonomous, API-using agents, we're hitting a wall. These agents are active and adaptive, yet they often lack a verifiable identity, a complete audit trail, or clear revocation logic. This makes them a black box, which creates risk for companies of all sizes and scopes. Noel warns that the biggest risk isn't only technical... it's "institutional paralysis". Companies that can't verify their AI's behavior will refuse to deploy them for anything important. This will likely create a two-tier system where only organizations with resources to build custom solutions benefit from AI, while everyone else operates in an environment of mistrust. His solution is a "trust fabric for AI", which is essentially a unified layer that sits below the application and above the model. This fabric would bring together the technical primitives that already exist, like persistent identity, scoped permission sets, verifiable provenance, and credential chains, to make agent behavior verifiable and secure. I really like Noel's take and highly recommend reading the full article (link in the comments section) #AI #AgenticAI #Trust #DataGovernance #AIEthics
21
2 Comments -
Dr Haluk D.
Amazon • 6K followers
A 12-month professional, voluntary collaboration has finally borne fruit. This morning, I was honored to see the University of Washington News (https://lnkd.in/gRRpDqjA) and the UW Milgard School of Business (https://lnkd.in/gKJ_pwbm) highlight our Responsible GenAI Framework. Why do we care about responsible generative AI? Because GenAI systems are no longer experimental curiosities. They shape decisions, influence behavior, generate content at scale, and increasingly affect trust, fairness, privacy, and accountability in organizations and society. Without intentional design and governance, these systems can amplify bias, create opaque risks, and erode confidence rather than enable innovation. This framework is about making GenAI useful, ethical, and sustainable at the same time. It emphasizes responsibility not as a constraint, but as an enabler of long-term value, credibility, and impact. I am grateful to the collaborators, reviewers, and institutions who supported this work, and excited to see responsible GenAI move from aspiration to practice. To learn more: - Read the RGAF Framework (PDF): RGAF Framework https://lnkd.in/gTaxnndV #GenAI #ResponsibleGenAI International Society of Service Innovation Professionals (ISSIP) Milgard CBA
23
1 Comment -
George Oates
Flickr Foundation • 1K followers
Yes, enjoyed this very much from Jed Sundwall. He's speaking plainly about how we must always think of the who will use our data and and how in the context of the "perfectly FAIR, utterly useless data" we keep spewing around everywhere. I've always been a fan of figuring *use* as a priority, in favour of the hard-to-know impact, especially when you hear factoids flying around that a very large portion of stored data—likely somewhere between 60-80%—is rarely or never accessed after initial storage.
4
-
Marc Brooker
Amazon Web Services (AWS) • 18K followers
Barbarians at the Gate (https://lnkd.in/gkEBX79s) is a very interesting new paper from some great folks at Berkeley, with some exciting results showing the potential for AI in systems research. But I think the authors aren't quite asking the hardest problem about where this takes systems as a field. I wrote a new blog post about it: https://lnkd.in/gxS3s4v4 The core point of the paper is that AI can be effective at automatically driving the 'number go up' part of systems research. Given an evaluator, AI can automate progress on implementation. I believe this is 100% correct. The paper also talks about how more focus will have to go into developing evaluators. Again, I 100% agree, but think they're understating their case. Building evaluators (basically tests) is, I believe, going to become the focus of systems and software over the next decade. Which leaves researchers with the hard problem of choosing which problems to go after. That's never been easy in academic systems, and is only going to become more of a challenge. But the best systems research is driven by choosing problems well! This vision for AI could lead to lots more of the least interesting systems research. The "number go up" kind. Which could make it harder than ever to identify the truly innovative and visionary work. We're already seeing that elsewhere in science. So, mostly, I think the paper is right, but not yet asking the biggest question: what should systems research be?
276
12 Comments -
Aatish Nayak
Harvey • 4K followers
It was a pleasure sitting down with Sachi from Sierra and Leigh Marie & Josh Coyne from Kleiner Perkins to discuss what its like shipping AI at enterprise scale. We talk about the new product principles for AI-native products incl: - Why context complete workspaces are more important than raw AI capability - The impossibility of planning roadmaps in this market - Building user trust through UX - When you need (and don't need) Forward Deployed Engineers - Re-finding PMF every few months - And why vibes are sometimes all you need. Thanks for having me on Builders Ep 2, and hope this is a useful and practical guide for anyone building in AI. Full episode link here: https://lnkd.in/ejRvsUSs P.S. Recommend all podcasts provide Chick-Fil-A to their guests :)
104
2 Comments -
Carmella (Surdyk) Weatherill
3K followers
We're excited to announce the new AI-first Colab Enterprise experience in Vertex AI and BigQuery. This powerful platform streamlines complex data science workflows, allowing you to simply prompt an agent with a request like "train a model to predict income." The agent then autonomously generates and executes a complete plan—from data loading and cleaning to model training and evaluation. It's a game-changer for accelerating your path from data to insights.
-
Alice Friend
2K followers
As an American, and a Californian, I was born into a culture of can-do optimism, with a faith in our collective ability to overcome the odds and not only survive but solve big problems. We’ve done it as a society over and over in our history. We dream, we work, we fight, we compromise and cooperate, we build, and often we rebuild. We are at our best dreaming and defying impossibility, being generous with each other and welcoming the stranger. I believe in this version of us and spent much of my career in national service because of that belief. But we are going through a very pessimistic time. We have problems as big as ever, but we don’t seem to have faith in ourselves to solve them. I spend most of my time now thinking about artificial intelligence and the problems we could solve with it. Nothing is a panacea, and it will take all of us to turn a technology into new scientific discoveries, deeper knowledge and better ways to learn it, a medical system that saves lives and supports caregivers, more meaningful work and greater prosperity. Those things are hard. But we need to do them. That’s a choice we have and can make. We have always invented our way out of problems. I am really struck by our fearfulness. Countries with far fewer advantages than we have are far more bullish on the possibilities of technology. It goes beyond AI. That’s just one small focal point. Where else do we have openings and opportunities. Where else can we roll up our sleeves and come together to fix problems? Where are we deciding to be defeated before we begin? What other possibilities are in front of us? We have proven over and over that we can rise, ultimately, to any occasion. It’s what our history has passed down to us. I’ve been thinking a lot about lines from an old Ani Difranco spoken-word poem: “Our foremothers and forefathers… came through Hell and high water/ so that we could stand here and behold breathlessly this sight/ how a raging river of tears/ is cutting a Grand Canyon of light.”
59
1 Comment -
Bassel Haidar
Booz Allen Hamilton • 5K followers
My brilliant colleague, Alison Smith, published "The Great Unscaling - Part 1". Her central thesis is deceptively simple but devastating: 𝗚𝗲𝗻𝗔𝗜 𝗯𝗿𝗲𝗮𝗸𝘀 𝘁𝗵𝗲 𝗲𝗰𝗼𝗻𝗼𝗺𝗶𝗰𝘀 𝘁𝗵𝗮𝘁 𝗯𝘂𝗶𝗹𝘁 𝗲𝘃𝗲𝗿𝘆 𝘀𝗼𝗳𝘁𝘄𝗮𝗿𝗲 𝗴𝗶𝗮𝗻𝘁 𝗼𝗳 𝘁𝗵𝗲 𝗹𝗮𝘀𝘁 𝟮𝟬 𝘆𝗲𝗮𝗿𝘀. For two decades, software meant zero marginal cost. Add a customer? No incremental expense. This made "scale at all costs" the only strategy. With GenAI, every inference costs real money. Scaling is no longer free. 𝗧𝗵𝗲 𝗺𝗼𝘀𝘁 𝗽𝗿𝗼𝘃𝗼𝗰𝗮𝘁𝗶𝘃𝗲 𝗶𝗻𝘀𝗶𝗴𝗵𝘁: While everyone chased horizontal scale through intangible assets, Meta, Alphabet, and NVIDIA were quietly building physical infrastructure. They saw something coming that others missed. Now we're watching it play out in real time: Companies either go vertical (owning the full stack for specific outcomes) or risk falling into what Alison calls "𝗧𝗵𝗲 𝗖𝗼𝗺𝗺𝗼𝗱𝗶𝘁𝘆 𝗧𝗿𝗮𝗽." 𝗧𝗵𝗲 𝗰𝗮𝘀𝗲 𝘀𝘁𝘂𝗱𝗶𝗲𝘀 𝗵𝗶𝘁 𝗵𝗮𝗿𝗱: • Metropolis didn't sell parking software; they bought a $1.8B parking operator to control the full stack • Plaid didn't try to be "data connections for everyone"; they became the indispensable operating system for fintech • The pattern is clear: 𝘀𝗽𝗲𝗰𝗶𝗳𝗶𝗰𝗶𝘁𝘆 𝗯𝗲𝗮𝘁𝘀 𝗴𝗲𝗻𝗲𝗿𝗮𝗹𝗶𝘁𝘆 𝗶𝗻 𝘁𝗵𝗲 𝗔𝗜 𝗲𝗿𝗮 Alison argues we need new metrics entirely: • Forget Total Addressable Market → Focus on Value Capture Ratio • Forget Monthly Active Users → Measure outcome-based KPIs • Add Cost per Outcome to your dashboard (or you're flying blind) For those of us implementing Agentic AI: whether in software development, business processes, or delivery models; this reframes everything. The question isn't "how do we scale this across everything?" It's "where can we become indispensable?" This is Part 1 of a 3-part series. If the next two parts are anything like this one, they'll be essential reading. 𝗢𝘂𝘁𝘀𝘁𝗮𝗻𝗱𝗶𝗻𝗴 𝗮𝗻𝗱 𝗶𝗻𝘀𝗽𝗶𝗿𝗶𝗻𝗴 𝘄𝗼𝗿𝗸, 𝗔𝗹𝗶𝘀𝗼𝗻! You've given us the framework for the next decade. Read it here: https://lnkd.in/edddWfwq #AI #GenerativeAI #AgenticAI #Strategy #ThoughtLeadership Booz Allen Hamilton
24
3 Comments -
Matthew Faenza
Boom • 604 followers
Running evals on the new Claude 4.5 Haiku this morning—seemed like a good time to share what we’ve built for evals at Boom. The core implementation: an automated evaluation pipeline that runs every supported model against an ever growing set of test cases for our critical path inference operations. Quality, latency, and cost benchmarked in real time against production workloads. But here’s what this actually delivers for our customers: **Guaranteed uptime, not just promised uptime.** When Anthropic or Google has an outage, our system automatically routes to the next best model. Our customers’ workflows don’t stop. They often don’t even notice. We’ve decoupled their reliability from any single provider’s reliability. **Immediate access to model improvements.** Claude 4.5 Haiku dropped last week. By this morning when I finally got a few minutes to add it to our model directory, I had complete performance data across our stack. If it improves customer outcomes, we can integrate it today. Our customers get the benefit of frontier model advances within hours, not months. **Protection from silent degradation.** The harder problem isn’t when services go down—it’s when they stay up but quality degrades. Our eval system monitors this continuously. If we detect accuracy dropping on a live service, we route traffic away before it impacts customer results. The chart shows what I’m looking at right now: Haiku absolutely crushing latency on most of our critical path operations (e.g. 11s vs 26-64s for other models) while maintaining 99%+ accuracy. That’s a customer experience improvement we can ship immediately. The power is in having evals that actually matter. Test cases from realistic production scenarios. Metrics that map to customer outcomes. When you can run any model against your ground truth in minutes, you stop guessing and start knowing. Model selection becomes an empirical question, not an intuitive or philosophical one. Your customers care about outcomes, not which model you’re using.
27
2 Comments -
Josh Klahr
9K followers
"AI Needs a Semantic Model" - Jamie Davidson, May 20, 2025 https://lnkd.in/dx7U4B_7 Loved this piece from our friends over at Omni - whether you want trusted and easy BI or AI - you need a #semanticmodel. Semantic models serve as the essential bridge between raw data and meaningful insights, ensuring that both AI and BI systems interpret information consistently and accurately. By embedding organizational context and definitions into the data layer, semantic models enable AI to provide trustworthy answers and empower business users to make informed decisions. Excited to see the excitement and interest in this space, and to share more about Snowflake's investment in this area at the upcoming #snowflakesummit.
145
2 Comments -
Bret Taylor
OpenAI • 139K followers
Building on top of large language models is fun, but getting consistent performance and reliability is extremely challenging. I love this post from Kimberly Patron about how Sierra uses health and performance-based traffic routing and request hedging to get meaningfully better performance and tail latency from high scale LLMs https://lnkd.in/g4qWRCb5
284
13 Comments -
Vinny DeGenova
Shopify • 3K followers
New from the Wayfair Tech Blog! My team built and shipped a multimodal GenAI system to validate product dimensions at scale. The problem sounds simple - make sure the sofa dimensions on the website match reality. But, across 30M+ products from 20K+ suppliers, it gets complicated fast. Conflicting sources, inconsistent conventions, and edge cases galore (how do you measure an inflatable palm tree?). We use Gemini to cross-reference text, images, and supplier data. Results: 85%+ precision and 70%+ recall, up from <50% precision and 7% recall with earlier classical ML approaches. ➡️ Full write-up: https://lnkd.in/e7MxEjxW #MachineLearning #GenAI #Ecommerce #DataQuality #MultimodalAI
66
5 Comments -
Adam McCabe
Convictional • 1K followers
We wrote a new essay detailing our journey towards AI powered analytics in Convictional: https://lnkd.in/gX4QCgqm TLDR: With LLMs, true self-serve analytics seemed finally within reach. No more over-busy dashboards built for everyone and exactly right for no one. Even better, the LLM could help answer questions about the data. However, a serious trust problem is introduced. Most business users are already skeptical of data and its interpretations, and when you throw in LLM hallucinations (bad SQL, incorrect join logic, or poor interpretation of data definitions) the margin of trust shrinks even further. One or two errors and the solution is written off. Data is a crucial input to decision making, so we’ve been committed to finding a solution that can provide analytics at the levels of accuracy needed. Over the last year we tried a number of approaches, with our initial achieving an unacceptable ~50% accuracy rate, and our current productionized technique approaching 100% accuracy. The key was acknowledging the human role in the solution. We found that using a semantic layer, defined by humans and queries by the LLM was the unlock. By no longer asking the LLM to write the SQL, and instead rely on robust pre-defined metrics and configurations, it could focus on actually answering the user’s questions instead of resolving SQL. If you’re using dbt’s semantic layer (or plan to) and interested in trying it out, just touch base and we can help get you set up! cc Jake Beresford Matthew H. Chequers, Ph.D.
14
-
Tom Gowan
Spire • 2K followers
As renewables become a larger share of generation, the energy industry’s reliance on accurate, low-latency weather forecasts continues to grow. At Spire, our proprietary GNSS-RO satellite data are assimilated through our data assimilation system to produce improved initial conditions, which drive state-of-the-art AI weather models. Downstream AI models then convert these weather forecasts into solar and wind power generation forecasts. As AI weather models increasingly commoditize, Spire’s proprietary space-based observations become even more valuable and directly enable higher-skill forecasts and more reliable decision-grade insights.
27
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top contentOthers named Siddharth Patil in United States
-
Siddharth Patil
Irvine, CA -
Siddharth Patil
New York City Metropolitan Area -
Siddharth P.
Champaign, IL -
Siddharth Patil
New York City Metropolitan Area
27 others named Siddharth Patil in United States are on LinkedIn
See others named Siddharth Patil