Sign in to view Scott’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Scott’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Palo Alto, California, United States
Sign in to view Scott’s full profile
Scott can introduce you to 10+ people at Distributional
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
7K followers
500+ connections
Sign in to view Scott’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Scott
Scott can introduce you to 10+ people at Distributional
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Scott
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Scott’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Websites
- Personal Website
-
http://scottclark.io
- Portfolio
-
http://github.com/sc932
Activity
7K followers
-
Scott Clark shared thisI recently attended the kickoff for the Foundation of Science and AI Research (SAIR) and got to have a great chat with Chuck Ng about the bright future that AI is unlocking for scientific study and innovation. I believe AI is going to help accelerate interdisciplinary research in a way I could have only dreamed of in grad school. The conference talks were stacked with Nobel, Turing, and Fields winners (Terence Tao, Richard Sutton, Barry Barish, and more). All available online (link in comments). It was exciting and humbling to listen and chat about all of the exciting developments and what is imminently on the horizon. I can't wait to see what SAIR continues to do and I am excited to help play a (small) role.Scott Clark shared thisScott Clark is CEO of Distributional, former VP/GM at Intel, and CEO of SigOpt (acquired by Intel) — with a PhD in Applied Mathematics from Cornell. He's the rare person who moves fluently across science, industry, and entrepreneurship, and brings genuine depth to all three. SAIR co-founder Chuck Ng spoke with Scott on what happens when AI starts doing that across every scientific discipline at once — and why he calls it a rising tide. The idea is straightforward but powerful: as AI bridges gaps between fields, every discipline gets lifted. Connections that no single researcher could make alone become possible at scale. But Scott is clear on what matters most as these systems grow more powerful — the wisdom to point them in the right direction, and the observability to know whether they're actually getting there. 🔗 Watch the full conversation: https://lnkd.in/dgrGh2RT #AIforScience #SAIR #AI #ScientificDiscovery #EntrepreneurshipEntrepreneur Scott Clark on Collaboration, Observability & Why AI Is Science's Greatest TranslatorEntrepreneur Scott Clark on Collaboration, Observability & Why AI Is Science's Greatest Translator
-
Scott Clark shared thisI gave a talk recently on the AI observability landscape and why I think better analytics is the missing part of the hierarchy. It was a fun 10 min overview of the last 2+ years of what we have learned and built at Distributional. If you have agents in production and want to find behavioral signals your evals may be missing let us know if we can help! (The product is free and deploys on-prem).Scott Clark shared this"Instead of trying to catch things pre-production, what if we had a better understanding of what was actually happening in production?" Scott Clark originally started Distributional as a testing company, but soon shifted to analytics. Learn more about his journey in the highlights from his SF AI Engineers Meetup talk hosted by Fonzi AI: https://lnkd.in/gyZk4BRY
-
Scott Clark posted thisI'll be speaking at the SF AI Engineers meetup presented by Fonzi AI this Wednesday at the House of AI alongside Wassim Gharbi of Tesla and YK Sugi of Eventual. In my 10 min talk I'll speedrun through lessons learned, mistakes made, and useful takeaways from building agent analytics at Distributional. Come by to talk AI infra and what's next for the hierarchy of agent observability! Thank you Drew Moffitt and Lauren Cotta for hosting!
-
Scott Clark posted thisI'm going to be giving a few talks this week on adaptive agent analytics. If you're building agents and want to understand and improve behavior as part of your observability stack you should check them out (and deploy our free platform!) Details below, links in comments. Talk 1: Built on Bedrock Demo Night. 5pm PT, 1/20 at the Amazon Web Services (AWS) Loft in SF, 525 Market Street. Talk 2: The Hidden Signal in Production AI Logs with Jason Liu. 10am PT, 1/21 online on Maven.
-
Scott Clark shared thisICYMI: Peeyush Aggarwal from Deloitte released a great report on holistic agent observability over the holidays. This is a must read for anyone looking to scale and improve agentic systems. The key insight is bifurcating observability into two paths: "hot path" and "warm path." Hot path gives you immediate feedback about operational signals with monitoring and alerting tools so that you can take quick action like firing off guardrails or fallback systems. Warm path aggregates agent behaviors using analytics to pull out patterns and signals that can be used for ongoing improvements of the agentic system itself by looking at a broader picture of overall usage. Both are a valuable part of a complete agentic observability solution. This reminds me of Thinking, Fast and Slow by Daniel Kahneman; you need both systems to optimally function. One can help you fix isolated, immediate issues and the other can help you find the signal in the noise to continuously improve over time.Scott Clark shared thisMaking it explainable, predictable, and controllable! Deloitte’s latest report highlights a critical shift: from passive monitoring to active governance. Here’s how forward-thinking teams are leading it: From "What happened?" to "Why?" Deep traceability, thought efficiency metrics, and warm-path analysis don’t just log errors—they reveal the reasoning behind them. Now you can trace a hallucination to its source, not just react to it. From observing to predicting Token spikes. Reasoning loops. Response drift. By analyzing patterns across sessions, we can now anticipate agent runaway, cost explosions, or performance decay—before they hit. From alerts to control Real-time guardrails. Hot-path intervention. Multi-agent validation. The human shifts from doer to supervisor, stepping in only when it matters—ensuring safety, compliance, and ethics by design. This isn’t just observability. It’s operational maturity for the agentic era, monitoring business metrics and correlation between different metrics ! Because trust isn’t built in days, but years of hard work and can be broken with one simple action ! #AI #AgentOps #Deloitte #AIObservability #ResponsibleAI #FutureOfWork #TechLeadership Galileo Distributional Sulabh Soral Atindriyo Sanyal Vikram Chatterji Scott Clark Ayush Chopra Ramesh Raskar Chris Pease Colin Payne
-
Scott Clark shared thisIt was great to chat with Helen Yu in the NVIDIA booth at AWS re:Invent about the future of agent analytics and how Distributional is teaming up with NVIDIA AI to bring understanding and continuous improvement to production AI applications. Check out our interview from the expo floor at AWS re:Invent here: https://lnkd.in/gdjpJUCx
-
Scott Clark reposted thisScott Clark reposted thisAre your AI agents still stuck in slideware or earning trust in production? At Amazon Web Services (AWS) re:Invent, I spoke with Scott Clark, CEO of Distributional and an NVIDIA Inception startup, about “agent analytics”; the third pillar of enterprise AI: build the models, run them, and truly understand them. In this special CXO Spice edition, we explored: ✅ Distributional’s origin story ✅ A customer’s “firehose” moment from POC to production ✅ How NVIDIA Inception and the NeMo Agent Toolkit close the confidence gap for AI agents in regulated industries If you’re building or deploying AI agents, it’s time to move beyond demos and gain real-world confidence. Learn more about what NVIDIA Inception can unlock for your next growth stage: https://nvda.ws/48xgtEh Watch our conversation here: https://lnkd.in/gz-edhPv Stay current with the latest trends in #Technology and #Innovation, Subscribe to 👉#CXOSpiceNewsletter https://lnkd.in/gy2RJ9xg Or 👉 #CXOSpiceYouTube here https://lnkd.in/gnMc-VpjAgent Analytics: The Third Pillar of Enterprise AIAgent Analytics: The Third Pillar of Enterprise AIHelen Yu
-
Scott Clark shared thisWe're demoing from the NVIDIA booth in the Inception Zone at re:Invent all week. If you're in Vegas this week come by and say hi!
-
Scott Clark shared thisJust got to AWS re:Invent and am excited to demo Distributional's newly relaunched product from the Inception Showcase in the NVIDIA booth all week! If you are in Vegas, stop by the NVIDIA booth and say hi! Or ping me if you want to meet up for coffee or a drink. Huge thanks to the great partnership and product teams at NVIDIA for this incredible opportunity Farshad Saberi Movahed, PhD Ashutosh Joshi Sydney Sykes Howard Wright Abhishek Sawarkar Vivienne Zhang and the incredible partnership team at Andreessen Horowitz (Efrat Noy Sharon Williams) who set up the initial meeting with Jensen Huang that kicked this all off!
-
Scott Clark liked thisI really enjoyed this conversation with Josh Grant. All our roles are changing rapidly. It’s important to take a step back and rethink what our jobs actually should be.Scott Clark liked thisThere is a version of a GTM interview where a journalist asks a practitioner to explain their job to an audience that has never done it. This is not that. George Bonaci is VP of Growth at Ramp. As a peer, I respect him and the Ramp team immensely. I sat down with him for an unfiltered conversation about growth, marketing, AI, and org design. No frameworks. No safe takes. Just two operators talking about how much the world is actually changing. His answers are uncomfortable for some. For others, they'll feel like permission. "The actual decision surface in most marketing roles is shockingly thin. Most companies disguise production as strategy — because headcount is how old-fashioned leaders measure their importance." "Day one of building from scratch? I don't hire anyone." "You have to change what the team IS before you can change what it DOES." "Agents are the unit of production. Humans are the unit of strategy." The ones who win the next three years are already thinking this way. George didn't hold back. Full conversation below
-
Scott Clark liked thisScott Clark liked thisThe right analysis can make the next steps obvious 📊😎 One of the analysis signals that DBNL computes automatically for teams is topic over all requests. After publishing logs for 7 days, teams can see the topic breakdown of all user requests of a production agent. In the case of our example outing agent, some clear usage patterns / topics emerged. An obvious improvement was to add some pre-baked prompts to the agents starting page. These prompts make it more clear for users getting started and reduce keystrokes for common queries Distributional is analytics for production agents Understand what to improve and fix next 🚀 deploy 📊 analyze 📈 improve 🔁 repeat Accelerate your AI product flywheel #moptheslop
-
Scott Clark liked thisScott Clark liked thisWeek to remember! It’s incredible to have live data flowing through HomeCEO.AI ;) still waitlist https://lnkd.in/eW9zircW With live data, 3 AI systems are kicking in: operations, features, and data. - Operations AI takes the stream of logs, looks for bugs and automatically creates fixes. We’re down from 4 bugs per user visit to 1 - in days!! This caught everything from permissions, safe guards on redirects, OOM for run functions, you name it. I’m going to have to address the probabilistic nature of AI interfaces and chat tokens to get that 1 down to near 0. I do see where humans still provide architecture wisdom in products created by AI. - For user suggested features, readability has dramatically improved across devices. Moreover user feedback is setting the priorities for the next set of feature entrancements. It dramatically needs to level up proactive communications with users. - For data, now that we have specific areas of interest the AI data creation pipelines have been churning away to 50x the data we had near people. I would never have guessed the Massachusetts area. An amazing shout out to the Gen1 folks taking it for a spin and helping me out! You are rock stars! I saw where the product delighted some and let others down…I know the next step and will keep up leveling until it delights all of you! Happy Friday 🍾
-
Scott Clark liked thisScott Clark liked thisThis week has been an incredibly difficult one for the entire Elephants community—our employees, our customers, and our neighbors. We have been deeply sustained by the outpouring of support we’ve received since Tuesday morning’s fire. Your messages remind us how meaningful Elephants has been to so many Portlanders as a restaurant, an employer, a nonprofit partner, and a place where people come together. We are grateful to the brave and capable team at Portland Fire & Rescue and Police for their swift response this week—both in containing Tuesday morning’s fire and apprehending the individual responsible on Wednesday afternoon. We hope that the individual involved receives the mental health services they need, both for their own wellbeing, and for the safety of our broader community. At the same time, accountability matters. Incidents like this require us not only to address the immediate situation, but also to examine the broader context in which small businesses operate. For more than four decades, Elephants Delicatessen has fed, nourished, and hosted the Portland community. We employ hundreds of Portlanders and strive every day to operate as a responsible and civically minded operation. We are honored to be called a beloved institution and to serve as a gathering place—a “third place”—for so many people in our city. We also have deep sympathy for neighbors experiencing hardship. But compassion alone cannot come at the expense of safety and stability. The continued presence of businesses like ours—and the livelihoods they support—is threatened when difficult decisions are delayed, accountability becomes uncertain, and the safety of employees and guests is compromised. Small businesses cannot succeed in an environment where insecurity, uncertainty, and anxiety become part of everyday life. The tools currently offered to small businesses are woefully insufficient. As we work to respond thoughtfully and constructively, we urge our civic leaders to use this moment as a catalyst for meaningful change. Let’s not abandon compassionate care, but also weigh the needs of businesses and citizens more strongly. To our customers, neighbors, and local vendors, thank you for your decades of support at our NW 22nd Avenue location. We are assessing the damage and are going to do everything we can to rebuild. As we both mourn and celebrate our flagship store, we invite you to share memories, photos, and stories from Elephants on NW 22nd by emailing Info@ElephantsDeli.com. And the most meaningful way to support our team right now is simple: continue to show up. Visit our six other wonderful locations, order catering, shop our markets, and gather around the tables that make our community special. Your support ensures that Elephants—and businesses like ours—can continue to serve Portland for many years to come.
Experience & Education
-
Distributional
********** *** ***
-
***** ***********
** *** ** ** ** *** *** ************** *********** ***** ***********
-
******
********** *** ***
-
******* **********
*** ******* **** *** *** ************ ******* **** ** ******** ******* undefined
-
-
****** ***** **********
** **** ** ************ ** ******** ** ************* *******
-
View Scott’s full experience
See their title, tenure and more.
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Volunteer Experience
-
Member Board of Trustees
OMSI
- 3 years
Served as Treasurer, member of Executive Committee, and head of Finance Committee
-
-
Member of Industry and Innovation Council
Oregon State University College of Science
- 5 years 8 months
Publications
-
Parallel Bayesian global optimization of expensive functions
Operations Research
We consider parallel global optimization of derivative-free expensive-to-evaluate functions, and propose an efficient method based on stochastic approximation for implementing a conceptual Bayesian optimization algorithm proposed by Ginsbourger in 2008.
Other authorsSee publication -
Bayesian optimization for machine learning: A practical guidebook
arXiv preprint arXiv:1612.04858
We outline four example machine learning problems that can be solved using open source machine learning libraries, and highlight the benefits of using Bayesian optimization in the context of these common machine learning applications.
Other authorsSee publication -
ALE: a generic assembly likelihood evaluation framework for assessing the accuracy of genome and metagenome assemblies.
Bioinformatics
Researchers need general purpose methods for objectively evaluating the accuracy of single and metagenome assemblies and for automatically detecting any errors they may contain. Current methods do not fully meet this need because they require a reference, only consider one of the many aspects of assembly quality or lack statistical justification, and none are designed to evaluate metagenome assemblies.
In this article, we present an Assembly Likelihood Evaluation (ALE) framework that…Researchers need general purpose methods for objectively evaluating the accuracy of single and metagenome assemblies and for automatically detecting any errors they may contain. Current methods do not fully meet this need because they require a reference, only consider one of the many aspects of assembly quality or lack statistical justification, and none are designed to evaluate metagenome assemblies.
In this article, we present an Assembly Likelihood Evaluation (ALE) framework that overcomes these limitations, systematically evaluating the accuracy of an assembly in a reference-independent manner using rigorous statistical methods.Other authorsSee publication -
Solving Genomic Jigsaws
DEIXIS Magazine
See publicationA popular science style essay I wrote about my PhD research for a computational science magazine.
-
Left handed beta helix models for mammalian prion fibrils.
Prion
See publicationWe propose models for in vitro grown mammalian prion protein fibrils based upon left handed beta helices formed both from the N-terminal and C-terminal regions of the proteinase resistant infectious prion core. The C-terminal threading onto a beta-helical structure is almost uniquely determined by fixing the cysteine disulfide bond on a helix corner. In comparison to known left handed helical peptides, the resulting model structures have similar stability attributes including relatively low…
We propose models for in vitro grown mammalian prion protein fibrils based upon left handed beta helices formed both from the N-terminal and C-terminal regions of the proteinase resistant infectious prion core. The C-terminal threading onto a beta-helical structure is almost uniquely determined by fixing the cysteine disulfide bond on a helix corner. In comparison to known left handed helical peptides, the resulting model structures have similar stability attributes including relatively low root mean square deviations in all atom molecular dynamics, substantial side-chain-to-side-chain hydrogen bonding, good volume packing fraction, and low hydrophilic/hydrophobic frustration. For the N-terminus, we propose a new threading of slightly more than two turns, which improves upon the above characteristics relative to existing three turn beta-helical models. The N-terminal and C-terminal beta helices can be assembled into eight candidate models for the fibril repeat units, held together by large hinge (order 30 residues) domain swapping, with three amenable to fibril promoting domain swapping via a small (five residue) hinge on the N-terminal side. Small concentrations of the metastable C-terminal beta helix in vivo might play a significant role in templating the infectious conformation and in enhancing conversion kinetics for inherited forms of the disease and explain resistance (for canines) involving hypothesized coupling to the methionine 129 sulfur known to play a role in human disease.
Honors & Awards
-
2016 Forbes 30 under 30
Forbes
http://www.forbes.com/30-under-30-2016/enterprise-tech/
-
Department of Energy Computational Science Graduate Fellowship
-
Four year full fellowship for PhD work. One of ~20 awarded each year nationally.
View Scott’s full profile
-
See who you know in common
-
Get introduced
-
Contact Scott directly
Other similar profiles
-
Hao Li
Hao Li
MBZUAI (Mohamed bin Zayed University of Artificial Intelligence)
17K followersLos Angeles, CA
Explore more posts
-
Niall Murphy
5K followers
YellowDog.ai just set a 10x benchmark uplift in scale computing, delivering 40,000 tasks per second (TPS) and managing 100,000 compute nodes in the cloud. What's even more interesting, that's 2x IBM Symphony and opens an intriguing pathway for these until-know closed/captive systems.
10
-
Michael Ryaboy
inference.net • 5K followers
Thanks to the resurgence of RL, LLMs are finally able to reliably coordinate tools and reasoning to do high-precision retrieval. Companies like Happenstance, Clado, and Mintlify have already shifted to agentic search, and it's only a matter of time until anything less feels broken to users. Link to full blog in comments.
20
2 Comments -
Joe Kristiansen
Tracware • 568 followers
Over the last 48 hours, 3,000+ people saw the CMEF demo. People tried it. Most asked the same question: "What do I actually DO with this information?" Fair question. So I wrote it down: 9 pages (spaced) on what CMEF is, how it works, and what happens after it catches something. Inside: What a "drift corridor" is (and why the Ukraine question is one) The anatomy of a 3-pass run (with actual entropy scores and flags) 5 concrete governance options when CMEF flags an answer Why "Eastern sources only" makes modern models hallucinate citations This isn't marketing. It's the manual. If you're deploying AI in compliance, healthcare, finance, or anywhere trust matters—this is how you measure what's actually coming out of your models. Before anyone sees it. Authors note - This is a repeat of an earlier post. Why you ask? Because my sense of humor doesn't always track. I left an AI artifact in the original report, because it was hilarious and a way to thumb my nose at the fear porn. However, it has been rightly pointed out that I am not in every room that report may be. So here is the Drift Corridor, minus my own personal humor.
2
1 Comment -
Chris Talley
Interconnection.fyi • 2K followers
The most recent Queued Up report from Berkeley Lab is out. Steven Zhang and I are proud to be listed as coauthors on this edition. The Interconnection.fyi team worked closely with Joseph Rand and the rest of the LBNL team to provide the data that powers this year's analysis. Supporting this level of research is core to our mission of bringing transparency to the wholesale energy markets. The 2025 edition (covering data through the end of 2024) highlights some significant shifts in the landscape: Active Capacity: 2024 closed with nearly 2,300 GW of generation and storage seeking interconnection. Changing Mix: Active natural gas capacity increased by 72% year over year, while solar and storage saw slight decreases in total queue volume. The Backlog: 408 GW of capacity already has an executed or draft interconnection agreement but has not yet reached commercial operations. Timelines: For projects built between 2018 and 2024, the median duration from request to operation has doubled compared to the early 2000s. This report is the definitive annual benchmark for the industry and provides a vital baseline as we begin to see the implementation of FERC Order 2023. The analysis in this report is based on our EOY 2024 data snapshot. Since then, we have continued to track and update these queues every single day. If you want a live view of how these numbers have shifted in the months since this snapshot was taken, follow Interconnection.fyi and subscribe to our Substack. You can find the full slide deck, interactive maps, and raw data files at the link in the comments.
41
3 Comments -
Shawaiz Ahmed
InstrumentiX • 2K followers
this is a huge signal for applied neuroscience. sources report sam altman has hired an award-winning biomolecular engineer, mikhail shapiro, to join his "merge lab". shapiro is known for his pioneering work at caltech. this isnt just another neuralink rival. it looks like a completely different approach. shapiro's research focuses on non-invasive techniques using tools like ultrasound and gene therapy to interact with cells. a stark contrast to neuralinks surgical implants. altman himself has expressed interest in less invasive methods. this is the kind of parallel path in applied neuroscience that is vital. the field is ready for major breakthroughs, and we need multiple innovative strategies to deliver new treatment options and understand the brain.
16
2 Comments -
Srihari Sriraman
Nilenso Software • 1K followers
Chris Lattner's take on CCC (Claude C Compiler) is a great summary of where AI's capabilities are at, in terms of generating code: https://lnkd.in/gGYapXVa Would suggest reading this along with Drew Breunig's take on Claude being an electron app: https://lnkd.in/g7xUsbss
11
-
Jacob Effron
Redpoint Ventures • 16K followers
At OpenAI Jerry Tworek helped drive o1, o3, and Codex as VP of Research. Then he left to pursue “types of research that are hard to do at OpenAI.” This week on Unsupervised Learning, I sat down with Jerry to discuss where AI research is headed and what he learned from seven years at the forefront of the field. We hit on: ▪️ Why Jerry updated his AGI timeline after building reasoning models and what fundamental capability current models are missing ▪️ The real limits of scaling reinforcement learning ▪️ Why he left OpenAI after creating some of its biggest breakthroughs ▪️ Why Anthropic did so well in coding ▪️ Inside OpenAI's pivotal 51-49 decisions ▪️ What makes great AI researchers Why current models can never be AGI without continual learning Jerry's biggest update is his belief that static models can never achieve AGI. When models hit walls, they become "hopeless" and stay stuck. True intelligence always finds a way by probing problems until solved, but current models lack the ability to update their beliefs based on failure. As Jerry puts it: "Intelligence always finds a way... which the current models do not really." The fundamental limits of scaling RL While scaling RL delivers predictable benefits, you fundamentally get what you train for. Models excel at specific RL skills but struggle with generalization beyond their training distribution. Scaling primarily means adding more data, and this loop is slow and labor-intensive. The question isn't whether RL works, but whether there's research that could give us better generalization with less data. Why Jerry left OpenAI to chase the next paradigm shift After helping introduce reasoning models, which Jerry describes as a "tectonic shift," he wants to chase that high again. He's now pursuing what he believes is missing in how the world trains models, seeking freedom to explore the most important unsolved problems differently than everyone else has tried. Listen here👇 Spotify: https://bit.ly/4brrgTE Apple: https://bit.ly/4a2zjnc YouTube: https://lnkd.in/ekyv-znp
46
3 Comments -
Jeff Jonas
Senzing, Inc. • 15K followers
AI is advancing at breakneck speed. What I imagined might be years away -- compute finding non-obvious insights across science papers yielding novel discovery -- has arrived: “What distinguishes Zochi is its ability to identify non-obvious connections across papers and propose innovative solutions that address fundamental limitations rather than incremental improvements." https://lnkd.in/gqtz2aSj
77
19 Comments -
Ayush Gupta
Genloop • 6K followers
#TuesdayPaperThoughts Edition 60: The Art of Scaling RL Compute for LLMs This week's #TuesdayPaperThoughts explores "The Art of Scaling Reinforcement Learning Compute for LLMs" from researchers at Meta, The University of Texas at Austin, University of California, Berkeley, and Harvard University. While pre-training has had its moments with predictable scaling laws, RL training has remained more art than science until now. Key Takeaways: 1️⃣ Predictive Scaling Framework: The work introduces sigmoidal compute-performance curves for RL training that separate asymptotic performance (A) from compute efficiency (B). This framework enables extrapolation from smaller-scale runs to predict performance at larger compute budgets. The team validated this with a massive 100,000 GPU-hour run where predictions from just the first 50k hours closely matched final performance. 2️⃣ Not All Recipes Scale Equally: Methods that look promising at small compute budgets can hit lower performance ceilings at scale. The study reveals that design choices like loss type and FP32 precision shift asymptotic performance, while factors such as loss aggregation and normalization primarily modulate compute efficiency. This explains why some widely-adopted methods plateau unexpectedly. 3️⃣ ScaleRL Recipe: Through systematic ablations consuming 400,000+ GPU-hours, the team developed ScaleRL—combining PipelineRL with 8-step off-policyness, truncated importance sampling (CISPO), FP32 logits computation, and adaptive prompt filtering. ScaleRL achieves A=0.61 asymptotic performance, outperforming DeepSeek's GRPO, Qwen's DAPO, and other prevalent methods on both ceiling and efficiency. The timing couldn't be better. With RL compute budgets exploding (10× increase between model generations for o1→o3 and Grok-3→Grok-4), the field desperately needed this systematic approach. Research Credits: Devvrit K., Lovish Madaan, Rishabh Tiwari, Rachit Bansal, Sai Surya Duvvuri, Manzil Zaheer, Inderjit Dhillon, David Brandfonbrener, Rishabh Agrawal Paper Link: In comments
21
1 Comment -
Alan Lockett
Self-employed • 1K followers
Just read this paper out of UC Berkeley (Vafaii, Galor, & Yates). Fascinating study. In just the first few paragraphs they summarize a range of references that unify variational autoencoders, diffusion models, and predictive coding (the neuroscience version) within the Bayesian learning framework. They go on to derive a recurrent spiking neuron just from the two principles of free energy minimization and a poisson firing rate, demonstrating experimentally that it has many neurally plausible properties. https://lnkd.in/g2XJSrxa
13
-
Ritvik Pandey
Pulse • 15K followers
Great to see how a computational physics research lab is using Pulse's API to extract LaTeX equations and mathematical notation from their 40+ year archive of papers. Their team built a specialized pipeline that preserves the semantic structure between equations, preserving relationships between variables, theorems, and proofs. What's remarkable is how Pulse correctly identifies and extracts nested mathematical structures (matrices within summations within integrals) while maintaining proper symbol relationships. Our mission is to turn previously inaccessible computational knowledge into structured, searchable data that accelerates novel research. If you're a research team interested in trialing out Pulse, send me a DM for some trial credits! Sample LaTeX output attached below.
64
1 Comment -
Krishna Caldas
Mutiro Labs • 1K followers
I was listening to this https://lnkd.in/giycVEkG and had a thought I can't shake: What if foundation models are AI's genetic code? DNA gives us instincts... babies cry, calves walk to milk without learning. Foundation models feel similar: pre-loaded with language and reasoning. But evolution doesn't create one perfect organism. It creates fish, mammals, birds... each specialized. Maybe AI works the same way? Not one AGI, but multiple species and individuals that learn and evolve. If foundation models are our "genetic code"... we need to be careful what we embed there. These become instincts for all future AI. Like Sutton says, learning is fundamental and the next crucial piece: how do we take what AI systems learn and feed it back to the next generation? In biology, successful traits get passed down... what's the AI equivalent? Maybe we're not just building better AI... maybe we're designing artificial evolution.
11
-
Simon Hoskins
Hostech • 306 followers
I’ve tried and used many tools surrounding LLMs, such as LangChain/Graph, LlamaIndex, vector stores like Qdrant, and full AI platforms like Azure AI Foundry. I don’t understand why none of them offer RAG applied to their own documentation - it seems like such an easy win. Many of these technologies are new and evolving quickly, so getting grounded information from places like ChatGPT can be difficult. Trawling through documentation feels unnecessary in the age of LLMs. Maybe someone can enlighten me.
18
9 Comments -
M. Alejandra Parra-Orlandoni (mapo)
Pasteur Labs & ISI • 5K followers
Ever wondered how JAX makes machine learning so fast? Just dropped Part II of our series, "From JVP to VJP," to explain. At Pasteur Labs, we use JAX to build complex simulation systems — and automatic differentiation makes it all possible. This post breaks down JAX's clever approach to computing gradients efficiently. If you've used jax.grad() and wondered what's happening behind the scenes, this demystifies the magic. ✨ Part of our ongoing series on building next-gen simulation systems. Thanks Niklas Heim for the excellent post! 👀 Read it and our other Insights posts: https://lnkd.in/eHi25bBE
25
1 Comment -
Manav Thiara
Sage • 3K followers
I’ve been experimenting with structured knowledge graphs, optimizing semantic search, and distilling reasoning frameworks that sit on top of LLM outputs. Despite all the progress, no one has yet solved how large language models can offer meaningful guidance about the future. The real breakthrough may come not from bigger models, but from architectures that build persistent agentic memory, systems that learn, recall, and reason more like the human brain. DM me if you see something worth a deeper conversation.
51
-
Brittany Walker
CRV • 5K followers
Voice is a super interesting modality right now - maybe the first modality we're seeing move to open source models across a number of scale ups / enterprises. Reliability concerns, high costs, and open source model performance are pushing engineers to do their own fine tuning vs. relying on third-party vendors of proprietary models. Many of these orgs have already been collecting their own first-party data and now with third-party vendors like Extrian, David AI, etc they can train really high quality models. RL has been insanely hyped, but it's been unclear how long it will take scale ups and enterprises to actually lean in. Voice AI might be hitting that inflection point faster than expected.
39
3 Comments -
Alex Leung
Powder • 1K followers
Go to sleep. Wake up to results. The human writes Markdown. The AI runs 100 experiments overnight. On one GPU. Karpathy's autoresearch is what ML research looks like now. You don't tune hyperparameters anymore. You design the arena and let the agent iterate inside it. Same pattern we landed on building our own refinement agent. Strict targets to optimize towards. Tight constraints, fast feedback, let it loop. Our accuracy for certain document formats went from 22% to 100% overnight. Engineer the environment. Constrain the scope. Let the agent run.
3
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content