Sign in to view Brian’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Brian’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Boston, Massachusetts, United States
Sign in to view Brian’s full profile
Brian can introduce you to 10+ people at Red Hat
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
13K followers
500+ connections
Sign in to view Brian’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Brian
Brian can introduce you to 10+ people at Red Hat
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Brian
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Brian’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Activity
13K followers
-
Brian Stevens reposted thisIt was a great pleasure to host CNBC at our Red Hat AI booth during the India AI Impact Summit this week in Delhi, and to share what we have been building together with customers, partners and the open source community. The momentum has been building as we continue to deliver a high performance AI platform that promises to deliver any model, using any accelerator, and running on any cloud. Daniel Aw Fytos Charalambides Navtez Bal Paul Whittard Garry Gray Philip Yeap Abhishek Shukla Mukesh Mehta Ausim Khan Mangesh Surve Misha Joshi Arpita Sengupta Ravi Goyal Joel Jackson Joe Fernandes Brian Stevens Vincent Caldeira James Lovegrove Tushar Katarki Jeff DeMoss Erwan Granger Karl Eklund Jeff Winn Vijay Chebolu Prasad Mukhedkar Tarun Ghai Kiran Challapalli abhishek vijra Elisa NavarroBrian Stevens reposted thisTHREE LEADERS. ONE BOOTH. THE REAL TALK ON AI IN INDIA. From the Red Hat booth at Bharat Mandapam during the India AI Impact Summit 2026, CNBC-TV18's Global AI Lens brings you insights from the team powering enterprise AI across India and Asia-Pacific. Steve Shirkey, Director, APAC AI Platform Misha Joshi, Senior Director, Head of Services (India/South Asia) and Ausim Khan, Director, Partner Ecosystem (India & South Asia) share Red Hat's vision for AI in India. Platform. Services. Ecosystem. The full stack of AI transformation—decoded. Global AI Lens | India AI Impact Summit 2026 HCLTech Red Hat d-Matrix Sid Sheth Dr. Ravi Gupta Pradip Thaker #GlobalAILens #IndiaAI #AIImpactSummit #CNBCTV18 #RedHat #BharatMandapam #EnterpriseAI
-
Brian Stevens reposted thisBrian Stevens reposted thisI’m so happy to share with you that this weekend (November 1st) there will be a great vLLM meetup in Beijing! We (Red Hat) together with our partners and friends in the community, such as the Ant Group, AMD, ByteDance and MetaX, co-host this event and we invite the most seasoned people in the community such as Michael Goin - the lead vLLM maintainer to give us a general update and llm-d’s latest features, and many others to share insightful topics - quite a few around distributed inference solutions and different accelerator support ways. So if you happen to be in Beijing this weekend, welcome to join in person! You can either drop me a message or register through “vLLM community” WeChat account! The meetup also has live-streaming, we will provide the live-streaming link shortly so please stay tuned;) Brian Stevens Brent Holden Vincent Caldeira Michael Goin Saša Zelenović Grant Shipley Tushar Katarki Christopher Nuland Jeff Winn Steve Shirkey Li Ming Tsai Andreas Spanner Chris Butler Vinod Pathangay Ken Komazawa Lisa Li
-
Brian Stevens shared thisThe pace of advancements in AI is stunning, and innovations in open source are driving the industry’s emerging capabilities in inference platforms, agentic architectures, and data integration. While that alone is a reason to celebrate, it's hard to keep up and a day doesn’t go by where I’m not investing time to learn something new. At Red Hat we decided to turn that inside out, and share a day-of-learning with like-minded colleagues. We’ve curated a virtual event of 12 session across 4 tracks: Inference and Optimization – Optimal deployment with vLLM, LLM Compressor, and Speculators. Model Customization – Connecting models to enterprise data Agentic AI – Building open and flexible agents with MCP, LlamaStack, and more. Scaling Across Hybrid Cloud – Learn distributed inference and scaling strategies with OpenShift AI. Register here: https://lnkd.in/e_UZ5qAG, and happy learning! Red Hat AI Day of Learning: Your Path to Enterprise-Ready AI October 16, 2025 | 10:00 a.m. - 11:30 a.m ESTRed Hat AI Day of Learning: Your Path to Enterprise-Ready AIRed Hat AI Day of Learning: Your Path to Enterprise-Ready AI
-
Brian Stevens reposted thisBrian Stevens reposted thisRed Hat AI Day of Learning: Your Path to Enterprise-Ready AI October 16, 2025 | Starts at 10:00 a.m. EST | Virtual Choose your own AI learning journey with 12 breakout sessions across 4 tracks: 1️⃣ Increase Efficiency with Fast, Flexible, and Efficient Inferencing 2️⃣ Simplified and Consistent Experience for Connecting Models to Data 3️⃣ Accelerate Agentic AI Delivery and Stay at the Forefront of Innovation 4️⃣ Flexibility and Consistency When Scaling AI Across the Hybrid Cloud Developers, engineers, and technical practitioners can mix and match sessions to focus on the topics that matter most and leave with practical skills to apply right away. Register here: https://lnkd.in/eCSDvYVyRed Hat AI Day of Learning: Your Path to Enterprise-Ready AIRed Hat AI Day of Learning: Your Path to Enterprise-Ready AI
-
Brian Stevens shared thisEldar Kurtić is amazing. Proud to be on his team.Brian Stevens shared thisWhen a brilliant mind from Bosnia and Herzegovina leads innovation at a global powerhouse like Red Hat, it’s impossible not to take notice. We’re proud to introduce another strong addition to our Engineering Stage this September — Eldar Kurtić, whose work pushes the boundaries of AI efficiency and deployment. ↳ Eldar Kurtić is a Senior Researcher at Red Hat and Institute of Science and Technology Austria, specializing in efficient inference techniques for large language models (LLMs), with a particular focus on sparsity and quantization. His work centers on developing methods to accelerate inference within the vLLM engine, bridging cutting-edge research with practical deployment solutions. At Kiss the Future AI Summit, Eldar will lead a hands-on workshop titled: “Beginner-Friendly Introduction to LLM Quantization: From Zero to Hero” He will cover: ↳ What quantization is — and why it matters ↳ How quantization fits into the architecture of LLMs ↳ Today’s leading quantization techniques for deployment ↳ How to quantize your own models ↳ Accuracy trade-offs and tuning for optimal performance ↳ Real-world inference cost and performance implications Whether you’re an ML engineer, AI researcher, or just getting started, this is your gateway to mastering LLM quantization. 🎟️ Did you know you can get a ticket just for the Engineering Stage? Head to Entrio now and secure your spot! The End of Hype. The Start of Impact. | 25th and 26th September #KissTheFutureAISummit #KisstheFuture #AISummitSarajevo2025 #EldarKurtic #RedHat
-
Brian Stevens reposted thisMore than happy to be part of this story! Thank you Blum Institut for the invitation.Brian Stevens reposted thisWhen a brilliant mind from Bosnia and Herzegovina leads innovation at a global powerhouse like Red Hat, it’s impossible not to take notice. We’re proud to introduce another strong addition to our Engineering Stage this September — Eldar Kurtić, whose work pushes the boundaries of AI efficiency and deployment. ↳ Eldar Kurtić is a Senior Researcher at Red Hat and Institute of Science and Technology Austria, specializing in efficient inference techniques for large language models (LLMs), with a particular focus on sparsity and quantization. His work centers on developing methods to accelerate inference within the vLLM engine, bridging cutting-edge research with practical deployment solutions. At Kiss the Future AI Summit, Eldar will lead a hands-on workshop titled: “Beginner-Friendly Introduction to LLM Quantization: From Zero to Hero” He will cover: ↳ What quantization is — and why it matters ↳ How quantization fits into the architecture of LLMs ↳ Today’s leading quantization techniques for deployment ↳ How to quantize your own models ↳ Accuracy trade-offs and tuning for optimal performance ↳ Real-world inference cost and performance implications Whether you’re an ML engineer, AI researcher, or just getting started, this is your gateway to mastering LLM quantization. 🎟️ Did you know you can get a ticket just for the Engineering Stage? Head to Entrio now and secure your spot! The End of Hype. The Start of Impact. | 25th and 26th September #KissTheFutureAISummit #KisstheFuture #AISummitSarajevo2025 #EldarKurtic #RedHat
-
-
Brian Stevens shared thisPersonally or professionally, every conversation I have w/ Chris Wright is always fun. This one was no different where we dove into vLLM and all things AI+Inference. Cheers.Brian Stevens shared thisHow do we take AI from research labs to robust, scalable enterprise production? Red Hat CTO Chris Wright and Red Hat AI CTO Brian Stevens dive deep into production-quality inference, the role of open source projects like vLLM, and the journey to practical enterprise AI. They discuss parallels with Linux's early days and the community effort needed to build the future AI stack. A must-listen for tech leaders navigating AI: https://red.ht/43ZcQWo.
-
Brian Stevens reposted thisBrian Stevens reposted thisOne of the voices in AI that I follow for his insightful commentary is Mitko Vasilev. He was commenting on the demand for sovereign AI infrastructure and the need for private, secure, manageable, cost-efficient, and scalable AI. The 2025 reality check is that you can get vLLM integrated in Kubernetes from the top contributors to #linux, #kubernetes, and #vllm. If you have real enterprise-grade needs for control, privacy, an integrated stack, security response, and support without praying to Linus, Red Hat offers a full-stack AI platform where you can use your choice of hardware, cloud, and models partnered with industry leaders. Companies like Mastercard need sovereign AI platforms all over the world, and are setting a great example of what great #EnterpriseAI looks like in their deep dive on thenewstack https://lnkd.in/eKXdmXy8 Learn more about #RedHatAI from the #RedHatSummit keynote with Brian Stevens: https://lnkd.in/efD5Q9tY Building blocks of AI: Meta #llama #llamastack Qwen #qwen DeepSeek AI #deepseek Google #gemma Mistral AI #mistral Ai2 #molmo Microsoft #phi NVIDIA #nemoton #h200 #gh200 #nvaie #dgx #gracehopper #nim IBM #granite AMD #instinct Google #gcp #tpu Amazon Web Services (AWS) Amazon #neuron Intel Corporation #gaudi #openvino Hugging Face Dell Technologies Hewlett Packard Enterprise Supermicro Lenovo #llm #llmcompression #quantization #kubeflow #kserve #mcp #a2a #multinode #multigpu #sdg #syntheticdatageneration #finetuning #RAG #InstructLAB #featurestore #feast #mlops #llmops #datascience #jupyter #ResponsibleAI #AgenticAI #EnterpriseAI #SovereignAI #RedHatOpenShiftAI #RedHatAIInferenceServer #RHELAIKubernetes Powers Mastercard's AI-Workbench for Secure InnovationKubernetes Powers Mastercard's AI-Workbench for Secure Innovation
-
Brian Stevens liked thisBrian Stevens liked thisThe next chapter for AI infrastructure is here: llm-d is entering the Cloud Native Computing Foundation (CNCF) Sandbox. As production inference becomes a standard cloud-native capability, CoreWeave is proud to continue its role as a founding contributor alongside Red Hat, IBM, Google and NVIDIA. llm-d matters because production inference places new demands on infrastructure: it needs more intelligent orchestration, economics aligned to application value, and the flexibility to run across public cloud, private data centers, and the edge. By entering the CNCF Sandbox, llm-d is moving onto neutral ground, helping make production inference more accessible, portable, and efficient for the broader industry. Read the full deep-dive from Urvashi Chowdhary, VP of Product Management, AI Services at CoreWeave: https://hubs.la/Q0488r_n0Why llm-d in CNCF Matters for Production Inference | CoreWeave BlogWhy llm-d in CNCF Matters for Production Inference | CoreWeave Blog
-
Brian Stevens liked thisBrian Stevens liked thisKubeCon + Cloud Native Con EU 2026 - what an event! This year goes down in history as the biggest #KubeCon of all time with 13,500 people coming together in Amsterdam to keep cloud native moving 🚲. And the news was just as big - Red Hat contributed #llmd live on stage (will never not get a thrill when that happens), #kyverno officially graduated 🎓 and NVIDIA became a Platinum member of the CNCF, as cloud native technologies continue to underpin every major AI advancement. Congratulations to the amazing Cloud Native Computing Foundation (CNCF) teams on an incredible event and special thanks to the marketing and PR rockstars Jennifer R. Eliza Power Audra Montenegro Helena Spease Diogenese Topper Natasha Woods Kristi Piechnik Kaitlin Thornhill Haley White Grace Lucier who hustle non stop during these shows. From karaoke to #klausGPT this community knows how to innovate - and have fun doing it. Until next time!
-
Brian Stevens liked thisBrian Stevens liked thisBig milestone for llm-d joining CNCF as a sandbox project today and a big step forward for open, efficient, performant AI inference at scale. Congratulations to the team and the broader community and let's keep going! https://lnkd.in/ea3KjkUM https://lnkd.in/es8tC_i8 https://lnkd.in/eBh7t7xW Carlos Costa Vita Bortnikov Brian Stevens Robert Shaw Stephen Watt
-
Brian Stevens liked thisBrian Stevens liked this**The inference conversation happens in Boston — and you're invited.** If you've ever muttered "why is my P99 latency so brutal" under your breath while staring at GPU metrics... this one's for you. The Open Accelerator, Red Hat AI, IBM and NVIDIA are teaming up to bring the **vLLM Inference Meetup** to Cambridge on March 31st. We're talking deep technical sessions, live demos, and the kind of hallway conversations that actually move the needle. **What's on deck:** 🔧 A hands-on **pre-event workshop** (4:00 PM) where you'll deploy Llama 3.1 8B with vLLM, benchmark it, scale it, and then watch llm-d's cache-aware routing improve that tail latency for you 🎤 Talks from **vLLM maintainers and committers** on everything from model compression and speculative decoding to agentic AI and distributed inference at scale with Kubernetes 🤝 A sneak peek at what we're building at **The Open Accelerator** — because the best part of open source and communities is what happens when they come together to help young founders and the ecosystem 🍕 Networking, food, drinks, and the a chance to dive deep into conversations about inference optimization strategies with people who are building and implementing them Whether you're a vLLM contributor, an ML engineer running inference in production, a platform team wrangling GenAI workloads, or just deeply curious about efficient serving — pull up a chair. 📍 Cambridge, MA 📝 Registration required — spots are approval-based and close 24 hours before the event Grab your spot here: https://luma.com/4rmkrrb7 Looking forward to seeing you! Bring your questions, your laptop, and your opinions about KV cache management. 🚀 🦞While everyone is "clawing" onto AI agents which are grabbing stuffed animals and honestly, respect. But while OpenClaw is perfecting the art of snagging a plushie, the vLLM and llm-d communities are out here solving the production inference challenges that actually keep ML engineers up at night. Come hang with the folks doing that work!!🤝 #vLLM #LLMInference #OpenSource #RedHatAI #TheOpenAccelerator #GenAI The Open Accelerator Stefanie Chiras, Ph.D. Saša Zelenović
-
Brian Stevens liked thisBrian Stevens liked this#FreedomOfChoice - That is what I advocated at Red Hat for 10+ glorious years to customers and am excited about exercising a similar freedom starting today for myself -- the choice to do what I want to do going to forward. Announcing my retirement from what I #have to do and looking forward with excitement to what I #want to do which includes any one of the following in no particular sequence: - Active membership and leadership roles in Boards of Directors - Taking calculated risks in exploring new ventures in emerging technologies - Realizing business outcomes with the strategic assembly of technologies - Advancing Quantum Computing for a safer digital world - On the Path to get certified as a Director NACD (National Association of Corporate Directors) - Documenting our familial roots for future generations - Promoting classical music in emerging generations of children - Taking care of our senior parents with gratitude ❤️ ❤️ And the best part about this is that I can do what I want to do, when I want to do it and enjoy doing it. Hey #Life! --- Hi there!! --- Here I come!!
-
Brian Stevens liked thisBrian Stevens liked this🚀 Calling all curious, smart, and self-driven students! 🚀 Are you ready to dive into the fast-paced world of AI and innovation right here in the heart of Boston’s Seaport/Fortpoint district? The Open Accelerator is looking for highly collaborative and fast-learning minds to join as a Program Manager Intern! This isn't your average internship. You'll be right at the forefront of the Massachusetts AI and tech ecosystem, helping us bridge the deployment gap for technical entrepreneurs and gearing up for our highly anticipated Residency launch. It's an exciting, rapidly changing environment where your ideas will directly shape the future of emerging AI startups. If you have a passion for building new things, analyzing data, and thriving in a space where you can truly be yourself, we want you on the team. You'll get to network directly with the local AI startup community, conduct first-party research, and get hands-on experience that actually matters, your work will have a direct and profound impact to our success. Plus, being based in the Seaport means you're never too far from some seriously good culinary spots when it's time to step out of the office and celebrate a project milestone.🌮☕️🍝 Ready to make an impact and accelerate your career? Check out the details and apply below! 👇 🔗 Apply here: https://lnkd.in/enRAy5NF The Open Accelerator #Internship #BostonTech #AIStartups #TheOpenAccelerator #Hiring #StudentOpportunities #TechCareers #MAAIhub
-
Brian Stevens liked thisBrian Stevens liked thisThe first vLLM Meetup in Hong Kong has successfully concluded! A heartfelt thank you to our partners, the contributors who traveled from across the globe, and everyone who follows our community's growth. Seeing old friends and meeting new ones in person are the happiest things for us;) Here are the brief summary and some feelings of myself after the meetup: Hong Kong - International Port: Besides Hong Kong local developers, it was inspiring to see developers join us not only from regional hubs like Shenzhen, Hangzhou, Shanghai, and Beijing, but also from Singapore and as far as Silicon Valley. This truly reinforced Hong Kong’s role as a vital international port for innovation and talent. The Multimodal Frontier: The vLLM community is evolving at breakneck speed. Multimodality was the undisputed highlight of this meetup; having core contributors in the room led to high-energy discussions about the future direction of the field. A Full-Stack Ecosystem Perspective: the agenda covered the entire vLLM landscape, not just the software itself. We have accelerator vendors to share their experiences supporting vLLM from a hardware perspective, alongside frontier model providers who shared their engineering innovations for running SOTA models on vLLM - which made the audiences really enjoyable. Deep Discussion on Technical Details: during the social networking time, discussion spanning from PD disaggregation to EPD disaggregation, from RL to Agentic AI, and the recent buzz around OpenClaw. We tend to agree - as OpenClaw drives token consumption to new heights, the vLLM mission like "Easy, fast, and cost-efficient LLM serving for everyone" is more critical than ever. Architectural Excellence: The developers admire Red Hat's contribution to vLLM, especially for the reconstruction of vLLM codebase, which made it easy to support more features and maintain. This is the foundational work for the community and benefits every developer within it. We love vLLM! If you are interested and looking for the next one in Greater China, register here: https://lnkd.in/g-r_RPwE - which is on March 15 in Beijing! Brian Stevens Brent Holden Karena Angell Vincent Caldeira Saša Zelenović Christopher Nuland Steve Shirkey Li Ming Tsai Robert Shaw Michael Goin Luka Govedič Michael Yang Andreas Spanner Chris Butler Vinod Pathangay Ken Komazawa Winifred Wong Albert Law Peter Ho
-
Brian Stevens liked thisBurford Capital has built a market-leading legal finance platform, and I look forward to contributing to its continued growth—particularly as the practice of law evolves and capital plays an increasingly strategic role in complex disputes. Onward!Brian Stevens liked thisWe are delighted to welcome Stephanie Southwick to Burford Capital’s US Commercial team. Stephanie brings deep expertise in legal finance and commercial and IP litigation from prior roles at Law Finance Group, Omni Bridgeway and Greenfield Southwick LLC. Based in the San Francisco Bay area, Stephanie will partner with companies looking to pursue strong legal claims without tying up their own capital. In a 5 Minutes With interview, Stephanie discusses the importance of non-recourse, non-dilutive capital; how legal finance can be used as a strategic liquidity and risk management tool; and what makes a case attractive from a financing perspective. Read the interview here: https://lnkd.in/eMiQkzdq
Experience & Education
-
Red Hat
*** *** ** ***
-
*******
***** ********* ****** ** ***** *********
-
*** ******
***** ********
-
********** *********** *********
** ******** ******* undefined
-
********** ** *** *********
** ******** *******
View Brian’s full experience
See their title, tenure and more.
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View Brian’s full profile
-
See who you know in common
-
Get introduced
-
Contact Brian directly
Other similar profiles
-
Jocelyn Goldfein
Jocelyn Goldfein
Zetta Venture Partners is named after the zettabyte (a trillion gigabytes!) Founded in 2013, we were the first AI-focused fund and we've been backing AI and infrastructure entrepreneurs since long before it was cool.<br><br>We are keenly interested in cloud and data infrastructure (like MotherDuck or Domo), tools and platforms for developers and data scientists building with AI (like Kaggle, Domino, Weaviate or Fixie) and applications powered by ML (like Tractable, Lilt, Skan, and too many more to name). <br><br>We lead or co-lead $1-5M rounds for pre-product-market-fit startups with B2B business models. We believe in verticals like financial services, insure tech, life sciences, health care, sustainability, cloud infra, devops, cybersecurity, manufacturing, supply chain and logistics. <br><br>More about us: https://zettavp.com<br><br>Before venture, I spent my career as an engineer and engineering leader. I led engineering teams in the high growth early years of VMware and Facebook (as well as a few startups including one of my own). I've worked across systems from OS'es to LAMP and mobile apps, to shrink-wrapped native software to developer tools and of course ML. The one "constant" has been high-growth and industry transformation. I'm passionate about scaling products, teams, and companies, and I care deeply about STEM education.
7K followersLos Altos, CA -
Chris Wright
Chris Wright
Experienced technology and strategy leader with a passion for open source software. Collaboration and continuous improvement are best tools for change.
16K followersGreater Boston -
Steven Sinofsky
Steven Sinofsky
Hardcore Software: Inside the Rise and Fall of the PC Revolution
649K followersSioux Falls, SD
Explore more posts
-
Christian Posta
solo.io • 13K followers
Have you heard of the llm-d project? It improves LLM performance and drives down costs of inference serving on Kubernetes. It's based around a few key projects including vLLM and kgateway. If you want a deep dive into how it works with a detailed step-by-step "packet flow" through llmd, take a look at this blog I wrote recently. 👉 https://lnkd.in/g2mbEh3m
44
3 Comments -
Andy Sacks
DeltaStream • 3K followers
Big news for teams building real time systems on Oracle Cloud. DeltaStream is now partnering with the Oracle Cloud team to make streaming analytics on OCI Streaming with Apache Kafka simple and operationally efficient. If you are building event driven systems on Oracle Cloud and want faster time to value from your streaming data, this partnership is worth a look. Read the full partner blog here: https://lnkd.in/gcv3Kyfg Thanks to the DeltaStream team (Hojjat Jafarpour, Rachel Pedreschi, Krishna Raman) and our partners at Oracle (Jason Wilcox, Abhishek Bhaumik, Chris Sullivan, Suresh Patibandla) for helping to kick off this collaboration. #PartnerEcosystem #OracleCloud #OCI #Kafka #RealTimeData #StreamingAnalytics #AIInfrastructure #DeltaStream #DataEngineering #Flink
70
6 Comments -
Saurabh Tiwary
Google • 24K followers
We're excited to announce new capabilities in Google Cloud #VertexAI Training designed to simplify and accelerate large-scale model development. Key highlights from our latest update: 🔹 Flexible, Self-Healing Infrastructure: Leverage fully managed, resilient Slurm environments with automated failure detection and performance-optimized checkpointing. 🔹 Cost-Effective Scheduling: Utilize Dynamic Workload Scheduler (DWS) for fixed future reservations or flexible on-demand capacity. 🔹 Integrated Frameworks: Access optimized recipes for the full model lifecycle (including SFT and DPO) and seamless integration with NVIDIA NeMo. Whether you're fine-tuning standard models or training massive custom foundational models from scratch, these new features are built to get you to production faster. Read the full announcement to see how organizations like Salesforce and AI Singapore are already leveraging these tools to elevate their models. 👇 https://lnkd.in/gVVq3WCw
185
2 Comments -
Susanta Ghosh
JPMorganChase • 2K followers
Today let's talk about Parallel Fan Out/ Concurrent Agentic Design pattern and when parallel agents go rouge What’s the Concurrent Orchestration Pattern? Imagine a scenario where multiple AI agents—each with its own lens or specialty—tackle the same task simultaneously. Instead of a single, step-by-step chain, tasks fan out to different agents in parallel. Then their outputs are merged or aggregated for the final answer. It’s the AI equivalent of a brainstorming session where everyone chips in together. This pattern thrives when you need diverse insights or speed—think ensemble reasoning or reaching a verdict faster. This pattern is commonly used in agentic RAG. Here’s how you can apply it in a Retrieval-Augmented Generation (RAG) system: Step 1: Break the user’s query into smaller sub-queries (e.g., “Define concept X,” “List use cases,” “Give examples”) and map out their dependencies. Step 2: Run those sub-queries in parallel—agents fetch context or process each in isolation (e.g., document retrieval, summarization, external tool usage). Step 3: Once all agents have results, aggregate the findings into a single, coherent response. This technique lets you parallelize independent parts, reducing latency while maintaining clarity—especially effective when sub-queries don’t depend on each other. Super Important : Try to avoid this pattern and use sequential execution even if it's slow, but it might yield a good result, below is the reason As Cognition [the company behind devin] warns in a famous blog post “Don’t Build Multi-Agents”, things can go sideways fast if agents don’t share context or coordination is weak. When agents operate in isolation—making decisions based solely on their own view—the final result might be fragmented, contradictory, or just plain incoherent. Think two agents building different puzzle pieces that don’t fit. The core issues: Context fragmentation: Each agent works in a silo, leading to mismatched assumptions. Implicit decisions: Agents’ outputs reflect unspoken choices that may clash when merged. Coordination complexity: Without strong orchestration, integration becomes error-prone. References : 1. Concurrent Execution Pattern : https://lnkd.in/gZybgm2s 2. LLM Compiler whitepaper : https://lnkd.in/gWwGJhbi 3. Coginition blog Don't build multi agents which shows parallel execution can yield agent drift : https://lnkd.in/gkDKC4TP
14
3 Comments -
Egen
106K followers
Is your AI governance just box-ticking or actually driving outcomes? In this video, Jesse McCrosky, Principal Architect at Egen, explains how effective AI governance means aligning values, assigning accountability, and building a culture that prioritizes transparency and upskilling, not just compliance. Watch the full interview on TFiR: https://lnkd.in/d8hhHztB #AIGovernance #ResponsibleAI #AI #TFiR #EthicalAI #EngineeringNewPossibilities
13
-
Matt Graham 👨💻
RapidDev • 40K followers
Most of you think of OpenAI as a pure software play. But what most of you don't know: they launched a consulting arm this summer with $10M minimum engagements. This embeds engineers directly into client organizations. So even the company that created ChatGPT can't just sell technology. They need humans to make it actually work. Think about that!! Guys, the most successful AI company in the world just admitted that software without implementation is worthless. This creates MASSIVE opportunity for builders like us. While OpenAI burns billions trying to figure out how to deliver value, businesses are desperate for working solutions right now. They don't want access to APIs. They want apps that solve problems. We've been doing this the whole time: taking technology and turning it into real business value. Speed wins. Results win. 🚀
34
18 Comments -
Mike Moss
Redis • 6K followers
Redis to acquire Decodable! "As the world pushes deeper into the age of LLMs, one of the biggest challenges developers will face is ensuring the right context goes to the right place at the right time. We’re helping to solve this by strengthening our capabilities as a real-time context engine that enables instant data access for agents.", Rowan Trollope Check out the announcement: https://lnkd.in/gNkggWWQ
122
-
Justin Borgman
Starburst • 14K followers
Starburst has always believed in providing customers with the freedom of choice. Nowhere is this more true than the emerging developments around interoperable compute among data platforms. Recently, I sat down with Snowflake’s Ryan C. Green to talk about the Open Semantic Interchange (OSI), an open, vendor-neutral standard for sharing semantic models across AI, BI, and analytics tools. OSI gives customers a common, open way to define and share business metrics. This helps them stay consistent across dashboards, notebooks, and machine learning models, no matter which tools they use. For Starburst, this furthers one of our core goals. Choice. Customers should be able to keep one trusted set of business definitions inside their Snowflake environment while still having the freedom to choose the tools that work best for them. OSI helps make that possible. It marks an important step in bringing Starburst’s engine to Snowflake-interoperable compute. The video from my chat with Ryan is coming soon 👀 In the meantime, you can get more details from our recent press release: https://lnkd.in/ev2XGB4u #OpenSemanticInterchange #OSI #Interoperability #AI #Data #Snowflake #Starburst
182
8 Comments
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content