Event Recap: The Future of AI Agents: Security, Speed, and Scaling đ
On Jul 30, we brought together founders, builders, and investors for an unfiltered deep dive into one of the most excitingâand chaoticâfrontiers in tech: AI agents and Infra. From infrastructure bottlenecks to security nightmares, venture trends to go-to-market wins, our panel didnât hold back.
The conversation spanned hard-won scaling lessons, the thorny realities of securing agent-to-agent workflows, and why the winners of the AI agent wave might look nothing like the platforms we use today.
đ¤ Shout out to our partner Revela, Descope and AWS for helping make this event possible! â
âRevela is AI-first dev shop partnering with Seed & Series A startups in SF to build product and infra. Helps accelerate roadmaps, cut inference costs, and scale to the next milestone. Book a free 1:1 with our CEO Mike here to see how we can help your team push more code and scale your AI infra. âFollow our LinkedIn. âLearn More
âDescope is a no/low-code platform for managing identity across customers, partners, and AI agents. Developers use Descope to secure APIs, MCP servers, and agents with authentication, authorization, consent, and tokens. âFollow on LinkedIn or X for ongoing dev updates. âLearn how Descope enables agentic identity. âVisit our AI demo microsite
Featured Speakers
âBrian Quinn, President & GM, North America at AppsFlyer; Managing Director Americas at App Annie; GVP at Kenshoo; Director at Experian, AT&T & Cisco
âRishi Bhargava, Co-founder of Descope ($53M raised), ex-Co-Founder of Palo Alto Networks, ex-GM & VP Software at Intel, ex-VP PM at McAfee
ââSuraj Patel, VP Ventures & Corporate Development at MongoDB; ex-Principal at Bow Capital
âSiddharth Bhai, Product Leader at Databricks, ex-Product Director at Splunk; Lead PM at Google; ex-Principal PM at Microsoft
âHaoyuan Li, Founder, Chairman, CEO at Alluxio (Series C+), Adjunct Professor at Peking University
đĄ TL;DR â Key Highlights
MCP is explodingâbut security is lagging: Tens of thousands of servers deployed, but few are production-ready.
AI agent autonomy will demand new access control models: The âcruise control to full self-drivingâ shift is coming faster than expected.
Infra bottlenecks are realâand solvable: Latency in cloud storage can cripple inference; caching and data access optimization are huge levers.
Go-to-market still decides who wins: Proprietary data, deep domain expertise, and clear ROI beat âjust another agentâ every time.
⥠MCP Mania: Growth Outpacing Security
Anthropicâs MCP protocol only launched in late 2024âyet by early 2025, there were tens of thousands of MCP servers in the wild. That hockey-stick adoption curve is exciting, but as Rishi Bhargava pointed out, âThe vast majority are local developer deployments. Production-ready remote servers? You can count them on two hands.â
The gap between experimentation and enterprise readiness is wide. Early MCP specs didnât even define authentication, making them unsuitable for exposing sensitive APIs. Aprilâs update added OAuth flows and a clearer split between authorization servers and resource servers, but adoption of those practices is still slow.
Suraj Patel added that in many industriesâlike automotive, healthcare, and logisticsâcore SaaS providers arenât building MCP servers at all. This creates opportunities for startups to be the integration layer, connecting âclosedâ software to the agentic world. But that opportunity is fragile: incumbents can close the gap quickly if they decide to launch first-party agents.
Brian Quinn warned of mismatched expectations: âC-levels believe theyâll get instant access to insights just by turning on MCP. The reality is, it takes iteration to get useful, accurate, and safe results.â
đ Rethinking Access Control for AI Agents
Traditional permission models assume a human is at the keyboard. In an agentic world, that assumption fails.
Rishi framed the shift as the same journey autonomous vehicles took:
Phase 1: Agents act only with tightly scoped, read-only permissions (the âcruise controlâ stage).
Phase 2: Incremental expansion into write actions and system integrations (lane keeping, assisted steering).
Phase 3: Full autonomyâagents can perform end-to-end workflowsâbut only with robust monitoring and fail-safes (full self-driving).
Siddharth Bhai described two dominant models emerging:
Run-as-owner â The agent inherits all the permissions of the invoking user.
Run-as-viewer â The agent can only access resources the viewer is explicitly authorized for.
The âviewerâ model reduces risk but is harder to implementâit requires unified governance across all data sources and AI endpoints. Many teams still default to ârun-as-ownerâ for speed, creating future security debt. As Siddharth put it, âWhatâs expedient isnât always whatâs rightâbut market pressures mean both models will be in play for years.â
đĄď¸ Data Governance as a Competitive Advantage
When agents start pulling customer data, the compliance stakes escalate fast. Brian Quinn described how AppsFlyer built its Privacy Cloud, a secure collaboration layer for marketing analytics. The approach allows advertisers to run AI models over campaign data without directly sharing sensitive identifiers.
In advertising, AI thrives on large, granular datasetsâbut privacy laws demand minimization. âItâs a paradox,â Brian said. âYou need more data for better models, but the law says use less.â
Suraj noted that MongoDB customers are increasingly designing governance into their AI pipelines from day oneâstructuring metadata, implementing role-based access, and separating inference data from training data. Databricks is taking a similar approach with its governance framework, ensuring that policies apply equally to raw datasets, embeddings, and model endpoints.
The emerging consensus: governance isnât just complianceâitâs a differentiator when selling into regulated industries.
âď¸ The Hidden Bottleneck: Cloud Storage Latency
Infrastructure talk got unusually animated when Haoyuan Li shared that some AI teams are running $50M GPU clusters at 30% utilization simply because their data layer canât feed the compute fast enough.
The common culprit: public cloud object storage. While S3 and GCS are cost-effective for massive datasets, theyâre too slow for inference pipelines that require rapid, repeated access. âItâs not unusual to see 10â30 seconds just to list a folder of files,â Haoyuan explained.
The fix is surprisingly straightforward: caching layers that sit between compute and storage. Salesforce, for example, saw nearly 1000Ă improvement in certain inference queries after deploying Alluxioâs caching layer. That kind of boost doesnât just save moneyâit enables product features that were previously impossible due to latency.
For founders, the lesson is clear: infra optimization isnât just for laterâit can be the difference between a functional MVP and a non-starter.
đĽ Multimodal Is HereâAnd Itâs Messy
The industry has moved from single-modality (text-only) to multimodal (text, image, video, audio) faster than many infra teams expected. This introduces complex challenges in storage, search, and retrieval.
Suraj emphasized the central role of metadataâthe descriptive layer that makes it possible to find the right asset without searching the entire dataset. Without rich metadata, embedding search breaks down, especially when mixing modalities.
Haoyuan noted that the multimodal journey mirrors the broader AI lifecycle:
Train or fine-tune the right model quickly.
Deploy to production and scale inference.
Avoid letting the data layer become the bottleneck.
Siddharth shared a tangible example: five years ago, real-time language translation was painfully slow and inaccurate. Today, multimodal translation apps can handle live conversation, switching seamlessly between audio, text, and video. âItâs a leapfrog momentâif the UX is right, adoption can go exponential almost overnight.â
đ¸ Where the AI Infra Money Is Flowing
Suraj walked through the venture funding timeline:
Late 2022âEarly 2023: Massive checks for model hosting, vector databases, and frameworks (LangChain, LlamaIndex).
MidâLate 2023: Shift to application-layer AI and infra that improves inference performance (e.g., Fireworks).
2024â2025: Growing demand for durable execution (handling long-running, multi-step tasks), identity/auth solutions for agents, and evaluation tooling.
Why the shift? Two reasons:
Many âcore infraâ plays have been commoditized by hyperscalers.
Application adoption patterns are changing too fast for single-point infra bets without deep defensibility.
Rishi added that security and identity in the agentic world are still âunsolved and unavoidableââa clear area for future investment.
đ¨ Security Threats in the Age of Open Agents
Letting agents connect to each other or your internal systems introduces a fundamentally different risk profile.
Rishi warned that âall the old modelsâfirewalls, DLP, user-based permissionsâwere built for human workflows. They donât map to agents that can chain actions across systems.â
Recent incidents, like an autonomous agent deleting a production database, underline the need for agent-aware security toolingâfrom permission models to anomaly detection tuned for non-deterministic behavior.
Siddharth likened it to onboarding a new human hire: âYou give them one task, watch how they do it, then slowly increase responsibility. Agents need the same staged trust modelâwith the ability to instantly roll back actions.â
đ Go-To-Market: The Real Differentiator
With hundreds of agent startups launching, Suraj was blunt: âIf you donât have proprietary data, deep workflow expertise, or a measurable ROI, youâre going to be competing with ChatGPT plug-ins for free.â
Brian illustrated what good GTM looks like: AppsFlyerâs creative analytics tool uses computer vision to analyze ad videos, identify the elements that drive conversions, and feed those insights back into creative production. Customers in gaming, e-commerce, and fintech have seen dramatic drops in customer acquisition costs.
Siddharth connected the dots back to his Google Ads days, where aligning sales efforts to customer KPIs unlocked both adoption and upsell. âYouâre not selling AIâyouâre selling a business outcome. The AI is just how you get there.â
Final Takeaway
The panelâs advice to founders was consistent:
Build for trust: Agent security, permissions, and governance are not optional.
Optimize infra early: Bottlenecks at scale are expensive and morale-killing.
Anchor on ROI: Whether itâs time saved, revenue gained, or cost reduced, show it fast.
A huge thank you to everyone who joined us for this deep, candid conversation. If you missed it, weâd love to see you at a future EntreConnect eventâwhether youâre building agents, investing in infra, or just curious about where AI is headed.
Letâs keep learning, building, and pushing the boundaries of whatâs possible. đ
đ Follow us on LinkedIn and Luma for future event invites and exclusive takeaways
đď¸ Join us at our next gatheringâweâd love to see you there!
đ¤ Interested in speaking at a future event? Reach outâweâre always looking for inspiring founders and industry leaders to share their stories & insights!
đŹ LinkedIn Challenge: Share, Learn, Connect
Thank you to everyone who participated in our LinkedIn Challenge! We're thrilled to feature the most engaging and inspiring post (link here), giving our community a chance to celebrate and learn from the experience. We also truly appreciate everyone who shared their best moments and insights with us!
đď¸ Upcoming Events
Aug 24, San Jose, CA | Startup Pitch Salon | Investor Feedback + Audience Vote. ââCalling for startup founders! We are excited to co-host a highly curated startup pitch event with WeShine, bringing together top early-stage founders, seasoned investors, and engaged audiences for an event of insight, feedback, and community.













