Results

Pitchit Comes Out of Stealth as LQaaS Space Continues to Heat Up

Pitchit Comes Out of Stealth as LQaaS Space Continues to Heat Up

The lead qualification as a service (LQaaS) sector has been heating up rapidly as more and more embedded AI capabilities streamline and automate the lead qualification process. One of the biggest barriers to fully automated lead qualification has been the preponderance of data silos that make it difficult to generate a truly predictable lead score with less than complete prospect data. But as AI breaks down data silos,  confidence continues to grow in using SaaS lead qualification tools that augment CRM systems.

But in some industries, speed is more important than deep qualification. Think of telecommunications firms trying to land new customers who might be in the process of switching providers - it is a race to grab attention and response and win over the customer. Time is of the essence…but, you also need to align with the buyer preferences and their own timeline and availability. 

Enter Pitchit. The company says it “automates the first 24 hours of manual labor required to qualify inbound leads.” As soon as a marketing-qualified lead enters a CRM, there is a 24-hour time window where a salesperson must manually qualify the lead as quickly as possible before the lead loses interest. Historically, this stage in the sales pipeline was labor-intensive and required direct phone calls, discovery meetings or live chat conversations to finish the qualification process.

The notion is to quickly engage with prospects and leads, and at the best moment hand off to a human. This enables the AI to do the heavy lifting, sifting through volumes of leads to find those ready to engage, both filtering by data-based qualification as well as readiness to engage with the brand. 

The company says it can sync leads from 7,000+ channels — social media, CRM, email, SMS, and more — to quickly quote pricing, book meetings, handle objections, and capture personally identifiable information (PII) before handing qualified leads off to a human sales rep. 

By focusing on purchase-ready leads, brands with a large volume of B2C sales across multiple consumer touchpoints can dramatically increase sales qualified lead (SQL) conversion rates, lower the time and cost associated with every sale, and maximize revenue per sales representative. In addition, by reducing sales friction, brands can offer their customers a faster and more personalized sales experience while reducing costs associated with lead qualification, like SDR headcount, training costs, etc. 

To date, Pitchit claims it has helped telecom and insurance sales teams qualify 531,000+ leads, save 4,400+ labor hours, book 4,000+ meetings, and close $280 million in customer revenue. Customers using Pitchit experienced a 250% increase in their lead qualification rate, on average. As they come out of stealth, the company has raised a $2.5m seed round.

For telecos and insurance providers, tools like Pitchit can be simple, cloud-based addition to the lead management stack. The company is honing its pricing, but for now the company gets paid based on conversions, which creates a low risk, high reward investment for businesses who have higher volumes of leads, historically low conversion rates, and short time windows to capitalize on consumer interest. 

Marketing Transformation Innovation & Product-led Growth Tech Optimization Future of Work Data to Decisions New C-Suite Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity ML Machine Learning LLMs Agentic AI Generative AI AI Analytics Automation business Marketing SaaS PaaS IaaS Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain CRM ERP finance Healthcare Customer Service Content Management Collaboration Chief Marketing Officer Chief Information Officer Chief Technology Officer Chief Digital Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Executive Officer Chief Operating Officer Chief AI Officer Chief Product Officer

Google Cloud Q2 revenue $10.37 billion, lands $1 billion in operating income

Google Cloud Q2 revenue $10.37 billion, lands $1 billion in operating income

Google Cloud revenue for the second quarter was better-than-expected at $10.35 billion, up from $8.03 billion a year ago. Analysts were modeling $10.2 billion for Google Cloud revenue.

It was the first quarter where Google Cloud topped operating income of $1 billion. Google Cloud reported second quarter operating income of $1.17 billion.

Alphabet, parent of Google, reported second quarter earnings of $18.37 billion, or $1.89 a share, on revenue of $84.74 billion, up 14% from a year ago. Wall Street was looking for second quarter earnings of $1.84 a share on revenue of $84.19 billion.

Analysts were expecting Google to report second quarter earnings of $1.84 a share on revenue of $84.26 billion.

Google’s second quarter earnings landed as the company’s plans to buy Wiz and HubSpot fell apart. Google has gone shopping in recent months to bulk up Google Cloud. Google Cloud also forged a pact with Oracle.

Speaking on an earnings conference call, Pichai said Nvidia's latest platform will be coming to Google Cloud. "We continue to invest in designing and building robust and efficient infrastructure to support our efforts in AI given the many opportunities we see ahead," said Pichai.

He added that Google is also looking to drive efficiency in its AI models and matching "the right model size to the complexity of the query to minimize the impact on costs and latency."

For Google Cloud, the Oracle partnership can embed its AI services into more enterprises. Customer references cited by Google Cloud for AI adoption include Bayer, Best Buy, Discover Financial and TD Bank to name a few. Also see what Equifax and Wayfair have done with Google Cloud.

 

Data to Decisions Digital Safety, Privacy & Cybersecurity Innovation & Product-led Growth Tech Optimization Future of Work Google Google Cloud SaaS PaaS IaaS Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain CRM ERP CCaaS UCaaS Collaboration Enterprise Service Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer

Meta launches Llama 3.1 450B and for Zuckerberg it's personal

Meta launches Llama 3.1 450B and for Zuckerberg it's personal

Meta released Llama 3.1 405B, an open "frontier-level model" that aims performs as well as proprietary models. For Meta CEO Mark Zuckerberg, the Llama cadence is designed to play the long game and bet open-source models ultimately win.

The company released Llama 3.1 405B, which will support synthetic data generation and model distillation. Those two features haven't been available in open-source models. Meta also said it released upgraded 8B and 70B Llama models with context length of 128K and better reasoning. The models are also multilingual and support multiple use cases.

For Zuckerberg, Llama is a mission. In a blog post, he said large language models (LLMs) will develop much like Linux did. First there was Unix and over time open-source Linux won. He said:

"I believe that AI will develop in a similar way. Today, several tech companies are developing leading closed models. But open source is quickly closing the gap. Last year, Llama 2 was only comparable to an older generation of models behind the frontier. This year, Llama 3 is competitive with the most advanced models and leading in some areas. Starting next year, we expect future Llama models to become the most advanced in the industry. But even before that, Llama is already leading on openness, modifiability, and cost efficiency."

Meta's latest Llama release and letter from Zuckerberg are designed to court developers that want to fine tune models, evaluate models for specific applications, pre-train and make models their own. The company said developers can leverage workflows and directions from partners such as Nvidia, AWS, Google Cloud, Microsoft Azure, Dell Technologies, Databricks and others.

Zuckerberg added that developers want affordable model options that can be fine-tuned on sensitive data while avoiding vendor lock-in. Llama is a high-profile effort, but not out of character for Meta, which also led the Open Compute Project.

The win for Meta's approach with Llama is that it can leverage the open-source community and build an ecosystem. Meta wants Llama to be a standard and can play neutral party since its business model isn't about selling access to LLMs.

Enterprises start to harvest AI-driven exponential efficiency efforts | Generative AI use cases, takeaways from projects underway and how the technology fits in with broader digital transformation.

In the end though, Meta's Llama efforts may be personal for Zuckerberg. He said:

"One of my formative experiences has been building our services constrained by what Apple will let us build on their platforms. Between the way they tax developers, the arbitrary rules they apply, and all the product innovations they block from shipping, it’s clear that Meta and many other companies would be freed up to build much better services for people if we could build the best versions of our products and competitors were not able to constrain what we could build. On a philosophical level, this is a major reason why I believe so strongly in building open ecosystems in AI and AR/VR for the next generation of computing."

More:

Data to Decisions Future of Work Innovation & Product-led Growth Tech Optimization Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Executive Officer Chief Technology Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer

CrowdStrike CEO called to testify before House committee

CrowdStrike CEO called to testify before House committee

CrowdStrike CEO George Kurtz is being called to the House to testify about the global IT outage that has hampered enterprises--notably airlines like Delta--for days.

House Committee on Homeland Security Chairman Mark E. Green, MD (R-TN) and Subcommittee on Cybersecurity and Infrastructure Protection Andrew Garbarino (R-NY) sent a letter to Kurtz requesting public testimony.

Kurtz's testimony will be closely watched, but he’s hardly the first technology CEO to testify and take lumps from lawmakers. CrowdStrike’s outage lands just as cybersecurity vendors are pushing platformization in a bid to consolidate IT budgets. While the cybersecurity industry sees innovation from startups and smaller companies, it’s still dominated by a few large vendors, including CrowdStrike. Some vendors, like Palo Alto Networks, advocate for platformization, aiming for even greater consolidation.

Green and Garbarino wrote:

"We write in response to the global information technology (IT) outage, which has been attributed to a “defect” in a CrowdStrike software update that impacted Microsoft Windows. While we appreciate CrowdStrike’s response and coordination with stakeholders, we cannot ignore the magnitude of this incident, which some have claimed is the largest IT outage in history. In less than one day, we have seen major impacts to key functions of the global economy, including aviation, healthcare, banking, media, and emergency services."

The letter also requests that CrowdStrike schedule a hearing with the Subcommittee on Cybersecurity and Infrastructure Protection no later than 5 p.m. July 24.

Delta has struggled with restoring services, but the CrowdStrike outage affected multiple industries. For its part, CrowdStrike has documented what went wrong and published a remediation and guidance hub.

Kurtz said in a statement:

"I want to sincerely apologize directly to all of you for the outage. All of CrowdStrike understands the gravity and impact of the situation. We quickly identified the issue and deployed a fix, allowing us to focus diligently on restoring customer systems as our highest priority."

Digital Safety, Privacy & Cybersecurity crowdstrike Security Zero Trust Chief Information Officer Chief Information Security Officer Chief Revenue Officer Chief Privacy Officer

IonQ's quantum computing bets: Quantum for LLM training, chemistry and enterprise use cases

IonQ's quantum computing bets: Quantum for LLM training, chemistry and enterprise use cases

IonQ outlined it roadmap for its quantum computing stack as well as various use cases as it aims to create enterprise grade infrastructure that could offload GPU workloads if successful.

In a webinar, IonQ CEO Peter Chapman said the company's strategy revolves around performance, scale and enterprise grade infrastructure. If those pillars are weighted equally, IonQ will generate commercial advantage for its quantum computers.

IonQ is approaching $100 million in annual bookings.

Chapman positioned IonQ as a company that's defining use cases and doubling down on the ones with the most promise. "I think we're homing in on exactly what those superpowers are," said Chapman.

Chemistry and quantum drug discovery have emerged as two primary use cases. The chemistry use cases were barely mentioned during IonQ's pre-IPO roadshow, but has emerged as a go-to enterprise application. "There is some brilliant work on the algorithmic chemistry coming," said Chapman.

Constellation ShortLists™

Generative AI is another area where IonQ can add value, said Chapman, who said IonQ has been focused on machine learning since inception. "One of the new things that we're doing this year is applying quantum to large language models," said Chapman. "We're just at the beginning of that journey, but I hope to have results within a year or so."

If successful, Chapman said IonQ "will be able to offload significant workloads from GPUs and significantly reduce energy requirements for data centers and LLMs."

The focus of IonQ's talk is about developing performance with its roadmap, lowering errors and then scaling. "We hope to find commercially significant applications, but the true promise of quantum will need a lot more qubits and faster gate speeds," said Chapman. "Just as importantly we'll need to network these quantum computers and reduce the costs to make them affordable. Future quantum computers are going to be made up of network individual machines so the cost per qubit comes down as computational power increases."

Chapman said the plan is not to over index on performance, scale or enterprise grade because the pillars have to be in balance.

IonQ SVP of Engineering & Technology Dean Kassmann walked through the IonQ roadmap, developing multi-core quantum infrastructure and systems that go beyond what can be simulated on a classical computer. Kassmann also emphasized modular designs, systems that will be put into data centers with classical infrastructure and software that just works.

The plan for IonQ is to get to a point where it can stack its qubits and technology advances together. "A milestone we have planned for 2025 is physical qubits that are photonically interconnected," said Kassmann. "We have a plan set out where we continually build up numbers of qubits by connecting modules together."

Kassmann added that IonQ is focusing on photonic interconnects between quantum systems to scale as well as smaller vacuum packages and refining components. "The work we're doing needs to be on-ramped into the larger engineering pipeline but represents our thought process to scale 2-, 5-, 10-years from now," he said.

Margaret Arakawa, CMO of IonQ, is a Microsoft alum that is focusing on democratizing quantum computing and scaling with its US manufacturing center in Seattle. To sell quantum--an enterprise category that has typically been more about R&D labs--IonQ plans to lead into use cases and solving big problems.

Some use cases include solving for corrosion in the US Navy, which is an expensive problem for the Department of Defense, Airbus optimization and flight orchestration. "We have an applications and algorithm team working on the most important groundbreaking areas a quantum computer can actually address," she said.

In the end, AI may wind up being the play for quantum computing. Arakawa and Chapman both spoke to it and the general idea is to focus on the part of AI that quantum computing can do better than supercomputers.

More quantum computing:

 

     

     

    Data to Decisions Tech Optimization Innovation & Product-led Growth Future of Work Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity Quantum Computing AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Technology Officer Chief Executive Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer

    Verizon sees genAI cost savings with edge revenue in the future

    Verizon sees genAI cost savings with edge revenue in the future

    Verizon is betting that its network--5G and fiber--can be the "backbone of the AI economy" with a boom of edge computing generative AI workloads on deck. The problem is Verizon hasn't seen real revenue yet but appears to be leveraging generative AI to become more efficient.

    Speaking on Verizon's second quarter earnings call, CEO Hans Vestberg outlined the genAI use cases on the revenue and cost side. Vestberg said:

    "We will power the best AI services for our customers. What set us apart with AI is our network's mobile edge computing capabilities and deep fiber footprint. By processing data closer to the source, we enable real-time AI application that requires security, ultra-low latency, and high bandwidth. Our network will be at the forefront of AI and mobile edge compute applications."

    For operations, Vestberg said AI is benefiting customer service by routing calls to agents, improving network design and personalizing offers. CFO Tony Skiadas said:

    "AI is an enabler of efficiencies. You can think about customer care, you can think about the personalization with myPlan. And we see efficiencies coming from there as well."

    In many respects, Verizon resembles many enterprises. Companies are leveraging generative AI to save money and become more efficient. The revenue side of the equation isn't quite baked yet. Enterprises have been expanding generative AI use cases, offering takeaways from projects underway and figuring out how the technology fits in with broader digital transformation.

    Progress Through The Five A's of AI | Margin Compression - Tech Vendors Are You Leading The Way Or In The Way? | Enterprises start to harvest AI-driven exponential efficiency efforts

    Verizon reported second-quarter revenue of $32.8 billion with net income of $4.7 billion. The results were roughly flat with a year ago. Verizon Business revenue was $7.3 billion, down 2.4% from a year ago.

    Verizon sees enterprises building out private 5G networks

    Vestberg said Verizon has more line of sight into the enterprise business and genAI than the consumer opportunity, which will depend on adoption of smartphones and features lead to device upgrades. Vestberg said:

    "I think where I'm most excited is that we have built the Verizon Intelligent edge network which will be the platform for GenAI. You are going to have to have a lot more compute storage at the edge of the network, and that's how we built the network already 2018 with fiber to all our main hubs and between our main subcenters. On top of that, we have cooling and power at those edges. As we go from the LLMs into commercial products for enterprises our network is set up for that. I'm very excited for that opportunity to go forward together with private networks. There are a lot of things coming into GenAI devices, our efficiencies and a business opportunity for us when it comes to AI."

    The problem for Verizon is that the genAI mobile edge payoff isn't here yet. Vestberg said Verizon doesn't have genAI in mobile edge computing to customers but it's clear in conversations with cloud vendors and enterprises that large language models will move to the end points.

     

    Data to Decisions Innovation & Product-led Growth Future of Work Tech Optimization Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Executive Officer Chief Technology Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer

    SAP Q2 cloud ERP revenue up 33%, sees restructuring hitting 9,000 to 10,000 jobs

    SAP Q2 cloud ERP revenue up 33%, sees restructuring hitting 9,000 to 10,000 jobs

    SAP in the second quarter said its cloud revenue was up 25% and CEO Christian Klein said its Business AI efforts are "enabling many deals." The company reiterated its outlook for 2024 and raised its 2025 operating profit guidance due to efficiency efforts.

    The company also said its restructuring efforts will impact 9,000 to 10,000 jobs. SAP said in January it expected to cut 8,000 jobs.

    For the quarter ending June 30, SAP said its SaaS and PaaS revenue was €4.02 billion, up 28%. Cloud ERP revenue was €3.41 billion, up 33%. The Cloud ERP Suite revenue line includes services that are included in RISE with SAP. SAP S/4HANA Cloud, SAP Business Technology Platform, and applications for HR and payroll, spend management, commerce, customer data, business process transformation, and working capital management also fall under Cloud ERP.

    SAP reported second-quarter revenue of €8.29 billion, up 10% from a year ago, with €918 million in net profit. Earnings per share were €0.76. Non-IFRS earnings per share were €1.10.

    In a statement, Klein said "our cloud growth momentum remained strong in Q2, with Business AI enabling many deals." He added that SAP is well positioned to see "accelerating topline growth through 2027."

    CFO Dominik Asam added:

    "We are staying squarely focused on delivering our outlook for this year. Our current cloud backlog growth during the second half of 2024, and especially in Q4 will be decisive to lay a solid foundation for our cloud revenue ambition for 2025. At the same time, we'll continue to execute against our transformation plan to achieve our 2025 free cash flow ambition despite a mid-triple-digit million cash out for restructuring spilling into next year."

    On an earnings conference call, Klein said SAP is landing and expanding as it moves large enterprises to the cloud. "Since Sapphire, we have seen a lot of additional interest from big customers, over 90 partner use cases for co-innovation, including use cases with big systems integrator," said Klein. "Sapphire also helped to significantly boost our pipeline. It's not just a lift and shift to the cloud. It's a holistic offering to increase competitiveness through a deep business transformation to replace the legacy ERP with our modular cloud ERP."

    Klein said he was confident that SAP was executing well and seeing a "land and expand" cadence at large enterprises like Exxon." He added that SAP is also investing in SMBs with partners. 

    By the numbers for the second quarter:

    • Current cloud backlog was up 28% to €14.81 billion.
    • Software license revenue fell by 28% to €0.2 billion.
    • SAP had repurchased 12,895,525 shares at an average price of €145.20 as of June 30.

    SAP also said that its restructuring program is expected to finish in early 2025. SAP said the restructuring will affect 9,000 to 10,000 positions, but the company expects headcount to be similar to 2023.

    As for the outlook, SAP projected €17.0 billion to €17.3 billion cloud revenue for 2024, up 24% to 27% from 2023 with €29.0 billion to 29.5 billion in cloud and software revenue for the year.

    For 2025, SAP projected non-IFRS operating profit of approximately €10.2 billion, up from previous guidance of €10 billion. The rest of SAP's 2025 outlook remains the same including cloud revenue of more than €21.5 billion and total revenue of more than €37.5 billion.

     

     

    Data to Decisions Digital Safety, Privacy & Cybersecurity Innovation & Product-led Growth Tech Optimization Future of Work SAP Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer

    Quantum Computing at Scale | Interview with Classiq CEO Nir Minerbi

    Quantum Computing at Scale | Interview with Classiq CEO Nir Minerbi

     

    Quantum computing software provider Classiq has been busy forging partnerships with everyone from Nvidia to BMW to Citi as it aims to expand enterprise use cases with a software layer that abstracts the underlying hardware.

    In the latest Constellation Insights interview, Editor-in-Chief Larry Dignan sits down with Nir Minerbi, Co-founder & CEO of Classiq Technology to discuss Classiq's end-to-end Quantum Software Platform that enables quantum computing at scale.

    On Insights <iframe width="560" height="315" src="https://www.youtube.com/embed/0RDe2hKmLRM?si=KulzLgUBaXmV43kf" title="YouTube video player" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>

    From Blue Screen to Blackout: Unpacking the CrowdStrike Catastrophe and Industry Implications

    From Blue Screen to Blackout: Unpacking the CrowdStrike Catastrophe and Industry Implications

    I’m not an early riser. By the time I woke up on Friday, the world had already experienced one of the most massive outages in recent history courtesy of CrowdStrike. My phone was buzzing with missed calls, texts, and inquiries from CxOs and media asking what happened, why it happened, and what it means for the cybersecurity industry. I spent the rest of my day on calls while following stories about people flying with paper boarding passes, frustrated passengers at airports, happy employees who got an unexpected day off, and anxious, helpless, and angry admins on the frontlines of this outage. I also spoke to The Wall Street Journal and TechCrunch for their coverage of this story. 

    Incident Details and Response from CrowdStrike

    Since the incident, we’ve learned that it was caused by a botched automated update from CrowdStrike, affecting only Windows machines, not Mac or Linux. We still don’t know why CrowdStrike didn’t catch this error before it reached millions of end users and bricked their devices. We’re patiently waiting for the root cause analysis (RCA).

    The initial response from George Kutz, the CEO of CrowdStrike, was less than ideal. He downplayed the incident that crashed most Fortune 500 Windows devices with a blue screen of death (BSOD) and caused an estimated $1 billion in damage, referring to it as an “inconvenience.” Though he later apologized, the damage was done. People never forget the first response. It’s possible CrowdStrike didn’t initially realize the gravity of the situation. CrowdStrike has been working hard to help customers restore their systems and regain their trust. Shawn Henry, the chief security officer of CrowdStrike, has given the most empathetic response I've seen.

    Broader Industry Implications

    While the cybersecurity industry sees innovation from startups and smaller companies, it’s still dominated by a few large vendors, including CrowdStrike. Some vendors, like Palo Alto Networks, advocate for platformization, aiming for even greater consolidation. In the coming days, we’ll hear many stories about CrowdStrike’s engineering culture, business practices, and leadership. What’s evident is what we engineers have always known: systems with a single point of failure (SPOF) are extremely vulnerable if not managed properly. Just ask public cloud providers powering businesses worldwide. A stringent DevOps process, significant investment in site reliability engineering (SRE), gradual rollouts, canary zones, ultra-fast rollbacks, and state-of-the-art resiliency plans are essential. Unfortunately, CrowdStrike is not a native cloud company. Their primary product, an endpoint detection and response (EDR) tool, is installed on devices (endpoints) they secure. For Windows machines, this requires low-level highly privileged operating system access to install and run the Crowdstrike software agent called Falcon.

    Source: Twiiter/X

    Microsoft’s Role and Customer Impact

    Some might argue that this wasn’t Microsoft’s problem since they’re just an operating system provider. I argue the opposite. The fundamental design of Windows that requires such access was the primary reason CrowdStrike could push a botched automated update deep into the operating system without adequate checks by Microsoft. Customers using Mac machines weren’t affected because Mac, with its underlying Linux architecture, doesn’t allow the same level of access to third-party vendors, yet it provides the same functionality. Microsoft acknowledged the broad economic and societal impacts of this outage but called it "infrequent"; it’s like saying terrorism is infrequent—it doesn’t lessen the severity. In fact, the very infrequent nature of such incidents makes them hard to detect and protect against. The CxOs and system administrators I talked to are upset and angry. They’re reconsidering whether to keep automatic updates enabled, and many are actively considering moving away from Windows unless Microsoft addresses the underlying architectural flaws related to installing third-party agents.

    Agentless Security and Future Directions

    This incident also sparked a discussion on the agentless approach many cybersecurity solution providers are adopting, particularly in cloud environments and OT devices like industrial controllers and medical equipment. The cloud can be secured without running an agent, and OT devices are often too complex and proprietary for agent installation. We’ll likely hear more about achieving security without invasive methods, such as installing agents deep into operating systems. Additionally, there will be discussions on modernizing DevOps, air-gapping updates, and various efforts to balance security with business continuity.

    A Wake-Up Call for the Industry

    This happened to CrowdStrike, but it could easily have happened to any other EDR vendor with agents running on customers’ devices. It could also happen to non-cybersecurity vendors requiring deeper OS access, like screen-sharing software or asset management agents. This is a reckoning moment for an industry reexamining its dependencies on technology for business continuity. Many customers I spoke to didn’t have the disaster preparedness they needed, including tabletop exercises, disaster drills, and post-incident command centers. Post-breach or post-incident resiliency plans are non-negotiable. Airline customers won’t remember CrowdStrike, but they won’t forget being stranded at an airport staring at a blue screen of death, unable to get a boarding pass. You own your customer experience and need to control that, not your vendors.

    Looking Ahead

    In the coming days, I look forward to a debate on:

    • Architectural approaches for third-party agents requiring deeper operating system-level access
    • Accountability of operating system vendors and third-party software vendors that can break devices causing significant business continuity challenges
    • The concentrated cybersecurity landscape, platformization, and single point of failure
    Digital Safety, Privacy & Cybersecurity Chief Information Officer Chief Information Security Officer Chief Technology Officer

    Enterprises start to harvest AI-driven exponential efficiency efforts

    Enterprises start to harvest AI-driven exponential efficiency efforts

    Enterprises are beginning to leverage AI and widen their profit moats just as some companies are seeing customers wobble. The efficiency gains are beginning to highlight how digital transformation and AI strategies are becoming self-funding, keeping expenses below the rate of inflation and optimizing processes.

    Not surprisingly, the companies that are using AI leverage happen to be in regulated industries and enterprises that have previously invested in data and digital transformation. Why? These companies tend to have strong data quality, governance and strategies.

    The payoff from these AI and data transformation efforts will be critical if and when customers pull back on spending. A growing number of CEOs sound like Shelley Simpson, CEO of J.B. Hunt.

    "While admittedly the market has been challenging, we have invested throughout this downturn to set us up for future growth and success across the business. We continue to focus on controlling expenses in the near term without jeopardizing our long-term potential, managing our headcount through attrition, while at the same time continuing to deploy and enhance our technology to increase the productivity of our people," said Simpson.

    This post first appeared in the Constellation Insight newsletter, which features bespoke content weekly and is brought to you by Hitachi Vantara.

    Here are 6 vignettes to ponder from enterprise technology's buy side and how companies are driving toward exponential efficiency. These scenes are part of an ongoing series looking at generative AI use cases, takeaways from projects underway and how the technology fits in with broader digital transformation.

    Progress Through The Five A's of AI | Margin Compression - Tech Vendors Are You Leading The Way Or In The Way?

    Bank of America

    Bank of America expects to spend almost $4 billion on technology initiatives this year focused on AI for its employees and clients. The bank's client insights tool has served up more than 6 million insights to financial advisers to give them proactive reasons to engage.

    CEO Brian Moynihan said Bank of America's mobile banking app has more than 47 million active users and digital sales are now 53% of total sales for its consumer business. On the cost side, Bank of America has saved money by streamlining client service requests using AI.

    And digital is more than cost. "Digital adoption and engagement continue to improve, and customer satisfaction scores remain near record levels, illustrating customer appreciation of our enhanced capabilities due to our continuous investment," said Moynihan.

    The leverage in Bank of America's digital strategy is also in what it calls its "operational excellence platform." Bank of America in its second quarter held its expense growth to 2% from a year ago and below inflation rates.

    Bank of America has also kept headcount roughly flat. Moynihan said the company has managed expenses well with the help of AI and digital. "We have huge cleanup stuff going on. We have the new initiatives that free up work," he said.

    PepsiCo

    In the second quarter, PepsiCo CEO Ramon Laguarta said consumers are being more selective as inflation takes a toll on household budgets. PepsiCo organic revenue growth was 1.9% in the quarter compared to 13% a year ago.

    The upshot is that PepsiCo is hitting the limits of price increases. PepsiCo saw its annual revenue surge nearly $7 billion between 2020 and 2023, but noted that "the impacts of persistent inflationary pressures and higher borrowing costs over the last few years have resulted in tighter household financial conditions. Consumers have become more value-conscious with their spending patterns and preferences across brands, packages, and channels."

    With consumers becoming tapped out, PepsiCo plans to fund investments going forward with productivity savings. PepsiCo has its process optimization game down.

    The company plans to:

    • Invest in more automation within warehousing, transportation and logistics.
    • Advance its digitization efforts and harmonize IT systems.
    • Optimize processes across the value chain.
    • Enhance analytics for trade promotions, consumer insights, supply planning and demand forecasting.
    • And optimize advertising and marketing spend.

    Laguarta said: "We're managing total PepsiCo operating margin and as you've seen, we keep improving the margin. This quarter had almost 100 bps of operating margin improvement, and it's been consistent for the last few years. We feel good about our productivity pipeline, it’s not tactical, it's super strategic and it's multiyear and it's based on automation, it's based on digitalization, simplification of the company, standardization, different service models to the business. So, there is a whole portfolio of productivity ideas that are multiyear in nature and we don't think that we will slow down our productivity in the coming years."

    UnitedHealth Group

    UnitedHealth reported better-than-expected second quarter results as it grows revenue and optimizes operations. CEO Andrew Witty said: "Our growing AI portfolio made up of hundreds of practical use cases will generate billions of dollars of efficiencies over the next several years. These investments enable us to improve consumer experience, enhance provider find and price care capabilities to meet people's needs and improve clinical back-office execution. We expect technology innovation to become an increasingly core driver of our growth over the next two to five years."

    UnitedHealth is still taking an earnings hit due to the Change Healthcare cyberattack, but its focus on efficiencies is minimizing the bottom line hit.

    Witty said the AI use cases identified by UnitedHealthcare will drive margins and cost savings through the next two years. Witty added that AI and process improvements go together and that AI will create a "fundamental reimagination of business processes" and "allow an existing process to run more efficiently.”

    "We actually take steps out of a process and really start to change things. I'd call out payment integrity as a front-runner in that particular regard," said Witty.

    A few examples:

    • UnitedHealth onboarded a record number of OptumRx customers but spent 9% less this year than last year due to digitization.
    • OptumHealth added nearly more than 1.5 million customers with zero increase in personnel headcount in risk-based businesses.
    • UnitedHealth is proactively engaging with three-fourths of its members using technology, up from 62% a year ago.

    Witty said UnitedHealth is doubling down on its own cost management efficiency and productivity and that focus "coincides with an extraordinarily and exciting moment around technological innovation, whether that's Generative AI, digitization, all wrapped together in our march toward a greater consumer focus within the organization."

    Omnicom Group

    Omnicom said it has expanded its genAI efforts for both clients and internal use cases.

    The advertising and public relations giant's genAI strategy revolves around becoming more efficient and selling products and services to clients. CEO John Wren said on the company's second quarter earnings call: "We also announced first-mover collaborations with Adobe, Amazon, Getty, Google and Microsoft's OpenAI to gain early access to their large language models. Just over a year later, we're seeing these generative AI platforms' tools and partnerships being activated throughout every area of our business from strategy to creative to production, media and precision marketing."

    For instance, Omnicom's TBWA unit launched Collective AI, a suite of tools for employees and clients that automates basic tasks and then provides insights. Collective AI uses TBWA's archives to train large language models. ArtBotAI is a content orchestration platform to create digital assets and personalized experiences.

    Omnicom also acquired Flywheel, which focuses on retail media and e-commerce. The idea is to combine Flywheels' commerce product and transaction signals with Omnicom's audience and viewership data, said Wren, who noted that Omnicom has also centralized production to go along with the AI investment.

    "From Gen AI to e-commerce to production, we are continuing to enhance our offerings to meet our clients' needs for better inform strategic insights using AI, creatively inspired content that can be personalized at scale and investments in targeted media that can be measured through quantifiable outcomes, all delivered in the most efficient and effective manner," said Wren. "We decided that the only way that we're going to efficiently and effectively grow, especially in this AI environment, which is going to change those legacy production businesses was in fact to centralize it."

    Goldman Sachs

    Goldman Sachs' CEO David Solomon has talked up the company's AI initiatives before. He said the company has leveraged AI for multiple use cases ranging from coding to equity research and content.

    "We are focused on how you can create use cases that increase your productivity," said Solomon. "If you look and you think across the scale of our business, I think you can think of lots of places where the capacity to use these tools to take work that's always been more manual and allow the very smart people to do that work to focus their attention on clients."

    Solomon said the AI spending boom is real and can drive productivity and revenue gains. Solomon said: "I am particularly encouraged by the ongoing advancements in artificial intelligence. Recently, our Board of Directors spent a week in Silicon Valley where we spoke with the CEOs of many of the leading institutions at the cutting-edge of technology and AI. We all left with a sense of optimism about the application of AI tools and the accelerating innovation in technology more broadly. The proliferation of AI in the corporate world will bring with it significant demand-related infrastructure and financing needs, which should fuel activity across our broad franchise."

    WD-40

    While many of the examples revolve around AI, it's worth noting that there are plenty of companies doing the data strategy heavy lifting to take advantage of the technology later. WD-40 is a good example.

    WD-40 CFO Sara Hyzer said on the company's third quarter earnings conference call that it is updating its ERP system, launching Salesforce and looking to "drive productivity via enhanced systems."

    Hyzer said WD-40 went live with the first phase of its ERP system across much of its business including the US, Latin America and Asia regional distributor businesses. The implementation resulted in some minor disruptions in the third quarter, but most of the critical issues have been resolved.

    "We know there is still work to do and have several enhancements that are already being worked on, which is not unexpected at this phase of the project. Most importantly, we have gained numerous learning moments from this implementation, allowing us to make process improvements and become more proactive," said Hyzer.

    The other move for WD-40 is to roll out Salesforce in the US to drive sales efficiencies and reduce costs. "We also know that use of data analytics and automated tools, leveraging data is increasing and can be a real enabler for the business," said Hyzer. "The foundational work we are doing now around data governance, centralizing our data architecture and data quality management will allow our people to leverage our data quicker and drive better decision-making."

    Insights Archive

     

    Data to Decisions Future of Work Next-Generation Customer Experience Tech Optimization Innovation & Product-led Growth New C-Suite Digital Safety, Privacy & Cybersecurity AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Executive Officer Chief Financial Officer Chief Information Officer Chief Experience Officer Chief Technology Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer