Results

SAP retools for generative AI, cuts 8,000 jobs, sets 2024, 2025 ambition

SAP plans to restructure and cut 8,000 positions as it focuses on artificial intelligence and uses the technology to become more efficient. SAP said it expects to add back new roles and exit 2024 with headcount at similar levels.

The enterprise software giant ended the year with nearly 108,000 full-time employees.

In a statement, SAP said it plans to retool to focus on growth markets and business AI. SAP said its restructuring will be covered by "voluntary leave programs and internal re-skilling measures." Restructuring charges will be about €2 billion with most of that being recognized in the first half of 2024 with minimal cost benefits due to reinvestment.

On a conference call with analysts, SAP CEO Christian Klein said:

"The tech industry is moving fast. We need to keep leading the way as a top enterprise application company and further advance to become the number one Business AI company as well. This is why out of a very strong position we are now accelerating the development of the company with the clear goal to grasp the opportunities of GenAI."

"We are stepping up our investment in Business AI to drive automation as we see significant growth opportunities lying ahead and want to improve our operating leverage," said SAP CFO Dominik Asam.

As for SAP's 2024 outlook, the company said €17 billion to €17.3 billion in cloud revenue, up 24% from €13.66 billion in 2023. The company said software and cloud revenue will be €29 billion to €29.5 billion with €7.6 billion to €7.9 billion non-IFRS operating profit.

SAP also outlined its 2025 outlook including cloud revenue of more than €21.5 billion and total revenue of more than €37.5 billion. Non-IFRS operating profit by 2025 will be €10 billion.

With the restructuring plan and outlook, SAP also reported fourth quarter and fiscal 2023 results. The company is retooling to drive cloud sales and move customers to S/4HANA. Customers have had a mixed reaction and SAP's 2023 net Promoter Score is 9. According to Simplesat, the average NPS for SaaS businesses is 30 and enterprise software's average NPS is 44.

SAP said fourth quarter revenue was €8.47 billion with cloud revenue of €3.7 billion. SAP S/4HANA cloud revenue was €1.03 billion. Earnings per share were €1.01 a share. Adjusted earnings were €1.41.

For fiscal 2023, SAP reported revenue of €31.26 billion with a profit of €5.93 billion.

Klein said that SAP can differentiate with its AI platform. He said:

"We are developing strong organic product, a strong organic AI platform so that our copilot tool can speak not only finance but can solve some of the hardest problems our customers facing across the company. We are going to infuse it right into the business processes. When you look at what we already can do in particular sales and optimizing inventory, it can take out a ton of CapEx and OpEx of the P&L or balance sheet of our customers. And then when you listen to our partners like Microsoft or NVIDIA, where we just closed another partnership, I mean they are keen actually now to combine their copilot with our copilot to extend our AI platform. When you have content from over 30,000 customers and access to the most mission-critical data, the algorithms become smarter every day. We can actually solve some problems."

Other takeaways to note:

  • SAP said Vodafone is betting on RISE with SAP and is using Signavio and Datasphere along with other apps. Volkswagen is an expanded customer win for SAP SuccessFactors.
  • Total cloud backlog is €44 billion, up 39% from a year ago.
  • SAP's top 1,000 customers are now on average using 4 SAP cloud solutions, up from 3 last year.
  • SAP's Cloud ERP suite represents 82% of the company's SaaS and PaaS revenue.
Data to Decisions Digital Safety, Privacy & Cybersecurity Innovation & Product-led Growth Tech Optimization Future of Work Next-Generation Customer Experience SAP AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer Chief Executive Officer Chief AI Officer Chief Analytics Officer Chief Product Officer

Verizon sees enterprises building out private 5G networks

Verizon said large enterprises are building out private 5G networks for connectivity and edge computing use cases.

The wireless giant disclosed the 5G private network buildout on its fourth quarter earnings call.

Verizon CEO Hans Vestberg said on the company's earnings conference call:

"We continue to see interest from large enterprises running complex logistics and operations like ports, automotive and heavy industries. In November, Norfolk International Terminal contracted us to build a second private 5G network for them. Audi, already one of our partners in smart car development, has contracted us to build a private network for their automotive tech testing environment. And Nucor, one of the country's largest steel companies, has us building private networks for three of its sites with more to come over the next year."

Vestberg added that private 5G networks are a new source of revenue for Verizon's enterprise unit. Verizon has built out 5G private networks for the NFL across stadiums.

"When we build relationships with these large enterprises and they see what our network can do for them, there is always potential for more business," said Vestberg.

What's notable about Verizon's 5G networks is that it can play into edge computing and connectivity to process more AI workloads locally.

Other takeaways from Verizon's fourth quarter earnings:

  • Vestberg touted a partnership with HCLTech on post-sale implementations and customer support for managed network services. He said HCLTech has enhanced "our customer service while saving Verizon money."
  • Regarding HCL Vestberg said: "On the cost program, '23 was a big year for us. We did a lot in customer care. We did a lot with the managed services with outsourcing with HCL. We implemented some really large IT systems. We continue to deploy our offshore centers and being even stronger using that as a platform on the basis that we created the Verizon Global Services. I'm really pleased with the platforms, and that means we're on track for the savings we talked about going into '24."
  • Total wireless postpaid phone net additions were 449,000 in the fourth quarter.
  • Verizon's adjusted fourth quarter earnings were $1.08 a share, on revenue of $35.1 billion, down 0.3% from a year ago. Consolidated net loss in the fourth quarter was $2.6 billion.
  • Verizon's business revenue was $7.6 billion, down 3.6% from a year ago. Verizon Business had 292,000 wireless retail postpaid net additions in the fourth quarter.
  • Verizon is projecting wireless revenue growth of 2% to 3.5% in 2024.

​​​​​​​

Next-Generation Customer Experience Tech Optimization Innovation & Product-led Growth Data to Decisions Chief Information Officer

The immovable opacity of AI meets the irresistible force of privacy

 

The power of Data Privacy law

Large language models (LLMs) and generative AI are developing at ever-increasing rates, alarming many commentators because it is so hard now to tell fact from fiction.  Deep fakes were a central issue in the recent Hollywood writers’ strike, with many creators and actors anxious to protect their personal identities against the possibility of being replaced by synthetic likenesses.

Naturally there are calls for new regulations. We should also be looking at how AI comes under the principles-based privacy regulations we already have.

There is a large and growing body of international principles-based data privacy. These are based on the idea of personal data, which is broadly defined to mean essentially any information which is associated or may be associated with an identifiable natural person. Data privacy laws such as the GDPR (not to mention 162 national statutes) operate to restrain the collection, use and disclosure of personal data.

Generally speaking, these laws are technology neutral; they are blind to the manner in which personal data is collected. And they apply to essentially any data processing scenario.  

So, this means that the outputs of AIs, when personally identifiable, are within the scope of data privacy laws in most places around the world. If personal data comes to be in a database by any means whatsoever then it may be deemed to have been collected.

Thus, data privacy laws apply to personal data generated by an algorithms, untouched by human hands.

Surprise!

Time and time again the privacy implications of automated person information flows seem to take technologists by surprise:

  • In 2011 German privacy regulators found that Facebook’s photo tag suggestions feature violated the law and called on the company to cease facial recognition and delete its biometric data. Facebook took the prudent approach of shutting down its facial recognition usage worldwide, and subsequently took many years to get it going again.  
  • The counter-intuitive Right to be Forgotten (RTBF) first emerged in the 2014 “Google Spain” case heard by the European Court of Justice.  The case was not actually about “forgetting” anything but related specifically to de-indexing web search results. The narrow scope serves to highlight that personal data generated by algorithms (for that’s what search results are) is covered by privacy law. In my view, search results are not simple replicas of objective facts found in the public domain; they are the computed outcomes of complex Big Data processes.

While technologists may presume (or hope) that synthetic personal data escapes privacy laws, the general public would expect there to be limits on how information about them is generated behind their backs by computers. In many ways, judgements produced by algorithms raise more concerns that traditional human judgements.

What’s next?

The legal reality is straightforward. If an information system comes to hold personal data, by any means, then the organisation in charge of that system has collected personal data and is subject to data privacy laws.

As we have seen, analytics and Big Data processes have been brought to heel by data privacy laws.

Artificial Intelligence may be next.

Responsibility for Simulated Humans

Large language models are enabling radically realistic simulations of humans and interpersonal situations, with exciting applications in social science, behaviour change modelling, human resources, healthcare and so on.

As with many modern neural networks, the behaviour of the systems themselves can be unpredictable. A recent study by researchers at Stanford and Google revealed “simulacra” (that is, robotic software agents) built on ChatGPT spontaneously exchanging personal information with each other, without being scripted by the software’s authors.

That is, the robots were gossiping, behind the backs, as it were, of the humans who developed them.

If this level of apparent autonomy is surprising, then bear in mind widespread reporting that nobody knows exactly how Deep Neural Networks work.

Bill Gates calls AI the most powerful technology seen in decades.  Given how important it is, can society accept that AI leaders and entrepreneurs can’t tell us what’s going on under the hood? 

Ignorance is No Excuse

Well-established privacy law shows that AI’s leaders might have to take more interest in their creations’ inner workings. Regulators might not find it acceptable that AI operators can’t necessarily tell how personal data arises in their systems. By the same token, they cannot even be sure what personal data is being generated internally and retained.

If a large language model generates personal data, then the people running the model are in principle accountable for it under data privacy rules. And it may not matter to regulators if the knowledge personal data is distributed through an impenetrable neutral network of parameters and weights buried in hidden layers

Privacy law requires that any personal data created and held by an LLM must be collected for a clear purpose, the collection must be proportionate to that purpose, and it must be transparent.  Personal data created in an LLM must not be used or disclosed for unrelated purposes (and in Europe, the individuals concerned have further rights in some cases to have the data erased).

I am not a lawyer, but I don’t believe that the owner of a deep learning system that holds personal data can excuse themselves from technology-neutral privacy law just because they don’t know exactly how the data got there.  Nor can they get around the right to erasure by appealing to the weird and wonderful ways that knowledge is encoded in neutral networks.

If an AI’s operator cannot comply with data privacy law, then a worst case scenario could see an activist data protection authority finding the technology to be unsafe, and ruling that it be shut down until such time as personal data flows can be fully and properly accounted for.

Data to Decisions Digital Safety, Privacy & Cybersecurity New C-Suite Innovation & Product-led Growth Next-Generation Customer Experience Future of Work Tech Optimization ML Machine Learning LLMs Agentic AI Generative AI Robotics AI Analytics Automation Quantum Computing Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain Leadership VR Security Zero Trust business Marketing SaaS PaaS IaaS CRM ERP finance Healthcare Customer Service Content Management Collaboration Chief Analytics Officer Chief Data Officer Chief Digital Officer Chief Information Officer Chief Information Security Officer Chief Privacy Officer Chief Technology Officer Chief Executive Officer Chief AI Officer Chief Product Officer

Oracle launches OCI Generative AI service, plans to embed throughout databases, applications

Oracle said its generative AI managed service on Oracle Cloud Infrastructure (OCI) is generally available and the enterprise software giant said it plans to infuse it throughout its database and application offerings.

The company said its OCI Generative AI service will integrate large language models (LLMs) from Cohere and Meta Llama 2 for multiple business cases. Those two options fall short of the number of models offered by other hyperscalers, which have taken a mall approach to models, but the way Oracle is deploying the generative AI service may appeal to enterprises.

In a blog post, Greg Pavlik, SVP OCI, said:

"We took a holistic approach to generative AI as we thought through the complete picture of what enterprises truly need to successfully implement generative AI. We’re also increasingly adapting models to real-world enterprise scenarios."

More:

OCI generative AI service can be used in Oracle Cloud or on-premises via an OCI Dedicated Region. That twist may appeal to regulated industries that Oracle already caters to, said Constellation Research analyst Andy Thurai. He said:

"While OCI's managed LLM as a service, via API access, is a compelling option, it is currently limited to just Cohere and Meta's Llama 2. Currently, the use cases are also very limited to text generation, summarization, and semantic similarity tasks.

Oracle's option to use the generative AI service in the Oracle cloud and on-premises via OCI dedicated region is a somewhat unique proposition that might be interesting to some large enterprise customers -- especially the ones in regulated industries.

In terms of overall generative AI offerings, Oracle is far behind all three cloud providers. However, the option to integrate generative AI in Oracle's ERP, HCM, SCM, and CX applications running on OCI could make this offering more attractive, if priced right."

Doug Henschen, Constellation Research analyst, said:

"It's notable that Oracle is hosting two moderately sized foundation models that  promise lower-cost operation than the large public models. Use of Cohere models will be indemnified by Oracle while Llama 2 is an open source option that will enable customers to build custom models. Being hosted on OCI, both options keep data and model training entirely inside Oracle's cloud, avoiding cross-cloud calls to external public models."

Here are the key points of the announcement at Oracle's CloudWorld Tour:

  • OCI Generative AI service supports more than 100 languages with improved GPU cluster management.
  • Oracle will embed AI across its applications and converged database with pre-built services instead of a tool kit.
  • Enterprises can consume the service via API calls.
  • Customers can refine the OCI Generative AI service models with retrieval augmented generation (RAG) techniques. To that end, OCI Generative AI Agents is in beta and can combine with a RAG agent for customization.
  • The OCI Generative AI Agents beta supports OCI OpenSearch, but Oracle said it will support search and aggregation tools to Oracle Database 23c and MySQL HeatWave with vector search tools.
  • Oracle said it will deliver prebuilt agent actions across its suite of SaaS applications.

Constellation ShortList™ Artificial Intelligence and Machine Learning Cloud Platforms | Trust In The Age of AI

In addition, Oracle launched OCI Data Science AI Quick Actions, a no-code feature of the OCI Data Science service that will enable integration with multiple LLMs and open-source models.

The company said it also improved its existing AI offerings for vision, speech, document understanding and translation.

Data to Decisions Innovation & Product-led Growth Future of Work Tech Optimization Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity Oracle ML Machine Learning LLMs Agentic AI Generative AI Robotics AI Analytics Automation Quantum Computing Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain Leadership VR business Marketing SaaS PaaS IaaS CRM ERP finance Healthcare Customer Service Content Management Collaboration Chief Information Officer Chief Executive Officer Chief Technology Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer

Supply chain transformation critical as resilience worries stack up

This post first appeared in the Constellation Insight newsletter, which features bespoke content weekly and is brought to you by Hitachi Vantara.

Confidence in the global supply chain's continuity and resilience is waning and it's a topic we're likely to hear more about as companies report fourth quarter earnings in the weeks ahead due to geopolitical and climate disruptions. Get ready for another supply chain transformation investment cycle.

The Dun & Bradstreet Global Supply Chain Continuity Index was 47.9 in the first quarter of 2024, down from 51.1 in the fourth quarter of 2023. Why? Geopolitical tensions in different parts of the world, trade disputes and climate-related issues are disrupting trade routes causing both higher delivery costs and delayed delivery times.

Enterprises are trying to re-route supplies through regions insulated from geopolitical issues. As a result, supply costs are increasing due to longer routes, higher freight charges and insurance costs.

There are two ways to look at what is emerging as the next supply chain crisis. First, higher supply chain costs will crimp profits as well as boost inflation. On the bright side, another crisis is going to accelerate the rewiring of the increasingly intelligent supply chain as technologies such as artificial intelligence are used.

My working theory is that these intelligent supply chains are a precursor to broader value chains. Supply chain software such as Manhattan Associates and Blue Yonder are working to integrate data, deliver insights and partner.

Raj Subramaniam, President and CEO of FedEx Corp., said it's imperative companies optimize supply chains. Speaking at the National Retail Federation (NRF) conference, Subramaniam said "you want to optimize your supply chain end-to-end and deliver a better, seamless ecommerce experience. Our mission is to make supply chains smarter for everyone."

FedEx had launched the fdx platform, which integrates FedEx capabilities to give merchants conversion tools, control of shipments, increased visibility, streamlined returns and data on carbon emissions impact. "Physical networks get us where we need to be, but digital supply chains make that journey faster, more precise and — equally critical — more seamless," said Subramaniam.

UPS has launched its own supply chain platform for businesses in November.

Walmart CEO Doug McMillon outlined the company's supply chain transformation at CES 2024. McMillon said Walmart built a general merchandise supply chain with a distribution center as it scaled. In the late 1990s, Walmart built a supply chain for e-commerce. Ultimately, Walmart had three different supply chains.

"For a few years now, we've been working to link them. We're building an intelligent, connected and more automated network, one that already moves more than 100 billion individual items every year. Now we can do that in a smarter fashion. I've been with this company for over 30 years and there's never been a period of transformational change like the one we've started with our supply chain," said McMillon.

Constellation Brands John Kester, senior vice president of operations services at the company, said the company has been streamlining its supply chain by making it more digital and proactive. Speaking at Constellation Brands recent investor day, Kester said "what's even more exciting is the future unlocks the digital tools can deliver beyond the base business case."

Specifically, Kester said Constellation Brands is creating planning signals throughout the supply chain. Those signals will reduce inventory, create more accurate cycle and safety stock levels and improve on-time and full shipments. Data from suppliers to warehouses to distributors to retailers will connect.

In many ways, supply chain transformation projects are insulated from macroeconomic issues. Since supply chain transformation can simultaneously save money and solve sustainability compliance requirements, there's usually a solid return on investment case to be made. Even if a long-term lens is needed for supply chain transformation projects, there are plenty of shorter-term returns available.

Here are a few takeaways on supply chain transformation from Suresh Kumar, Walmart's Chief Technology Officer and Chief Development Officer.

  1. Intelligent supply chains require systems integration and scale. Kumar said Walmart's three supply chain systems worked well individually. The power was in reimagining the entire systems and reconfiguring parts.
  2. A flexible data architecture. It almost goes without saying that if a company doesn't have its data game down, the supply chain transformation will suffer. Kumar noted multiple data points that drive its automated supply chain.
  3. Intelligence, via artificial intelligence, machine learning, data and analytics, needs to be able to forecast what customers want and when they want it and orchestrate the movement of different products that are stored in different ways.
  4. Forecasting is critical for both suppliers and customers. Suppliers need to see demand changes in near real time as well as customers. "We built an industry leading forecasting system that is smart. It's automated and it uses a patent pending machine learning model that predicts customer behavior and helps us accurately forecast how much of a product is needed and where," explained Kumar. "The system incorporates dozens of different types of data. Like historical sales data, but also things like weather forecasts, the overall popularity of an item compared with last year, but also how an item is trending on social media."
  5. AI has to model and orchestrate the movement of inventory at scale. "We also built artificial intelligence into how we orchestrate the optimal movement of our inventory. The main job is to have the product where our customers needed the most," said Kumar. "The AI system also redistributes autonomously. If the demand for an item spikes in one area of the country, our automated system redistributes the merchandise within the network so that customers can get it when they want. When the customer places an order. Our AI system predicts how long it will take based on several factors including how many associates and drivers are available but the distance to the home in real time."
  6. Customer experiences offline and online need to be connected. Kumar said the transformation of the supply chain is ongoing and critical to Walmart's vision of adaptive retail. "It's only going to get better with the connected supply chain," said Kumar.

From our underwriter

Supply chain disruption — from human and technology errors to weather and other crises — has always been a challenge. Manufacturers are using dashboards to visualize performance measurements, preventative maintenance and process optimization. But while they are a critical starting point, supply chain dashboards are not enough. Today’s manufacturers need a way to move from data collection through the supply chain to decision support and actionable insights. Get the full story.

Data to Decisions Matrix Commerce Sales Marketing Revenue & Growth Effectiveness Next-Generation Customer Experience Future of Work Innovation & Product-led Growth New C-Suite Tech Optimization Supply Chain Automation Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software IoT Blockchain ERP Leadership Collaboration M&A Chief Information Officer Chief Procurement Officer Chief Supply Chain Officer Chief Sustainability Officer Chief Experience Officer

Supermicro ups Q2 outlook as AI server demand heats up

Supermicro said its second quarter results will be stronger than expected as demand for its AI, cloud and storage rack systems surged.

The company said that its second quarter revenue will be between $3.6 billion and $3.65 billion compared to its previous guidance of $2.7 billion to $2.9 billion. Earnings will be $4.90 to $5.05 a share on a GAAP basis compared to previous guidance of $3.75 a share to $4.24 a share. Non-GAAP earnings for the second quarter will be $5.40 to $5.55 a share.

Supermicro didn't detail what was driving its better-than-expected results, but the likely reason is demand for generative AI systems. Supermicro said its liquid cooled rack systems were seeing strong demand. In addition, AMD has launched its GPUs for generative AI workloads to compete with Nvidia. All of those developments likely add up to strong AI server sales for Supermicro.

The server maker’s outlook indicates that the booming demand seen by Nvidia is not filtering down to server companies. Both Dell Technologies and HPE have signaled strong pipelines for generative AI systems.

Speaking at a recent investment conference in December, CFO David Weigand said the slate of emerging technologies for AI workloads and next-generation data centers is playing into Supermicro's strengths. Weigand said:

"Super Micro's strength is the fact that we are approximately 50% engineers. We are an engineering company, and that fact is helping us now because technological development has sped up. You have multiple platforms with Nvidia, AMD and Intel, ARM solutions, as well as others. There are a lot of emerging technologies. And this is really playing into Supermicro's model, which is our building block solutions that we architected the server technology from the ground up."

Weigand added that Supermicro's in-house designed servers enable it to integrate new technologies for custom workloads. "We design something unique for a customer that not only gives them the very best cost performance metric, but it also gives the lowest total cost of ownership, because we have designed our servers to be low power consumption and to manage heat," said Weigand. "With GPUs going past 1,000 watts and dissipating a thousand watts of heat and CPUs dissipating in excess of 500 watts of heat, heat is becoming more important. Our ability to provide liquid cooled solutions and what we call green computing or the lowest amount of heat dissipation allows us to have a competitive edge."

Data to Decisions Tech Optimization Innovation & Product-led Growth Future of Work Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity Big Data AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer Chief Executive Officer Chief AI Officer Chief Analytics Officer Chief Product Officer

Arizona State, OpenAI to collaborate on ChatGPT education use cases

OpenAI has inked its first partnership with Arizona State University in an effort to bring ChatGPT Enterprise to courses, tutoring and research.

The partnership is notable since ASU is planning to build a personalized AI tutor for students. Generative AI in education has been a hot topic as some universities have moved to ban its use. Other institutions have embraced generative AI. Meanwhile, students are using tools like ChatGPT and educational services like Chegg and Khan Academy have already partnered with OpenAI, which just launched its GPT Store

Previously: Why Chegg is using Scale AI to develop proprietary LLMsEducation gets schooled in generative AI | Coursera: Generative AI will lead to reskilling, upskilling boom

In addition, the education technology stack is looking to embed generative AI.

According to ASU, the plan is to begin use of ChatGPT Enterprise with faculty and staff. ASU said it is focusing on enhancing student success, finding new research avenues and streamlining processes. ASU created an AI accelerator within its enterprise technology department last year. 

On the privacy front, ASU said it will safeguard user data. ASU CIO Lev Gonick said in a statement:

"The goal is to leverage our knowledge core here at ASU to develop AI-driven projects aimed at revolutionizing educational techniques, aiding scholarly research and boosting administrative efficiency."

Data to Decisions Innovation & Product-led Growth Future of Work Tech Optimization Next-Generation Customer Experience Digital Safety, Privacy & Cybersecurity AI GenerativeAI ML Machine Learning LLMs Agentic AI Analytics Automation Disruptive Technology Chief Information Officer Chief Executive Officer Chief Technology Officer Chief AI Officer Chief Data Officer Chief Analytics Officer Chief Information Security Officer Chief Product Officer

Samsung's Galaxy S24 launch becomes showcase for Google Cloud AI

Smartphones are increasingly about foundational models, generative AI features and the ability to leverage AI locally. The latest example is Samsung's Galaxy S24 launch, which also served as a showcase for Google's Gemini Pro and Imagen 2 on Vertex AI.

The consumer electronics giant unveiled the Galaxy S24 Ultra, Galaxy S24+ and Galaxy S24 and touted Galaxy AI experiences. Features included Interpreter, which can translate live conversations, Chat Assist, to ensure communication comes off well, Note Assist, which will feature AI-generated summaries, and other features baked into the camera.

With the Samsung launch, two of the primary Android flagship devices will come equipped with generative AI experiences. If you've been following any of the recent hardware launches the next battle is on device model processing. The Google Pixel 8 Pro is designed to show off Google’s models and AI processing. Amazon's Alexa event also had a heavy LLM spin, Apple touched on running AI and machine learning models locally and PC makers are betting (more like praying) that there will be an upgrade cycle due to model training. Microsoft's Surface event was really about a barrage of Microsoft 365 Copilot launches. Samsung said it will be the first Google Cloud partner to deploy Gemini Pro and Imagen 2 on Vertex AI via the cloud to smartphone devices.

According to Samsung, the Galaxy S24 Ultra will be equipped with Snapdragon 8 Gen 3 Mobile Platform for Galaxy, an optimized chipset for AI processing. The Galaxy S24 starts at $799.99 for Galaxy S24, the $999.99 Galaxy S24 Plus, and $1,299.99 Galaxy S24 Ultra. All devices have AI features. 

Janghyun Yoon, Corporate EVP and Head of Software Office of Mobile Experience Business at Samsung Electronics, said Google Cloud and Samsung teams worked together on the Galaxy S24 launch and conducted "months of rigorous testing and competitive evaluation."

While Samsung and Google touted consumer features on the Galaxy S24, the long-term takeaway for enterprises is that they'll eventually be able to leverage the processing power in smartphones for generative AI applications. Local AI processing is more secure, efficient and cost effective.

Bottom line: Smartphones are going to compete on generative AI. Smart enterprises will figure out ways to use local processing for personalized individual use cases.

Constellation Research's take

Constellation Research analyst Holger Mueller said:

"The interesting thing on Galaxy S24 is how many AI features are Google's including Circle to Search were seamless. Google's ability to push capabilties on Android seems to finally be working beyond Pixel devices to Samsung flagship smartphones. Samsung committed to 7 years of support with Google on the backend is definitely an inflection point."

Andy Thurai at Constellation Research added:

"The Samsung-Google and Google Cloud partnership is a win-win for both companies. Google's partnership with Samsung allows them to take Apple head-on. It will be interesting to see what Microsoft will do. Samsung will also deploy Gemini Nano, an LLM that is purpose-built for mobile devices. Because Samsung uses Android as its OS, this partnership and technology alliance was fairly easy.

While Microsoft/Azure is trying to capture the search market that Google owns with its AI advancements, Google is trying to go after the mobile market. Gemini Pro and Imagen 2 on Samsung Galaxy S24 will certainly challenge iPhone.

I wouldn't be surprised if Apple and Microsoft explore an alliance on mobile as they both need each other. While Apple has done some AI-related things such as facial ID unlock with facial recognition, A15 bionic AI chip, and some basic Siri auto-correct and photo editing, etc., the walled garden of Apple's ecosystem hasn't done much on the AI front."

Next-Generation Customer Experience Innovation & Product-led Growth Data to Decisions Future of Work Digital Safety, Privacy & Cybersecurity Tech Optimization Google Google Cloud SaaS PaaS IaaS Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain CRM ERP CCaaS UCaaS Collaboration Enterprise Service ML Machine Learning LLMs Agentic AI Generative AI AI Analytics Automation business Marketing finance Healthcare Customer Service Content Management GenerativeAI Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer Chief Executive Officer Chief AI Officer Chief Analytics Officer Chief Product Officer

Quantinuum raises $300 million, valued at $5 billion

Quantinuum raised $300 million in equity investment putting the quantum computing company's valuation at $5 billion.

Honeywell merged its quantum unit with Cambridge Quantum Computing in 2021 and launched Quantinuum as a stand-alone company. Honeywell remains Quantinuum's largest shareholder.

In a statement, Honeywell said that the funding round was led by JPMorgan Chase with participation from Mitsui & Co., Amgen and Honeywell. Mitsui said it will help expand Quantinuum's reach in Asia.

Quantinuum, which is focusing on quantum use cases such as cybersecurity, computational chemistry and simulation, has raised $625 million since inception. 

According to Honeywell, Quantinuum will use the funds to "accelerate the path towards achieving the world's first universal fault-tolerant quantum computers, while also extending Quantinuum's software offering to enhance commercial applicability." Quantinuum is also working to develop Quantum Natural Language Processing, an effort to bridge quantum computing and generative AI.

Quantinuum counts JPMorgan Chase as one of its customers using Quantinuum's H-Series quantum processors and the company's software development kit, TKET. Other Quantinuum customers include Airbus, BMW Group, Honeywell, HSBC, Mitsui and Thales.

Here's a look at Quantinuum's development so far and roadmap ahead. 

Data to Decisions Innovation & Product-led Growth Tech Optimization Quantum Computing Chief Information Officer Chief Technology Officer

Hitachi Vantara names NetApp alum Tanase as chief product officer

Hitachi Vantara has named Octavian Tanase as chief product officer effectively immediately. Tanase will report to Hitachi Vantara CEO Sheila Rohra.

Tanase was most recently senior vice president of hybrid cloud engineering at NetApp where he integrated the company’s software portfolio with offerings from AWS, Microsoft Azure and Google Cloud.

At Hitachi Vantara, a Constellation Insights underwriter, Tanase will oversee the storage, infrastructure and hybrid cloud management company's product vision, strategy, development and execution.

In a statement, Rohra said the addition of Tanase will help Hitachi Vantara position for generative AI and the "explosive growth of processing required for data." Tanase said his goal "is to help the company expand its leadership position by harnessing the power of generative AI and other emerging technologies to drive even greater innovation in its portfolio."

The addition of Tanase comes just a few weeks after Hitachi Vantara, a subsidiary of Hitachi, appointed Tony Gonnella as CFO. Gonnella had been CFO at Cortex and Unit 42 at Palo Alto Networks.

On the product front, Hitachi Vantara recently launched Hitachi Unified Compute Platform (UCP) for GKE Enterprise to manage hybrid cloud operations. UCP is delivered via Google Distributed Cloud Virtual and can distribute data workloads between Google Cloud Anthos and in-house infrastructure securely.

Data to Decisions Tech Optimization Digital Safety, Privacy & Cybersecurity Innovation & Product-led Growth Future of Work SaaS PaaS IaaS Cloud Digital Transformation Disruptive Technology Enterprise IT Enterprise Acceleration Enterprise Software Next Gen Apps IoT Blockchain CRM ERP CCaaS UCaaS Collaboration Enterprise Service Chief Information Officer Chief Technology Officer Chief Information Security Officer Chief Data Officer