Meta launches latest chip for AI workloads
The second version of the Meta Training and Inference Accelerator (MTIA) highlights how cloud hyperscale players are creating their own processors for large language model (LLM) training and inferencing.
Five years ago, Chief Information Officers (CIOs) were on top of the world. These executives played mission-critical roles in driving multi-million dollar projects that delivered massive change. However, a global recession and the inability of CIOs to deliver on business value have tarnished their status. Today's CIOs are under pressure to deliver on requests for innovation, cost reduction, connectivity, and a growing demand for business intelligence. Just as previous technology and business shifts have changed the role of the CIO, the new, more consumer-oriented business models of the social revolution will favor a new breed of business and technology leader. These leaders will have to navigate myriad converging and disruptive technologies, align new initiatives to both business value and technology feasibility, and identify strategies to leverage existing investments to fund innovation. Constellation's research and advisory offerings arm the CIO w/ the knowledge, best practices, and strategies required to manage the four personas of the next generation CIO.
The second version of the Meta Training and Inference Accelerator (MTIA) highlights how cloud hyperscale players are creating their own processors for large language model (LLM) training and inferencing.
Google Cloud pitched an agent-oriented vision for generative AI at Google Cloud Next and highlighted a bevy of emerging use cases going from pilot to production.
The chipmaker's Gaudi 3 launch, announced at the Intel Vision conference, is the linchpin of Intel's plans to garner AI training and inference workloads and take share from Nvidia.
MongoDB expanded integrations with Google Cloud's Vertex AI, BigQuery, Google Distributed Cloud and Google Cloud Manufacturing Data Engine.
Google Cloud outlined a series of services and enhancements across its platform in a bid to make it easier for enterprises to bring their data to generative AI models, build applications and deploy them at scale.
JPMorgan Chase CEO Jamie Dimon issued his annual shareholder letter and provided an incremental update on the company's artificial intelligence efforts as well as private cloud buildout.
Two questions have haunted me for two decades: first, can we really address security without addressing networks? Second, are observability and security like oil and water? We are seeing a convergence of network, security, and observability—fueled by AI.
Enterprises need to focus on data lakehouse strategies in 2024 to properly take advantage of generative AI; model architecture will be critical to managing large and small models; fine tuning is more difficult than you'd think; and CXOs were weary of database vendors glomming on to genAI hype.
It's easy to conclude that generative AI is going to take jobs from humans. But there's another argument that genAI will be needed just to maintain and improve productivity levels because there will be fewer workers. There’s a demographic donut hole in the workforce that may be partially ameliorated by genAI.
Wipro has named Srini Pallia CEO effective immediately replacing Thierry Delaporte, who stepped down to pursue other interests.
Archetype AI has raised $13 million in seed funding and launched Newton, a foundational model that is built to understand the physical world via data signals from accelerometers, gyroscopes, radars, cameras, microphones, thermometers and other environmental sensors.
With the move, DataStax aims to create an integrated generative AI stack to create applications. DataStax will integrate Langflow with its DataStax Astra DB and Python libraries.