#ai-workloads

[ follow ]
#nvidia

HPE crams 224 Nvidia Blackwell GPUs into latest Cray EX

HPE Cray's EX154n systems will drastically increase GPU capacity, supporting up to 224 GPUs and 8,064 CPU cores per cabinet.

Nvidia reportedly forms unit to peddle IP to cloud providers

Nvidia is creating a business unit to sell its intellectual property and design services to cloud providers.
Cloud providers like AWS, Microsoft, and Meta have been developing their own custom silicon alternatives to Nvidia's GPUs for AI workloads.

HPE crams 224 Nvidia Blackwell GPUs into latest Cray EX

HPE Cray's EX154n systems will drastically increase GPU capacity, supporting up to 224 GPUs and 8,064 CPU cores per cabinet.

Nvidia reportedly forms unit to peddle IP to cloud providers

Nvidia is creating a business unit to sell its intellectual property and design services to cloud providers.
Cloud providers like AWS, Microsoft, and Meta have been developing their own custom silicon alternatives to Nvidia's GPUs for AI workloads.
morenvidia
#edge-computing

Windows Server IoT 2025 is all about that speed

Windows Server IoT 2025 enhances performance and speed for edge computing, supporting AI workloads and improving peripheral management capabilities.

Red Hat Device Edge 4.17 processes workloads in near real-time

Red Hat's Device Edge platform upgrade improves processing speed for time-sensitive workloads, enabling real-time responses necessary for critical applications.

Windows Server IoT 2025 is all about that speed

Windows Server IoT 2025 enhances performance and speed for edge computing, supporting AI workloads and improving peripheral management capabilities.

Red Hat Device Edge 4.17 processes workloads in near real-time

Red Hat's Device Edge platform upgrade improves processing speed for time-sensitive workloads, enabling real-time responses necessary for critical applications.
moreedge-computing

Cloud providers can hardly keep up with dazzling demand for AI

The demand for cloud capacity due to AI workloads is overwhelming major cloud providers.
Investor patience is running thin amid high expectations for cloud and AI growth.
#containerization

Okteto 3.0 brings remote deployments, faster build times, and deeper insights to platform engineering - Amazic

Okteto focuses on enhancing deployment velocity through innovative solutions.
The startup's developments include remote deployments and smart builds for efficiency.

Kubernetes at 10: The long road to mastery of persistent storage | Computer Weekly

Kubernetes celebrates its 10th birthday as the market-leading container orchestration platform.

Okteto 3.0 brings remote deployments, faster build times, and deeper insights to platform engineering - Amazic

Okteto focuses on enhancing deployment velocity through innovative solutions.
The startup's developments include remote deployments and smart builds for efficiency.

Kubernetes at 10: The long road to mastery of persistent storage | Computer Weekly

Kubernetes celebrates its 10th birthday as the market-leading container orchestration platform.
morecontainerization
#amd

Intel, AMD unite in new x86 alliance to tackle AI, other challenges

Intel and AMD's x86 Ecosystem Advisory Group aims to tackle AI workload challenges and enhance architecture interoperability among semiconductor rivals.

AMD's next generation of AI laptop processors have a new name too

AMD announces Ryzen AI 300 series for AI workloads, featuring improved architectures and high performance chips like Ryzen AI 9 HX 370.

Intel, AMD unite in new x86 alliance to tackle AI, other challenges

Intel and AMD's x86 Ecosystem Advisory Group aims to tackle AI workload challenges and enhance architecture interoperability among semiconductor rivals.

AMD's next generation of AI laptop processors have a new name too

AMD announces Ryzen AI 300 series for AI workloads, featuring improved architectures and high performance chips like Ryzen AI 9 HX 370.
moreamd
#artificial-intelligence

Interview: Nvidia on AI workloads and their impacts on data storage | Computer Weekly

Understanding the quality and relevance of data is crucial for successful AI projects.

Raspberry Pi gets AI vision superpowers

Raspberry Pi's new AI camera designed for advanced workloads establishes its role in the AI sector.

Do you really need that GPU or NPU for your AI apps?

Artificial intelligence (AI) technology integration is pervasive, from office software to smartphones, raising discussions on the need for separate accelerators like GPUs or NPUs.

Pure deepens Fusion as reorientation to storage for AI continues | Computer Weekly

Pure Storage aims to upgrade its Fusion control plane to provide easily provisioned storage capacity for AI workloads.

Storage technology explained: AI and data storage | Computer Weekly

AI and ML have a wide range of applications from simple chatbots to complex content generation.
Storage plays a crucial role in AI by providing data for training and storing large volumes of generated data.

Interview: Nvidia on AI workloads and their impacts on data storage | Computer Weekly

Understanding the quality and relevance of data is crucial for successful AI projects.

Raspberry Pi gets AI vision superpowers

Raspberry Pi's new AI camera designed for advanced workloads establishes its role in the AI sector.

Do you really need that GPU or NPU for your AI apps?

Artificial intelligence (AI) technology integration is pervasive, from office software to smartphones, raising discussions on the need for separate accelerators like GPUs or NPUs.

Pure deepens Fusion as reorientation to storage for AI continues | Computer Weekly

Pure Storage aims to upgrade its Fusion control plane to provide easily provisioned storage capacity for AI workloads.

Storage technology explained: AI and data storage | Computer Weekly

AI and ML have a wide range of applications from simple chatbots to complex content generation.
Storage plays a crucial role in AI by providing data for training and storing large volumes of generated data.
moreartificial-intelligence
#kubernetes

Edera is building a better Kubernetes and AI security solution from the ground up | TechCrunch

Edera aims to enhance security for Kubernetes containers and AI workloads with a new hypervisor, addressing longstanding challenges in multi-tenancy.

Rafay Systems: There's more to the cloud than Kubernetes & we're out to simplify it all - Amazic

Rafay Systems simplifies Kubernetes and expands to AI and non-Kubernetes workloads.

Cern: Challenges of GPU datacentre management | Computer Weekly

Cern awarded CNCF Top End User Award
Cern exploring GPUs for AI workloads
Kubernetes Scheduler supports GPU sharing

Edera is building a better Kubernetes and AI security solution from the ground up | TechCrunch

Edera aims to enhance security for Kubernetes containers and AI workloads with a new hypervisor, addressing longstanding challenges in multi-tenancy.

Rafay Systems: There's more to the cloud than Kubernetes & we're out to simplify it all - Amazic

Rafay Systems simplifies Kubernetes and expands to AI and non-Kubernetes workloads.

Cern: Challenges of GPU datacentre management | Computer Weekly

Cern awarded CNCF Top End User Award
Cern exploring GPUs for AI workloads
Kubernetes Scheduler supports GPU sharing
morekubernetes
#data-management

Google Cloud expands its database portfolio with new AI capabilities | TechCrunch

Google enhancing databases for AI workloads during Cloud Next conference in Tokyo, focusing on Spanner, Gemini-powered features, and data management for generative AI success.

Oracle Exadata Exascale: Big architecture shift for small workloads | Computer Weekly

Oracle's Exadata Database Service on Exascale aims to enhance performance and cost-efficiency with AI workloads and a shared architecture.

Google Cloud expands its database portfolio with new AI capabilities | TechCrunch

Google enhancing databases for AI workloads during Cloud Next conference in Tokyo, focusing on Spanner, Gemini-powered features, and data management for generative AI success.

Oracle Exadata Exascale: Big architecture shift for small workloads | Computer Weekly

Oracle's Exadata Database Service on Exascale aims to enhance performance and cost-efficiency with AI workloads and a shared architecture.
moredata-management

Tape is so dead, 152.9 exabytes worth of LTO shipped in 2023

Tape storage is not dead despite technological advancements and continues to be relevant in modern data storage strategies.

Microsoft's new Copilot+ PC brand heralds a "new era" in personal computing - here's why we're excited

Microsoft introduces Copilot+ PCs designed for AI workloads, featuring Neural Processing Units and Qualcomm Snapdragon chips for enhanced performance and AI capabilities.

Hosting 101: The do's and don'ts

Small businesses can benefit from dedicated hosting by making the right choices from the start and optimizing workloads for better performance.

NetApp upgrades AFF all-flash as it targets AI storage | Computer Weekly

NetApp refreshes AFF all-flash storage arrays targeting AI workloads and energy efficiency.

Tesla wants to monetize its cars to process AI workloads

Tesla considering using vehicle compute power for cash generation.

AI PCs are here but a killer application for biz users? Nope

Forrester Research emphasizes the lack of a 'killer app' making AI PCs essential for business users.

LANL powers up Nvidia's GH200-packed Venado super

Venado supercomputer focuses on AI workloads with impressive exaFLOPS performance using lower precision calculations.

Ethernet advances will end InfiniBand's lead in AI nets

Ethernet is set to become a better alternative for AI workloads with three imminent improvements predicted by Gartner.
By 2028, it is estimated that 45 percent of Gen AI workloads will run on Ethernet, up from less than 20 percent currently.

AMD and Microsoft cement relationship with cloud collaborations

Azure customers can run high intensity or AI workloads on powerful infrastructure
Customers don't need to house or maintain the infrastructure themselves
#AI workloads

Memory prices to rise next year, Gartner forecasts

The semiconductor market is expected to return to growth in 2024, driven by increasing demand for AI workloads and memory components.
Gartner estimates that global semiconductor revenues will rise 16.8% in 2024, following a contraction in sales for 2023.

Microsoft launches custom chips to accelerate its plans for AI domination

Microsoft announced two custom chips for accelerating AI workloads in its Azure cloud computing service.
Maia is designed for large language models like GPT-3.5 Turbo and GPT-4, while Cobalt is a CPU for conventional tasks.
Microsoft plans to use these chips internally and not sell them.

Microsoft launches custom chips to accelerate its plans for AI domination

Microsoft announced two custom chips for accelerating AI workloads in its Azure cloud computing service.
Maia is designed for large language models like GPT-3.5 Turbo and GPT-4, while Cobalt is a CPU for conventional tasks.
Microsoft plans to use these chips internally and not sell them.

Microsoft Announces New Maia 100 and Cobalt 100 Chips

Microsoft will release two custom chips next year: the Maia 100 designed for AI workloads and the Cobalt 100 CPU for general compute workloads on Microsoft cloud.
The chips are built in-house by Microsoft, allowing for customization of the entire infrastructure stack to maximize performance.
Microsoft has developed custom server racks with liquid cooling to accommodate the Maia 100 AI Accelerator.

Microsoft Announces New Maia 100 and Cobalt 100 Chips

Microsoft will release two custom chips next year: the Maia 100 designed for AI workloads and the Cobalt 100 CPU for general compute workloads on Microsoft cloud.
The chips are built in-house by Microsoft, allowing for customization of the entire infrastructure stack to maximize performance.
Microsoft has developed custom server racks with liquid cooling to accommodate the Maia 100 AI Accelerator.

Making sense of Nvidia's SuperNIC

Nvidia has introduced a new networking accelerator called SuperNIC, designed to boost AI workloads in Ethernet-based networks.
SuperNIC offers features such as high-speed packet reordering, advanced congestion control, programmable I/O pathing, and integration with Nvidia's hardware and software portfolio.
SuperNIC is not a rebrand of Nvidia's previous DPU, but a separate product designed to work with Nvidia's Spectrum-X offering.

Memory prices to rise next year, Gartner forecasts

The semiconductor market is expected to return to growth in 2024, driven by increasing demand for AI workloads and memory components.
Gartner estimates that global semiconductor revenues will rise 16.8% in 2024, following a contraction in sales for 2023.

Microsoft launches custom chips to accelerate its plans for AI domination

Microsoft announced two custom chips for accelerating AI workloads in its Azure cloud computing service.
Maia is designed for large language models like GPT-3.5 Turbo and GPT-4, while Cobalt is a CPU for conventional tasks.
Microsoft plans to use these chips internally and not sell them.

Microsoft launches custom chips to accelerate its plans for AI domination

Microsoft announced two custom chips for accelerating AI workloads in its Azure cloud computing service.
Maia is designed for large language models like GPT-3.5 Turbo and GPT-4, while Cobalt is a CPU for conventional tasks.
Microsoft plans to use these chips internally and not sell them.

Microsoft Announces New Maia 100 and Cobalt 100 Chips

Microsoft will release two custom chips next year: the Maia 100 designed for AI workloads and the Cobalt 100 CPU for general compute workloads on Microsoft cloud.
The chips are built in-house by Microsoft, allowing for customization of the entire infrastructure stack to maximize performance.
Microsoft has developed custom server racks with liquid cooling to accommodate the Maia 100 AI Accelerator.

Microsoft Announces New Maia 100 and Cobalt 100 Chips

Microsoft will release two custom chips next year: the Maia 100 designed for AI workloads and the Cobalt 100 CPU for general compute workloads on Microsoft cloud.
The chips are built in-house by Microsoft, allowing for customization of the entire infrastructure stack to maximize performance.
Microsoft has developed custom server racks with liquid cooling to accommodate the Maia 100 AI Accelerator.

Making sense of Nvidia's SuperNIC

Nvidia has introduced a new networking accelerator called SuperNIC, designed to boost AI workloads in Ethernet-based networks.
SuperNIC offers features such as high-speed packet reordering, advanced congestion control, programmable I/O pathing, and integration with Nvidia's hardware and software portfolio.
SuperNIC is not a rebrand of Nvidia's previous DPU, but a separate product designed to work with Nvidia's Spectrum-X offering.
moreAI workloads

Interview: Nvidia on AI workload demands and storage performance | Computer Weekly

AI workloads have varying I/O profiles impacting storage differently.

HP shows the AI PC ecosystem is starting to bear fruit - is it time for businesses to take note?

AI-ready PCs with dedicated hardware not likely in mainstream by 2024, but software advancements make them worth considering for added value.
#data-storage

Podcast: Storage for AI needs scale, hybrid cloud and multiple integrations | Computer Weekly

AI workloads have different performance and capacity requirements compared to traditional enterprise workloads, needing to handle large datasets from gigabytes to petabytes and utilizing multiple cloud platforms.
Importance of storage scalability, performance, API integration, and data wrangling capabilities in handling AI workloads across hybrid cloud environments.

Pure CEO: AI needs write speed and storage in place | Computer Weekly

Storage write speed is key for checkpointing in AI workloads.

Podcast: Storage for AI needs scale, hybrid cloud and multiple integrations | Computer Weekly

AI workloads have different performance and capacity requirements compared to traditional enterprise workloads, needing to handle large datasets from gigabytes to petabytes and utilizing multiple cloud platforms.
Importance of storage scalability, performance, API integration, and data wrangling capabilities in handling AI workloads across hybrid cloud environments.

Pure CEO: AI needs write speed and storage in place | Computer Weekly

Storage write speed is key for checkpointing in AI workloads.
moredata-storage
[ Load more ]