Topic / Trend Rising

AI Hardware & Infrastructure Evolution

The relentless demand for AI is driving massive investments in specialized chips (TPUs, HBM, custom silicio), data centers, and advanced cooling. There's a strong push for on-premise and edge AI solutions to manage costs and data sovereignty.

Detected: 2026-04-27 · Updated: 2026-04-27

Related Coverage

2026-04-27 DigiTimes

OpenAI Reportedly Taps Apple Suppliers for Hardware Expansion

OpenAI is reportedly exploring collaborations with key Apple suppliers, including MediaTek, Qualcomm, and Luxshare, to bolster its hardware initiatives. The rumor, reported by analyst Ming-Chi Kuo, suggests a strategic expansion in the AI infrastruct...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-27 DigiTimes

HCL-Foxconn and CTCI: A New OSAT Facility in India for the Chip Supply Chain

The joint venture between HCL and Foxconn has selected Taiwan's CTCI to build an OSAT (Outsourced Semiconductor Assembly and Test) facility in India. This strategic initiative aims to strengthen semiconductor manufacturing capacity in the subcontinen...

#Hardware #LLM On-Premise #DevOps
2026-04-27 DigiTimes

Taiwan's Supply Chain: An Irreplaceable Pillar for AI Hardware

The surging demand for AI hardware highlights the centrality of Taiwan's supply chain. This strategic dependence poses crucial considerations for companies planning on-premise LLM deployments, impacting availability, TCO, and data sovereignty. Unders...

#Hardware #LLM On-Premise #DevOps
2026-04-27 DigiTimes

AI in Smart Cockpits: The Challenge of Real Value and Edge Deployment

Integrating artificial intelligence into smart cockpits represents one of the next major technological challenges. The central question is not merely technical feasibility, but AI's ability to generate tangible and measurable value. This involves cri...

#Hardware #LLM On-Premise #DevOps
2026-04-27 DigiTimes

Naver Cloud and HanmiGlobal: Global Data Center Expansion for AI

Naver Cloud and HanmiGlobal have announced a joint global expansion of their data centers. This strategic move is set against the backdrop of the escalating competition for AI infrastructure, highlighting the need for dedicated computational resource...

#Hardware #LLM On-Premise #DevOps
2026-04-27 DigiTimes

Nio ventures into chipmaking to reduce reliance on Nvidia

Electric vehicle manufacturer Nio is investing in proprietary chip production, a strategic move aimed at reducing its reliance on external suppliers like Nvidia. This decision reflects a growing trend among companies to seek greater control over thei...

#Hardware #LLM On-Premise #DevOps
2026-04-26 DigiTimes

Wi-Fi 7 Accelerates Networking: Impact on Infrastructure and On-Premise AI

Taiwanese networking firms anticipate significant growth in Q2, driven by Wi-Fi 7 adoption. This technological evolution, with its promises of higher throughput and lower latency, is crucial for modern enterprise infrastructures. While not directly t...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-26 DigiTimes

Taiwan's Growing AI and Industrial Demand Reshapes Corporate Strategies

Taiwan is at the forefront of escalating demand for artificial intelligence and industrial capacity. This pressure is compelling companies to re-evaluate and reshape their operational and infrastructural strategies, with significant implications for ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-26 The Next Web

Sequoia and Mac Minis: Boosting On-Premise AI Beyond Investment

Sequoia Capital distributed 200 custom Mac Minis to attendees of its "AI at the Frontier" event. The initiative, led by Alfred Lin, a co-steward at Sequoia, aims to foster AI projects that fall outside traditional investment models, promoting local d...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-26 Tom's Hardware

AI Wave Propels Taiwan: Record Market Cap and TSMC's Role

Taiwan's stock market has surpassed the UK's, despite a significantly smaller economy. This surge is driven by the artificial intelligence boom, with TSMC alone contributing over 40% of Taiwan's total market value, highlighting the centrality of adva...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-26 DigiTimes

The HBM Competition: Samsung, Nvidia, and TSMC Vie for the Future of AI

The High Bandwidth Memory (HBM) market is at the heart of growing competition among tech giants. Samsung is leveraging its production capacity to secure crucial orders from Nvidia for its AI accelerators, while TSMC intensifies its pushback. This mar...

#Hardware #LLM On-Premise #DevOps
2026-04-26 DigiTimes

BizLink and Optical Interconnects: CPO Timing Uncertainties for AI

BizLink is intensifying its focus on optical interconnects, crucial components for high-performance AI infrastructure. However, the company notes uncertainties regarding the widespread adoption timeline for Co-Packaged Optics (CPO), a technology pois...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-26 DigiTimes

AI Data Centers Face Interconnect Limits, Boosting Optical Module Demand

The increasing complexity and computational demands of AI workloads, particularly for Large Language Models, are pushing data centers to the limits of their interconnection capabilities. This scenario is driving a surge in demand for optical modules,...

#Hardware #LLM On-Premise #DevOps
2026-04-25 TechCrunch AI

Maine Governor Vetoes Data Center Moratorium

Maine's governor rejected L.D. 307, a bill that would have established the first statewide moratorium in the United States on new data center construction until November 1, 2027. This decision has significant implications for the expansion of digital...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-25 The Next Web

Oracle Secures $16.3 Billion Bond for Mega Data Center

Oracle has secured $16.3 billion in financing for a single data center campus in Michigan, marking the largest single-facility technology debt package ever assembled. PIMCO anchored approximately $10 billion of the bond after US banks withdrew from t...

#Hardware #LLM On-Premise #DevOps
2026-04-25 Tom's Hardware

Maine Governor Vetoes Bill Banning Large New Data Centers, Citing Key Project

The governor of Maine has vetoed a legislative proposal aimed at prohibiting the construction of large new data centers within the state. The decision was driven by the belief that the bill should have included an exemption for a specific, well-suppo...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-25 The Next Web

From Web Intelligence to AI: Emerging Infrastructure Challenges

The web intelligence industry, a cornerstone of data-driven development, faces new infrastructure challenges. With the exponential growth of big data and advancements in AI, ensuring sustained data flow and adequate infrastructure becomes crucial to ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-25 The Next Web

The AI Skills Gap: A Challenge for On-Premise Deployment

Denis Brovarnyy highlights a growing gap between theoretical training and the practical skills required in the tech sector. As AI transitions from experimentation to enterprise implementation, ignoring this gap becomes costly. Companies urgently need...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-25 DigiTimes

Taiwan's Industrial Production Surges Driven by AI Infrastructure Demand

Taiwan's industrial production is experiencing significant growth, fueled by robust global demand for artificial intelligence infrastructure. This trend underscores the increasing need for specialized hardware to support the development and deploymen...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 The Next Web

US Government's Intel Stake Jumps to $36 Billion: The Impact of the CHIPS Act

The United States government has seen the value of its 9.9% stake in Intel surge to approximately $36 billion. Acquired for $8.9 billion last August by converting CHIPS Act and Secure Enclave funds into equity at $20.47 per share, the stake has gener...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 The Register AI

Ubuntu 26.04 LTS: Enhanced GPGPU and AI Tooling, Xorg Departs Desktop

Canonical has released Ubuntu 26.04 "Resolute Raccoon," its new LTS version. This release integrates GNOME 50 and Linux kernel 7.0, removing Xorg from the desktop installation in favor of Xwayland for X11 applications. A significant focus is placed o...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 TechCrunch AI

Sold-out Mac minis flood eBay as local AI demand surges

The increasing demand for running local AI models and tools has led to Apple's Mac mini being sold out. This shortage has triggered a secondary market on eBay, where devices are offered at marked-up prices, highlighting the interest in compact deskto...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 Tom's Hardware

NEO Semiconductor: 3D X-DRAM Validated, an HBM Alternative for AI Processors

NEO Semiconductor has validated the proof-of-concept for its 3D X-DRAM, an innovative memory technology for AI processors. The company secured funding to further develop this solution, which positions itself as a high-performance alternative to HBM. ...

#Hardware #LLM On-Premise #DevOps
2026-04-24 TechCrunch AI

Nothing introduces an AI-powered on-device dictation tool

Nothing has launched a new AI-powered dictation tool, designed to operate directly on the device. The solution supports over one hundred languages, highlighting an approach that prioritizes local processing over the cloud for AI functionalities, with...

#Hardware #LLM On-Premise #DevOps
2026-04-24 ServeTheHome

Meta Secures Millions of AWS Graviton Cores for AI

Meta has announced the acquisition of tens of millions of AWS Graviton CPU cores based on Arm architecture. This strategic move aims to bolster its "agentic AI" compute portfolio, highlighting the increasing demand for dedicated hardware resources to...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 Tom's Hardware

SoftBank and Intel Develop ZAM, a Low-Power Memory for AI

A SoftBank subsidiary, in collaboration with Intel, is developing ZAM, a new memory technology designed for AI workloads. The goal is to offer a lower-power alternative to current HBM memories. The project has received financial support from the Japa...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

Google Specializes TPU Chips for AI Training and Inference

Google has announced the specialization of its TPU chips, distinguishing versions optimized for AI model training and inference. This move reflects a growing industry trend towards dedicated AI infrastructures, with significant implications for on-pr...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 DigiTimes

Liteon Focuses on Edge AI for Strategic Reset by 2026

Liteon is reorienting its startup platform towards the Edge AI ecosystem, aiming for a significant growth reset by 2026. This strategic move underscores the increasing importance of locally distributed artificial intelligence solutions, addressing ne...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

SMIC Re-enters Advanced Packaging to Bolster AI Chip Strategy

SMIC is strengthening its AI chip strategy by re-entering the advanced packaging sector and expanding its team. This move underscores the growing importance of advanced integration technologies for the performance of AI-dedicated processors, a critic...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 DigiTimes

Google Ramps Up TPU Server Deployment: Impact on AI Supply Chain

Google is accelerating the deployment of new Tensor Processing Unit (TPU)-based servers, a move that is strengthening the position of Taiwanese suppliers in the supply chain. This development underscores the growing demand for specialized AI hardware...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

Taiwan Boosts Local AI Infrastructure: Foxconn Submits Proposal

Taiwan's Ministry of Digital Affairs (MODA) is encouraging local firms to invest in AI computing power. In this context, Foxconn has already submitted a project proposal, signaling a growing strategic interest in developing robust, localized AI infra...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 DigiTimes

Texas Instruments: Edge AI Opportunities Extend Beyond Robotics

Texas Instruments highlights how artificial intelligence opportunities at the edge (Edge AI) are growing well beyond traditional robotic applications. This expansion opens new scenarios for companies seeking local processing solutions, with significa...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

Zhen Ding Invests CNY 40 Billion to Expand High-End PCB Capacity in Huai'an

Zhen Ding, a key player in the printed circuit board (PCB) industry, has announced an investment of CNY 40 billion to expand its high-end PCB manufacturing capacity in Huai'an. The initiative, marked by a groundbreaking ceremony for its new HD campus...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

Intel's CPU Revival in the AI Era: An Early-Stage Recovery

Intel is experiencing a rebound in its CPU sector, specifically driven by the integration of artificial intelligence capabilities. This signal, though still in its initial stages, highlights the growing importance of AI in reshaping the hardware land...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 DigiTimes

Intel Bets on CPUs as the Backbone of AI Growth

Intel is strengthening its artificial intelligence strategy, positioning CPUs as a fundamental component for the expansion and adoption of AI technologies. This move underscores the persistent role of general-purpose processors in a GPU-dominated lan...

#Hardware #LLM On-Premise #DevOps
2026-04-24 DigiTimes

Largan, Sunny Optical Target FAU in Push Toward CPO and AI Optics

Largan and Sunny Optical are intensifying their efforts in developing Freeform Optical Units (FAU), crucial for advancing AI optics and Co-Packaged Optics (CPO) technologies. This strategic focus reflects the growing demand for high-speed, low-power ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-24 DigiTimes

AI Data Center Cooling: Asia Optical and Frore Systems Join Forces

Asia Optical and Frore Systems have announced a strategic collaboration focused on developing advanced cooling solutions for AI-dedicated data centers. This partnership aims to address the increasing thermal challenges posed by high-density AI archit...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 The Next Web

EV Market and Fuel Prices: Lessons for On-Premise AI Strategies

In Q1 2026, Tesla delivered 358,023 vehicles, a 6% increase but missing estimates, while US petrol prices surpassed $4/gallon. These market dynamics, influenced by geopolitical factors, highlight the need for tech decision-makers to carefully evaluat...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 Google AI Blog

Google's TPUs Tackle Increasingly Demanding AI Workloads

Google developed its Tensor Processing Units (TPUs) to accelerate increasingly complex artificial intelligence workloads. These specialized units are crucial for managing the growing demands of Large Language Model (LLM) training and inference. The a...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 Phoronix

Ubuntu 26.04 LTS: The New Foundation for On-Premise AI Infrastructures

Canonical has released Ubuntu 26.04 LTS, along with Ubuntu Server 26.04 LTS and its various flavors, powered by the Linux 7.0 kernel. This Long Term Support version represents a significant update for enterprises managing AI workloads, offering a sta...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 The Register AI

Datadog Introduces GPU Monitoring to Optimize AI Costs

Datadog has expanded its observability platform capabilities by integrating GPU monitoring. This new feature offers organizations managing AI workloads essential tools to analyze the efficiency and utilization of their most expensive hardware, thereb...

#Hardware #LLM On-Premise #DevOps
2026-04-23 Ars Technica AI

AI's Energy Footprint: Data Centers Could Outpace Nations in Emissions

Eleven data center campuses in the US, powered by new natural gas projects, could generate more greenhouse gases than Morocco's total emissions in 2024. These facilities, serving AI giants like OpenAI and Microsoft, bypass traditional grids for energ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 Phoronix

Ubuntu 26.04 LTS: ROCm Available via APT, But With Outdated Versions

Canonical and AMD collaborated to integrate ROCm libraries into the Ubuntu 26.04 LTS archive, promising simplified installation for AI/ML and HPC workloads. However, at the time of release, the versions available via `sudo apt install rocm` are sever...

#Hardware #LLM On-Premise #DevOps
2026-04-23 TechCrunch AI

Astronomical Research Fuels GPU Demand: Implications for the AI Market

Astronomers are increasingly adopting GPUs to analyze vast volumes of cosmic data, searching for patterns and anomalies. This growing reliance on hardware acceleration significantly contributes to the already high global demand for GPUs, a factor tha...

#Hardware #LLM On-Premise #DevOps
2026-04-23 The Register AI

AI's Demand Extends Chip Shortage to Traditional Servers

The escalating demand for AI solutions is creating a new wave of chip shortages, impacting essential components for general-purpose servers. Vendors are redirecting production capacity towards higher-margin AI server products, jeopardizing traditiona...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 The Register AI

Tesla Bets AI Future on Intel's Unfinished 14A Process

Elon Musk announced Tesla's plans to build proprietary AI chips, relying on Intel's 14A manufacturing process. This decision represents a significant gamble, as the 14A technology is still under development and not yet available. The initiative highl...

#Hardware #LLM On-Premise #DevOps
2026-04-23 Tom's Hardware

Nvidia H200: Sales Blocked in China and the Push for Local Industry

The U.S. Commerce Secretary confirmed that Nvidia H200 GPUs have not been sold to China. This move reflects restrictions imposed by the Chinese government, aimed at stimulating the development of its domestic semiconductor industry, with significant ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 TechWire Asia

Google Cloud Next 2026: New TPUs and AI Agent Platform

Google Cloud unveiled its eighth-generation Tensor Processing Units (TPU), the TPU 8t and TPU 8i, optimized for training and inference workloads. The company also introduced the Gemini Enterprise Agent Platform, a centralized environment for building...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

SpaceX and Tesla: Hardware Strategies Between GPUs and Custom Chips

SpaceX is exploring an expansion of its GPU capabilities, while Tesla is tapping Samsung for chip upgrades. These moves highlight the increasing importance of hardware control and computing power for tech companies, influencing on-premise deployment ...

#Hardware #LLM On-Premise #DevOps
2026-04-23 DigiTimes

SK Hynix to Shift Over Half of NAND Output to 321-Layer Chips

SK Hynix has announced a significant reorientation of its NAND memory production, dedicating over half of its volume to new 321-layer chips. This strategic move underscores the company's commitment to innovation in storage density, with direct implic...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

AI Demand Strengthens Semiconductor Equipment Cycle

The semiconductor industry is experiencing a recovery, driven particularly by the growing demand for artificial intelligence. This trend is strengthening the production equipment cycle, with companies like Lam Research benefiting from the recovery in...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

Lam Research: AI Sustains Demand for Semiconductor Equipment

Lam Research has reported sustained AI-driven momentum, leading to an improved outlook for the Wafer Fab Equipment (WFE) sector. This trend highlights the increasing demand for advanced hardware to support AI workloads, impacting the entire semicondu...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

TSMC Targets 2029 for A13 and A12 Nodes, Pillars of Future AI Chips

TSMC, a global leader in semiconductor manufacturing, has set 2029 as the target for the start of production for its next A13 and A12 process nodes. These advanced manufacturing processes are poised to become the foundation for the next generation of...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

Google Debuts TPU 8t and 8i as AI Workloads Diverge

Google Cloud has announced its new TPU 8t and 8i processors, designed to address the increasing diversification of artificial intelligence workloads. This move highlights the need for specialized hardware solutions, for both training and inference, a...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

SK Hynix Expands AI Memory Capacity with New HBM Packaging Hub in Cheongju

SK Hynix is building a new HBM packaging hub in Cheongju, South Korea. This initiative aims to significantly expand the production capacity of high-bandwidth memory, essential for powering the growing demand for artificial intelligence systems, both ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-23 DigiTimes

Strategic Hardware Investments: Zhen Ding's New Site and the AI Supply Chain

Zhen Ding Technology has commenced construction of a new facility in China, an event that underscores the importance of investments in the hardware supply chain. While specific details are limited, such initiatives are crucial for strengthening globa...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 The Next Web

SpaceX: Orbital AI Data Centers Between Ambition and IPO Filing Risks

SpaceX's confidential S-1 pre-IPO filing reveals that its plans for orbital AI data centers involve "significant technical complexity and unproven technologies," risking commercial non-viability. This statement contradicts Elon Musk's earlier claim i...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 ServeTheHome

Google Unveils New TPU 8i and 8t for AI Inference and Training

Google has announced its new eighth-generation Tensor Processing Units (TPUs), the TPU 8i and TPU 8t. Designed specifically for AI inference and training workloads, respectively, these proprietary solutions aim to optimize AI tasks within the Google ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 TechCrunch AI

Google Cloud Boosts AI Offering with New Chips: The Nvidia Challenge Continues

Google Cloud has introduced two new AI chips, the Tensor Processing Units (TPUs), promising superior performance and lower costs compared to previous generations. This move intensifies competition in the AI accelerator market, traditionally dominated...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 Ars Technica AI

Google Unveils Eighth-Gen TPUs for the 'Agentic Era'

Google has introduced its eighth generation of Tensor Processing Units (TPUs), diverging from the industry's widespread adoption of Nvidia accelerators. These new chips, designated TPU 8t for training and TPU 8i for inference, are engineered for the ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 The Next Web

Google Redefines TPUs: Separate Architectures for Training and Inference

Google announced the general availability of its seventh-generation TPU, Ironwood, and unveiled the eighth, comprising TPU 8t (for training) and TPU 8i (for inference). This new strategy involves dedicated chips, designed by Broadcom and MediaTek res...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 The Register AI

Grafana: Free AI Assistant for On-Premise and Open Source Deployments

Grafana has announced the free availability of its AI assistant, specifically targeting Open Source communities and users managing on-premise deployments. The initiative, unveiled at the Barcelona user conference, strengthens the company's commitment...

#Hardware #LLM On-Premise #DevOps
2026-04-22 Phoronix

Intel LLM-Scaler: vLLM 0.14.0-b8.2 Introduces Arc Pro B70 Support

Intel's LLM-Scaler initiative continues with the vLLM 0.14.0-b8.2 update. This version officially introduces support for the Arc Pro B70 graphics card, extending AI inferencing capabilities on Intel Arc hardware. The update aims to optimize performan...

#Hardware #LLM On-Premise #DevOps
2026-04-22 The Register AI

Google Accelerates AI: New TPUs and Arm-based Axion for Training and Inference

Google unveiled two new proprietary AI accelerators at the Cloud Next conference: one for training and one for inference, featuring Arm-based Axion cores. This strategic move highlights Google's commitment to developing custom silicio to optimize per...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 Tom's Hardware

AMD and the Evolution of AI Upscaling: Implications for Local Hardware

AMD's Software Development Kit (SDK) hints at the introduction of new 4x and 6x multipliers for AI-driven frame generation. This driver-level optimization underscores the growing trend of leveraging local GPU compute power for complex workloads, a cr...

#Hardware #LLM On-Premise #DevOps
2026-04-22 The Register AI

Energy Costs and AI: UK Firms Offshoring Workloads

One in five UK firms have already moved their AI workloads abroad, driven by high energy costs. This trend, extending beyond traditional call center offshoring, raises concerns for the British government, which relies on artificial intelligence as an...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 DigiTimes

Quartz Components: AI and Automotive Drive Demand for TXC and Taitien in 2026

TXC and Taitien, suppliers of quartz components, anticipate increased sales in the first quarter of 2026. This growth is fueled by rising demand in AI optical communication and the automotive sector, highlighting the critical role of foundational com...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 DigiTimes

MediaTek and Marvell: A Strategic Partnership for Future TPU Generations

The collaboration between MediaTek and Marvell for the supply of Tensor Processing Units (TPUs) for the next three generations marks a significant step in the AI hardware landscape. This strategic agreement highlights the growing importance of specia...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 DigiTimes

ASM International: 16% Growth in 1Q26 Reflects Booming AI Market Expansion

ASM International reported a 16% revenue increase in the first quarter of 2026, a figure highlighting strong demand in the artificial intelligence sector. This outcome underscores how the semiconductor supply chain is a fundamental pillar for the dev...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 DigiTimes

Apple Taps Apple Silicio Veterans for AI Hardware-Software Balance

Apple is re-aligning internal resources to address the challenges of the artificial intelligence era. The company has entrusted two key figures, previously responsible for the success of Apple Silicio, with the task of balancing software and hardware...

#Hardware #LLM On-Premise #DevOps
2026-04-22 ArXiv cs.CL

2D Early Exit Optimization: New Horizons for On-Premise LLM Inference

A two-dimensional early exit strategy revolutionizes LLM inference by coordinating layer-wise and sentence-wise exiting. This incremental method generates multiplicative computational savings, surpassing single optimizations. Tested on 3B-8B paramete...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-22 DigiTimes

Taiwan Suppliers Anticipate Renewed Focus on Hardware Innovation for AI

Taiwan's suppliers expect a renewed drive towards innovation, with significant implications for AI hardware. This trend is crucial for companies evaluating on-premise deployment strategies for Large Language Models, impacting TCO, data sovereignty, a...

#Hardware #Fine-Tuning
2026-04-21 The Register AI

CPU Monitoring: Task Manager's Legacy and On-Premise Challenges

Task Manager's CPU meter, based on simple kernel calls, represents a bygone era. Today, for on-premise Large Language Model deployments, granular hardware monitoring beyond the CPU is essential, including VRAM, throughput, and latency. This visibilit...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-21 DigiTimes

Geopolitical Dynamics and Digital Autonomy: The Role of Self-Hosted AI

Recent geopolitical measures and the affirmation of independent economic goals, as reported by DIGITIMES, highlight the importance of sovereignty and control. This context is mirrored in the tech sector, where companies are increasingly evaluating se...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-21 DigiTimes

AI Drives Memory: ASML's HBM Revenue Surpasses Logic in 1Q26

In the first quarter of 2026, ASML's revenue from memory production exceeded that from logic, signaling the surging demand for High Bandwidth Memory (HBM) fueled by artificial intelligence. This trend highlights AI's impact on the semiconductor suppl...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-21 DigiTimes

Strategic Collaboration to Enhance On-Premise LLM Deployments

Industry experts are urging greater collaboration among companies, institutions, and governments to accelerate the development and adoption of self-hosted LLM infrastructures. The goal is to strengthen data sovereignty, optimize TCO, and ensure granu...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-21 DigiTimes

Google's AI chip push: A new phase in the battle with Nvidia

Google is intensifying its development of dedicated AI chips, aiming to capitalize on the expanding inference boom. This move marks a new phase in the competition with Nvidia, highlighting the importance of specialized hardware solutions for AI workl...

#Hardware #LLM On-Premise #DevOps
2026-04-21 DigiTimes

AI Demand Inflates Silicio Valuations: Impact on TSMC and Nvidia

The surge in artificial intelligence demand is exerting significant pressure on the silicio supply chain, influencing the valuations of industry giants like TSMC and Nvidia. This scenario presents new challenges for enterprises evaluating on-premise ...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-21 Phoronix

AMD GAIA: Portable AI Agents for Local Deployments

AMD is enhancing GAIA, its cross-platform software solution built around the Lemonade SDK, for running local AI agents on AMD hardware (CPUs, GPUs, NPUs). The latest update introduces portability for custom AI agents, facilitating easy import and exp...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 DigiTimes

AI Reshapes Memory Supply: Procurement Strategies Under Scrutiny

The advancement of artificial intelligence is profoundly altering the memory supply chain, prompting the Global Electronics Association to issue a warning. Traditional procurement strategies, no longer adequate for the specific demands of AI workload...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 DigiTimes

Amazon's Chip Journey: Trainium and its Leading AI Customers

Amazon has invested for over a decade in developing proprietary chips, culminating in Trainium. This analysis reveals how Anthropic and OpenAI have emerged as key customers for this technology, highlighting the growing adoption of custom hardware for...

#Hardware #LLM On-Premise #DevOps
2026-04-20 DigiTimes

Vivo's High-End Push and On-Device AI: Implications for Edge Computing

Vivo is strengthening its high-end offering in Taiwan with the new X300 series, aiming for 40% sales growth. This market move highlights the increasing importance of advanced mobile devices, platforms increasingly capable of running artificial intell...

#Hardware #LLM On-Premise #DevOps
2026-04-20 DigiTimes

Samsung Improves HBM4 Production: Nvidia Praises 4nm Innovation

Samsung has made significant progress in the production yield of HBM4 memory, a critical component for next-generation AI accelerators. The company also implemented a 4-nanometer PMBIST process upgrade, which received positive feedback from Nvidia. T...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 The Next Web

Google Challenges Nvidia in AI Inference with Diversified Chip Supply Chain

Google is building a custom chip supply chain for AI inference, involving four partners (Broadcom, MediaTek, Marvell, Intel). The strategy, which includes Ironwood TPUs and future 2nm TPU v8 chips, aims to challenge Nvidia, offering new perspectives ...

#Hardware #LLM On-Premise #DevOps
2026-04-20 404 Media

Control and Sovereignty: From Indie Journalism to On-Premise AI Deployment

Maddy Myers, editor-in-chief of Mothership, founded an independent publication focused on gender and video games, highlighting the value of controlling one's platform and content. This principle of "owning your work" finds a significant parallel in t...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 The Register AI

AI Reshapes Britain's Datacenter Map: New Priorities Beyond London

The rise of artificial intelligence is fundamentally altering the geography of datacenters in the UK. Experts indicate a progressive shift away from London, driven by power shortages and planning constraints. Reduced reliance on low-latency connectio...

#Hardware #LLM On-Premise #DevOps
2026-04-20 DigiTimes

High-Performance Materials: A Pillar for On-Premise AI

Taiwanese textile firms are diversifying into aerospace and drones, leveraging advanced materials. This trend highlights the critical importance of such innovations for developing robust and high-performance hardware, essential for on-premise AI infr...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 DigiTimes

AI Boom Drives Taiwanese Chip Testing Firms to Record Results

The increasing demand for artificial intelligence solutions is significantly impacting the global supply chain. Taiwanese chip testing firms, a crucial link in AI hardware production, reported record financial performance in Q1 2026, highlighting the...

#Hardware #LLM On-Premise #Fine-Tuning
2026-04-20 DigiTimes

The AI Chip Race: ABF Substrates Sold Out for Key Suppliers

The escalating demand for AI chips is straining the supply chain, with ABF (Ajinomoto Build-up Film) substrates reported as sold out from key suppliers like Unimicron, Kinsus, and Nan Ya PCB. This shortage highlights a potential bottleneck in AI acce...

#Hardware #LLM On-Premise #Fine-Tuning
← Back to All Topics