UALink and the Battle for Rack-Scale GPU Interconnect

UALink and the Battle for Rack-Scale GPU Interconnect Abstract Scaling AI workloads to tens of thousands of accelerators requires more than raw FLOPS — it requires the ability to connect GPUs into a coherent compute fabric. Nvidia has long dominated this domain with NVLink and NVSwitch, which provide high-bandwidth, low-latency interconnects tightly integrated into its […]
Read More

Ethernet Fabrics and the Rise of Open AI Networking

Ethernet Fabrics and the Rise of Open AI Networking Abstract Artificial intelligence has transformed datacenter architecture. The challenge today is not just how fast a single GPU can compute, but how efficiently tens of thousands of accelerators can work as one coherent, distributed system. For the past two decades, InfiniBand has dominated this role, offering […]
Read More

How Large Language Models Are Quietly Changing RTL Design?

How Large Language Models Are Quietly Changing RTL Design? RTL design has always been one of the most crucial – and complex – steps in digital chip development. It’s where an engineer takes a specification and expresses it in hardware behaviour using languages like Verilog or System Verilog. As chips become more advanced – especially […]
Read More

Verifying AI Chips at RTL: How Foundation Models Are Transforming Design Assurance

Verifying AI Chips at RTL: How Foundation Models Are Transforming Design Assurance Modern AI chips are not general-purpose processors. Their RTL design logic is optimized not for instruction throughput, but for dataflow acceleration, parallel computation, and high-efficiency matrix operations. These chips integrate systolic arrays, sparsity-aware data paths, configurable tensor engines, and hierarchical memory systems. This […]
Read More

Making Automotive SoCs Truly Safe

Making Automotive SoCs Truly Safe Why functional safety changes the verification game? Most SoC verification teams know how to prove a chip meets its functional specification. You write and run simulations, close coverage, debug mismatches, and tape out. But in automotive design — especially when working to ISO 26262 — the job isn’t done when […]
Read More

LLM-Aided RTL Design and Verification in Pre-Silicon Flows

LLM-Aided RTL Design and Verification in Pre-Silicon Flows The integration of large language models (LLMs) into semiconductor pre-silicon workflows has moved from isolated academic prototypes to production-grade deployments in leading EDA environments. In this emerging paradigm, generative models are not limited to assisting with documentation or simple code generation; they actively participate in RTL creation, […]
Read More

Exploring Agentic Misalignment: A Deep Dive for Businesses

When AI Acts Like a Threatened Human: Exploring Agentic Misalignment: A Deep Dive for Businesses Artificial Intelligence (AI) is transforming how businesses operate, with AI agents taking on increasingly autonomous roles in areas like finance, legal, and IT. These agents, powered by advanced large language models (LLMs), can handle complex tasks with minimal human oversight. But […]
Read More

AMD’s Game-Changer: Instinct MI350/MI400

AMD’s Game-Changer: Instinct MI350/MI400 & Helios—Delivering Open, Scalable AI Infrastructure In our ongoing series on AI accelerators, we previously explored the hardware fueling the AI revolution. Now, we turn our focus to AMD’s latest unveilings—the Instinct MI350/MI400 GPUs and Helios rack-scale system. These innovations tackle critical bottlenecks in AI infrastructure, positioning AMD as a key […]
Read More

Deep-Tech Challenges in 2nm Physical Design Stage

Deep-Tech Challenges in 2nm Physical Design Stage 1. Physical Implementation Standard Cell Design & Placement GAA introduces discrete nanosheet widths (e.g., 2/3/4 stacks), forcing standard cells to be quantized, which reduces flexibility in sizing and optimization. Cell height (track-based design) is no longer easily scalable due to BEOL and BSPDN constraints – placement becomes non-uniform and harder to optimize. GAA-based […]
Read More

AI Accelerators: Powering Embedded Systems, Edge Computing, and On-Device LLMs

AI Accelerators: Powering Embedded Systems, Edge Computing, and On-Device LLMs Introduction AI accelerators—specialized hardware like NPUs, TPUs, GPUs, FPGAs, and ASICs—are revolutionizing how real-time AI operates in smart devices, from augmented reality (AR) glasses to autonomous vehicles. These accelerators are designed to handle compute-heavy tasks such as matrix multiplications and convolutions, enabling powerful AI inferencing […]
Read More