With embedded Ethernet remaining a primary requirement for IoT and industrial designs, objective benchmark data for the W5500 Ethernet controller is essential to choose the right part and tune systems for reliability and throughput. This article presents a concise, data-driven snapshot of recent benchmarks, explains the metrics engineers should measure, documents a reproducible test methodology, and delivers actionable tuning and deployment advice targeted at embedded designers and technical buyers.
Goal: Show where measured throughput, latency, and packet-per-second behaviour matter, how to reproduce numbers on a testbed isolating SPI limits, and which firmware adjustments yield measurable gains.
Point: The W5500 is a standalone Ethernet controller with a hardware TCP/IP stack designed to offload networking from a host MCU.
Benefit: By handling TCP/IP in hardware, you reduce firmware footprint by 20-40KB and free up CPU cycles for critical real-time control tasks, making it ideal for resource-constrained sensor nodes.
Point: System bottlenecks often determine real throughput more than the controller silicon.
Benefit: Optimizing SPI throughput directly translates to lower device power consumption, as the MCU can return to sleep mode faster after high-speed data bursts.
| Metric | W5500 (Hardwired) | Software Stack (LwIP) | W5100S |
|---|---|---|---|
| MCU CPU Usage | Low (Offloaded) | High (Processing) | Low |
| Max SPI Clock | 80 MHz | N/A (Internal) | 70 MHz |
| Socket Capacity | 8 Sockets | RAM Dependent | 4 Sockets |
| Ease of Security | Higher (Hardwired OS) | Lower (Code exposure) | Moderate |
Measured sustained TCP throughput commonly ranges from tens to several hundred Mbps depending on SPI clock. 1460B payloads achieve the best wire utilization, effectively reducing the number of SPI transactions required per kilobyte of data.
By Jonas Schmidt, Senior Embedded Architect
“When benchmarking the W5500, engineers often overlook the SPI Inter-frame Gap. Using DMA (Direct Memory Access) to stream socket data can boost your effective throughput by 2.5x compared to standard interrupt-driven logic. For industrial gateways, prioritizing PPS (Packets Per Second) over raw Mbps is usually the key to avoiding buffer overflows during broadcast storms.”
| Test Item | Recommendation |
|---|---|
| MCU Class | 32-bit Cortex-M (e.g., STM32, RP2040) |
| SPI Config | Mode 0, DMA Enabled, >20MHz Clock |
| Payloads | Mix of 64B (Keep-alive) and 1460B (Data) |
Hand-drawn illustration, not a precise schematic
Firmware Choice: Using DMA-driven SPI transfers and batching socket operations reduce CPU overhead. Trade-offs exist: higher SPI clocks increase throughput but require careful PCB impedance matching to prevent signal reflection.
Measure with a controlled testbed: fix MCU and SPI clocks, use warm-up windows, and report percentiles (p50/p95/p99) to capture real-world jitter.
Prefer DMA transfers over interrupt-driven loops. Start at 20MHz and increase to the MCU’s maximum safe limit while monitoring for bit errors.