Breaking News

GLOBAL Wi-Fi NEWS bulletin

Search

The Internet of Things (IoT)is in the midst of a quiet revolution, and at the center of this shift is Matter. Developed by the Connectivity Standards Alliance (CSA) to streamline smart home integration, Matter has been one of the most rapidly evolving technologies in the IoT space.

What is Matter, and Why Does it Matter?

Matter is a unified application layer that abstracts the underlying transport technologies—like Wi-Fi, Thread, and Ethernet— so smart devices can communicate reliably, regardless of their network protocol. With it, developers and device makers can now build secure, interoperable products that work seamlessly across ecosystems.

In its short, lifespan, Matter has evolved from 1.0 to Matter 1.4, which now works with 54 device types including energy management devices, enhanced entertainment controls, and solar EV chargers.

Historically, HANs (home area networks) used proprietary or utility-specific protocols. But with Matter’s broad industry support and open standard approach, it’s positioned to become the default protocol for future HAN deployments, allowing energy devices to talk natively to the meter and each other.

Matter is designed to solve the biggest problem in smart homes: devices that don’t talk to each other.

With Matter:

  • Your thermostat can work with any assistant you like.
  • Your EV charger can coordinate with your energy usage automatically.
  • Your HVAC, lights, plugs, and solar inverter can finally act like a team.
  • Your smart meter becomes a real energy orchestrator—not just a reader.

And it’s all local, secure, and fast. No cloud dependencies. No silos. Just devices that understand each other—no matter who made them.

Think of Matter as a USB for smart homes. You plug it in, and it just works. Now imagine every energy device in your home—the EV charger, heat pump, solar inverter, and your smart meter—are all speaking Matter. They don’t just coexist, but can collaborate to save you money, shift energy usage, and make your home smarter without lifting a finger.

The Evolving Role of Smart Meters

Smart meters are no longer just energy monitors. They’re evolving into interactive nodes in the energy ecosystem, capable of:

  • Reporting time-of-use (TOU) energy prices
  • Communicating real-time consumption data to smart devices
  • Acting as a HAN controller for Matter devices
  • Enabling demand response by controlling or signaling to HVACs, EV chargers, and more

In the near future, a smart meter will not just passively measure usage—it will coordinate with local energy-consuming and generating devices to optimize for both cost and sustainability.

Imagine an EV charger seeing a high peak price signal from your meter and delaying charging by 30 minutes to save you money—or your solar inverter exporting more power because the meter detected a TOU bonus. With more responsibility on smart meters comes the need for better communication across devices.

Ecosystems are Rallying Around Matter

If your device doesn’t support Matter, it can’t communicate or operate with other Matter-enabled devices. It’s sidelined while everything else coordinates in real time—EV chargers, HVAC systems, solar inverters, and more. You could try retrofitting it later, but that would be like trying to learn a new language after you’ve already moved abroad — you’re starting at a disadvantage.

Why Wi-Fi Is the Right Call for Smart Meters

Thread is great for things like motion sensors and door locks—low power, short bursts of communication. But smart meters are plugged in, always on, and live outside your house.

Here’s why Wi-Fi wins:

  • Always connected: Meters need reliable, IP-based, 24/7 communication.
  • No extra gear: Virtually every home already has Wi-Fi. Thread requires a border router, necessitating another box and more setup.
  • Range matters: 2.4 GHz Wi-Fi can reach the side of the house, backyard, or wherever the meter lives.
  • Save money: Today, some meters have $5–10 app processors to detect devices like EV chargers. But soon, all those devices will announce themselves with Matter and will not require heavy processing inside the meter.

Future-Proof Your Meters with Matter Over Wi-Fi or Risk Falling Behind

We all agree that retrofitting meters in the field is a nightmare. It’s expensive, slow, and not always possible. Waiting until Matter over Wi-Fi is the new standard means you risk deploying devices that can’t speak the dominant smart home language. You’ll be stuck with meters that can’t connect to anything, and users who are stuck paying more for energy.

Silicon Labs’ SiWG917 Wi-Fi Modules are engineered specifically to deliver strong, industrial-grade, single-band 2.4 GHz Wi-Fi that is perfect for outdoor meters and built for Matter. Known for its ultra-low power performance, built-in AI/ML accelerator, robust security, and flexible architecture, the SiWG917 is the ideal choice. To learn more about the power saving features of the SiWx917, read our whitepaper, Achieve Ultimate Energy Efficiency for Low-Power Wi-Fi.

Don’t let your meter be an outsider.

Build it with Wi-Fi.

Build it with Matter.

Build it ready for the world it’s about to enter.

Smart meters need to join in, not just watch from the sidelines. Wi-Fi gives them a seat at the table, and Matter makes sure they can talk to everyone.

So don’t play catch-up in two years. Get started today by visiting our Matter Developer Journey.

The electric grid is the backbone of our modern society here in North America. Ensuring its reliability and security is paramount, which is where the North American Electric Reliability Corporation (NERC) Critical Infrastructure Protection (CIP) standards come in. These standards provide a framework for securing the Bulk Electric System (BES) against cyber threats.

However, with the grid undergoing significant modernization and increased connectivity, meeting these stringent cybersecurity requirements presents a complex challenge for power utilities. More connected devices mean a larger attack surface, demanding a robust and phased approach to security.

Cisco’s Phased Approach to Industrial Threat Defense

Cisco recognizes that enhancing your security posture is a journey. We advocate for a phased approach, building foundational security elements that support subsequent steps, allowing utilities to improve security at their own pace while demonstrating value. The Cisco Industrial Threat Defense solution offers a modular and comprehensive set of capabilities designed to address the unique challenges of securing operational technology (OT) environments and achieving NERC CIP compliance.

How Cisco Solutions Help Address Key NERC CIP Requirements:

Cisco just published a solution brief describing the key NERC CIP requirements and how our portfolio can help utilities to comply. Here is a quick summary:

  1. Visibility and Categorization (CIP-002, CIP-015):
    • Cisco Cyber Vision: Provides deep packet inspection embedded in the industrial network to automatically discover and inventory all grid assets, their communication patterns, and vulnerabilities. This visibility is fundamental for categorizing BES Cyber Systems (CIP-002) and is a core component of Internal Network Security Monitoring (INSM) (CIP-015). It helps identify risks and deviations from expected behavior.
    • Splunk OT Security Add-On: Aggregates data from various sources, including Cyber Vision, to provide asset classification visibility (CIP-002) and supports monitoring for INSM (CIP-015).
  2. Electronic Security Perimeters (ESPs) and Access Control (CIP-005, CIP-007):
    • Cisco Industrial Routers and Secure Firewalls: Serve as the backbone for defining and enforcing ESPs. They offer comprehensive Next-Generation Firewall (NGFW) features, stateful inspection, application control, and integrated intrusion prevention (IDS/IPS) to manage electronic access and block threats at the perimeter (CIP-005, CIP-007). They can enforce unified security policies across distributed sites.
    • Cisco Secure Equipment Access (SEA): Provides a Zero-Trust Network Access (ZTNA) solution for secure remote access, crucial for managing vendor and remote user access to BES Cyber Systems. It enforces least privilege, just in time access and supports multi-factor authentication (MFA) as well as session monitoring/recording (CIP-005).
    • Cisco Catalyst Center and Identity Services Engine (ISE): Help manage security policies centrally across switching infrastructure, control physical port usage, and enforce access controls via IP ACLs or Security Group ACLs (CIP-007).
    • Splunk OT Security Add-On: Collects logs from firewalls, routers, switches, and access systems to monitor activity crossing the ESP boundary (CIP-005) and track ports, services, and system access control events (CIP-007).
  3. System Security Management & Vulnerability Assessment (CIP-007, CIP-010):
    • Cisco Catalyst SD-WAN Manager and Catalyst Center: Enable centralized management of network device configurations, helping prevent unauthorized changes and facilitating the deployment of ‘golden’ configurations (CIP-010). They also support security event monitoring on network infrastructure (CIP-007).
    • Cisco Cyber Vision: Identifies vulnerabilities in discovered assets and highlights those actively exploited by bad actors to help prioritize patching. Also monitors deviations from network communication baselines (CIP-010).
    • Splunk OT Security Add-On: Aggregates logs from various sources (firewalls, endpoints, etc.) to track ports/services, security events, malware alerts, and supports baselining efforts (CIP-007, CIP-010). It also helps track compliance with log retention requirements (CIP-007).
  4. Incident Reporting, Response, and Recovery (CIP-008, CIP-009):
    • Splunk: Acts as a central SIEM for collecting, correlating, and analyzing security events from across the network and security tools. It supports incident detection, investigation, and reporting, helping utilities meet the requirements for identifying and responding to cyber incidents (CIP-008).
    • Cisco Catalyst Center and Catalyst SD-WAN Manager: Provide monitoring and recovery capabilities for network equipment, supporting the restoration of network infrastructure in case of failure or attack (CIP-009).
    • Splunk OT Security Add-On: Provides dashboards to monitor notable security alerts (CIP-008) and brings in data from backup logs and Splunk environment status to support recovery plan requirements (CIP-009).
  5. Information Protection & Supply Chain Risk (CIP-011, CIP-013):
    • Cisco Network Infrastructure & Security Policies: Enforce network segmentation and access controls to protect BES Cyber System Information (BCSI) from unauthorized access (CIP-011).
    • Cisco Security and Trust Organization: Cisco’s commitment to security is embedded in its Secure Development Lifecycle (SDL), certified for IEC 62443-4-1. Trustworthy technologies like image signing and secure boot ensure product integrity. The Cisco Product Security Incident Response Team (PSIRT) handles vendor-identified incidents and provides vulnerability information, patches, and mitigation advice (CIP-013). Cisco is also an active contributor to relevant industrial security standards.

A Unified Approach for Enhanced Security

Navigating NERC CIP compliance requires a strategic, solutions-based approach. Cisco provides the building blocks and integrated solutions to help power utilities secure their critical infrastructure, enhance visibility, and meet regulatory requirements effectively. Have a look at our NERC CIP Compliance Solution Brief to better understand the requirements and see how Cisco can help.

I will be presenting a webinar on July17th together with experts from Burns & McDonnell to discuss the new Internal Network Security Monitoring (INSM) CIP-015 standard and solutions available to help Utilities comply. Save the date and register now.

New partnership delivers simple and easy WireGuard VPN service to USG FLEX H Series firewall customers at no additional cost

ANAHEIM, California (June 26, 2025) Zyxel Networks, a leader in delivering secure, AI- and cloud-powered business and home networking solutions, today announced a new partnership with Tailscale, the leading identity-native connectivity platform. This integration brings simplified, scalable, and secure VPN connectivity to Zyxel Networks’ USG FLEX H Series firewalls, empowering small businesses and advanced users to build private, peer-to-peer networks with ease and at no additional cost.

Now available on Zyxel Networks USG FLEX H Series firewalls running uOS v1.32 and above, Tailscale’s WireGuard-based mesh VPN network is fully integrated into the firewall’s management interface. Once enabled, users can set up secure remote access in minutes without the need for complex configurations.

Streamlined VPN experience

The USG FLEX H Series firewalls now support a full suite of VPN protocols, including IPSec, SSL and WireGuard, to secure a wider range of use cases and endpoints. The Tailscale integration allows users to take advantage of a streamlined setup experience, which requires no server configuration or port forwarding. Devices automatically connect to the private network upon a few simple clicks, making the process accessible even to those with minimal networking experience.

Protected access control

With the Tailscale integration, firewall users can now create a peer-to-peer mesh network where devices communicate directly with each other via encrypted tunnels, instead of routing through a central server. This architecture reduces latency, boosts performance and scales efficiently as businesses grow. Furthermore, the service supports a wide range of operating systems, including Windows, macOS, Linux, iOS and Android, as well as embedded devices, and cloud services directly, allowing firewall users to extend VPN protection across all major endpoints–including mobile devices and BYOD environments.

For added convenience and security, users can log in to Tailscale using their existing identity providers such as Google Workspace, Microsoft Entra ID (Azure AD), Okta, OneLogin, or JumpCloud. Multi-Factor Authentication (MFA) can also be implemented through these accounts, providing an extra layer of protection while simplifying access control and user management.

“VPNs are essential for protecting data and supporting hybrid workforces, but they have often been too complex for smaller organizations with limited IT resources,” said Ken Tsai, President of Zyxel Networks. “This integration is the answer our customers have been asking for a simpler, faster and smarter way to enable secure connections across all their devices, wherever they are.”

“VPNs have a reputation for being slow, painful to set up, and impossible to manage. We started Tailscale to fix that, and now we’re excited to bring that simplicity to more people through our partnership with Zyxel Networks,” said Avery Pennarun, co-founder and CEO of Tailscale. “Zyxel’s reach in the SMB market means we can help a lot more teams skip the ‘VPN headache and just get back to work—securely.”

As part of the partnership’s launch, Zyxel Networks is offering eligible USG FLEX H Series customers free access to Tailscale’s Starter Plan. For more information, please visit https://community.zyxel.com/en/discussion/29583

For more information about Zyxel, its connectivity solutions, and the Zyxel Authorized Partner Program, visit www.zyxel.com/us and follow us on FacebookX(formerly Twitter) and LinkedIn.

About Zyxel Networks

Zyxel Networks is a leading provider of secure, AI-powered cloud networking solutions for SMBs and the enterprise edge. We deliver seamless connectivity and flexible scalability through subscription services, all backed by robust security. With a reputation built on decades of unlocking potential and helping people adapt to the changing workplace, Zyxel Networks has earned the trust of over 1 million businesses across 150 markets.

Is your WiFi painfully slow when it should be blazing fast? WiFi channel overlap is likely the issue. It’s like trying to have a conversation at a rock concert – everyone’s shouting, but nobody can understand a word.

For network professionals, understanding and addressing WiFi channel overlap can make the difference between a high-performing network and one that constantly disappoints users. Poor channel planning can cut network capacity in half. For businesses, this directly impacts operations through slower connections, dropped calls, and frustrated users.

In this blog, we’ll explore what channel overlap is, how it affects performance across different frequency bands, and provide practical solutions to identify and fix these issues.

WiFi Channels Explained: 2.4GHz vs 5GHz vs 6GHz

WiFi channels function like highway lanes. When too many devices use overlapping frequencies, performance suffers dramatically. Regional regulations affect which channels are available in different countries across all frequency bands.

2.4GHz Band

  • Only channels 1, 6, and 11 are WiFi non-overlapping channels
  • Offers better range but suffers from significant congestion
  • In North America, channels 1-11 are available, while some regions permit up to channel 13 (Japan exclusively allows channel 14

The WiFi channel overlap chart above shows why channels 1, 6, and 11 are preferred – they don’t overlap with each other.

Pro Tip:

Channel 14 remains illegal everywhere except Japan. Stick with WiFi non-overlapping channels 1, 6, and 11 unless you’re keen on having a friendly chat with regulatory authorities.

5GHz Band

  • Provides 25 non-overlapping channels (quantity varies by country)
  • Less crowded but doesn’t reach as far
  • Supports multiple channel widths: 20/40/80/160MHz
  • Many 5GHz channels are labeled as “DFS” (Dynamic Frequency Selection). These channels share the same frequencies that weather and military radar systems use. WiFi devices must monitor for radar signals and switch channels if detected. Regulations for these channels vary by country.

6GHz Band (WiFi 6E and WiFi 7)

  • Delivers 59 non-overlapping channels in the US
  • Offers clean, uncluttered spectrum without legacy device interference
  • Supports channels up to 320MHz wide with WiFi 7
  • Not all countries have approved 6GHz for WiFi use, with varying amounts of spectrum allocated in different regions

Pro Tip:

Channel width impacts performance significantly. Think of it as road lanes – more lanes (wider channels) move more data but consume more of the limited frequency space.

How WiFi Channel Overlap Causes Interference

Two types of interference result from WiFi channel overlap.

Co-Channel vs. Adjacent Channel Interference

Type What Happens Impact
Co-Channel (CCI Access points on same channel coordinate transmissions Slower but remains functional
Adjacent-Channel (ACI) Overlapping signals corrupt each other Data loss and constant retransmissions

Older WiFi standards use CSMA/CA (Carrier Sense Multiple Access with Collision Avoidance), which works like a polite conversation. Before transmitting, devices first listen to see if the channel is clear. If they detect another transmission, they wait a random period before trying again.

Here’s why CSMA/CA works well with co-channel interference but fails with partial overlap:

  • With co-channel interference (same channel): Devices can properly “hear” each other and politely take turns. Performance slows as more devices join, but the system remains functional.
  • With adjacent channel interference (overlapping channels): Devices can’t properly detect transmissions on partially overlapping channels. They incorrectly think the channel is clear and transmit anyway, causing signal corruption when the transmissions collide. This leads to data loss and constant retransmissions.

Newer WiFi 6/6E/7 standards use OFDMA (Orthogonal Frequency Division Multiple Access), which divides the channel into smaller resource units that can be assigned to different devices simultaneously. This is like converting a single-lane road into multiple lanes, allowing more efficient handling of traffic. While OFDMA reduces co-channel interference significantly, it still cannot fully solve adjacent channel interference problems.

 

Pro Tip:

Adjacent channel interference creates far more problems than co-channel interference. When choosing between a busy non-overlapping channel or an empty overlapping channel, always select the busy non-overlapping option.

 

How to Check WiFi Channel Overlap

How to check if your WiFi has channel overlap issues, requires specialized tools. NetAlly’s solutions reveal what’s actually happening in your wireless environment:

  • AirCheck G3 PRO – Handheld analyzer that immediately displays channel usage patterns and identifies WiFi channel overlap
  • EtherScope nXG – Provides comprehensive analysis of both WiFi and wired networks

How to check WiFi channel overlap symptoms: Watch for fluctuating performance, unexpected disconnections, and the frustrating combination of strong signal but poor throughput.

Pro Tip:

Channel interference problems can also be caused by non-WiFi interferers. In this case you should use a spectrum analyzer like the NXT-2000 from NetAlly, which allows you to identify non-WiFi interference sources that standard analyzers completely miss and measures both channel occupancy and noise floor levels.

Best Practices for WiFi Channel Planning

Quick Channel Selection Guide

Band Best Practice Channel Width
2.4GHz Use exclusively WiFi non-overlapping channels 1, 6, 11 Stick to 20MHz for all setups
5GHz Start with non-DFS channels 40MHz works well in many cases
6GHz Spread usage across available channels 80MHz can be used in many instances

Effective channel planning means configuring access points, so they don’t interfere with each other. For 2.4GHz, always use channels 1, 6, and 11 because they don’t overlap. For 5GHz, start with channels that don’t require radar detection (36-48 and 149-165) before using DFS channels.

The 6GHz band works with both WiFi 6E and WiFi 7 devices and offers much more space for wider channels. For outdoor deployments, 6GHz requires checking a database (AFC) before transmitting to protect existing users.

Choose channel widths based on how many access points you have:

  • Home/Small Office: Wider channels (40/80/160/320MHz) give faster speeds when you have fewer access points
  • Medium Business: Mix of 40MHz on 5GHz and 80MHz on 6GHz balances speed and capacity
  • High-Density Environment: Narrower channels (20/40MHz) allow more non-overlapping channels to be used

For networks with multiple access points, plan their placement and channel assignments carefully to minimize overlap. A channel overlap chart like that available on NetAlly tools helps you see which channels will conflict with each other.

How to Fix WiFi Channel Overlap Problems

  1. Assess – Determine which channels your network and neighbors currently use with a WiFi channel overlap chart
  2. Implement – Switch to WiFi non-overlapping channels (e.g. 1, 6, 11 on 2.4GHz)
  3. Adjust Power – Lower power settings to prevent signals from access points using the same channel to overlap with each other
  4. Verify – Measure throughput in previously problematic areas

If forced to choose, even when it’s busier, go with the crowded channels instead of the overlapping ones.

To effectively resolve WiFi channel overlap issues and boost network performance, explore NetAlly’s purpose-built tools: the AirCheck G3, AirMagnet Survey PRO, EtherScope nXG, and CyberScope Air.

Wireless connectivity has become a pillar of modern vehicle design. But as OEMs and Tier 1 suppliers integrate more consumer-grade wireless technology into their vehicle platforms, they are faced with the same set of challenges confronting many electronics producers: how to create and manage efficient, repeatable, scalable and cost-effective testing procedures from the early stages of R&D through high-volume manufacturing.

The primary focus of automotive design teams is to enable innovation. However, without adequate test protocols at key points along the design chain, test engineering runs the risk of lagging behind, which can affect product quality, time to market and the user experience.

Wireless Connectivity Is in the DNA of Automotive Design 

As a mainstay for enhancing safety, convenience and the overall driving experience, the most prevalent wireless connectivity technologies include Bluetooth®, Wi-Fi and cellular networks that enable functions like hands-free calling, real-time navigation, over-the-air (OTA) software updates and remote diagnostics. These advancements have transformed vehicles into connected hubs, meeting consumer expectations for seamless digital integration.

 

The global market for automotive connectivity is projected to grow from $33.4 billion in 2024 to $190.3 billion by 2033, with a compound annual growth rate of 19 percent. This growth underscores the increasing demand for vehicles equipped with advanced connectivity features.

Looking ahead, wireless connectivity is poised to enable more sophisticated applications in future vehicle models. These include support for autonomous driving technologies as well as vehicle-to-everything (V2X) communication, which allows cars to interact with each other and with infrastructure to improve traffic flow and safety. As these innovations become more prevalent, the role of robust and reliable in-vehicle wireless connectivity will only become more critical.

In-Vehicle Wireless Design: A Fragmented Ecosystem 

Within the automotive design ecosystem, the different subsystems in the car will have different needs. The requirements for secure keyless entry are not the same as infotainment or embedded telematic control units (TCUs) that wirelessly connects the vehicle to the cloud.

Different wireless applications exist on their own using different technologies for connectivity, each with unique test needs. The consequence is clear: a one-size-fits-all approach to automotive testing is fraught with risk due to the number and diversity of wireless integration vectors.

Wireless Weaves a Tapestry of Innovation

Wireless technologies like Wi-Fi, Bluetooth and cellular networks were initially developed for consumer applications — smartphones, laptops, smart home devices — where rapid innovation, large production volumes and price competition have driven fast iteration and falling costs. When these same technologies are integrated into vehicles, the test and deployment landscape changes dramatically.

 

Automotive environments are far more demanding. OEMs in this space are risk-averse and often late adopters of new technologies, prioritizing safety, security and long-term reliability over cutting-edge features. Unlike consumer devices, which may be replaced every two to three years, automotive systems must function flawlessly over much longer product lifecycles and under harsher operating conditions. This means more rigorous test requirements — especially in RF performance, environmental durability and functional safety.

Let’s look at the key points and test challenges for several of the most important technologies:

  • Wi-Fi is widely used in vehicles for infotainment systems, enabling media streaming, OTA updates and real-time diagnostics. In-vehicle Wi-Fi testing poses unique challenges. Dense RF environments filled with Bluetooth, cellular and even radar signals can cause interference. Coexistence testing is critical to ensure robust, uninterrupted connectivity. With Wi-Fi 7 introducing technologies like 4096-QAM and Multi-Link Operation, the test complexity increases further due to tighter Error Vector Magnitude (EVM) requirements and the need to validate operation across multiple frequency bands.

 

  • Bluetooth has evolved from a convenience feature to a core connectivity technology for hands-free calling and voice control, in-cabin device syncing and tire pressure monitoring systems (TPMS). New capabilities like Bluetooth channel sounding enhance secure access by precisely measuring proximity, which is crucial for digital keys and anti-theft systems. But testing Bluetooth in automotive applications demands attention to low-latency performance, interoperability with various mobile devices, and challenging antenna design constraints exacerbated by the vehicle’s limited space.

 

  • Digital keyless entry often combines ultra-wideband (UWB), Bluetooth and near field communication (NFC) to enable secure, seamless vehicle access. Testing must account for complex factors like precise location accuracy (to avoid relay attacks), secure cryptographic protocols and sensitivity to physical placement and interference. NFC’s short range makes testing particularly sensitive to repeatability and precision in real-world conditions.

 

  • Cellular connectivity via 4G, and increasingly 5G, is the backbone for telematics, emergency services and vehicle-to-cloud functions. Automotive test strategies must account for whether connected modules are pre-certified or chip-on-board, with the latter requiring significantly more complex and expensive RF validation.

 

Overall, while consumer adoption helps reduce the cost of wireless components, automotive applications demand far more robust and specialized test strategies to meet the industry’s stringent requirements for safety, reliability and performance.

Why Car Makers Need a Wireless Test Partner

Unlike the consumer electronics industry, cost-sensitive and risk-averse automotive manufacturers are slower to invest in leading-edge wireless testbeds. As a result, some OEMs and Tier 1 suppliers rely on outdated methods due to limited internal RF test knowledge.

To close the gap, a wireless test partner such as LitePoint can simplify wireless test for labs and factories and provide the capability to test the latest wireless products with innovative test methodologies.

As a trusted wireless test company, LitePoint provides high-quality, efficient and reliable test solutions. LitePoint is also the only test company to cover all three critical technologies in the Car Connected Consortium (CCC) digital key 3.0 specification – UWB, Bluetooth and NFC.

Driving Performance with Trusted Test Partnerships

As vehicles incorporate more software-defined technologies, wireless testing is no longer optional; it’s a strategic imperative.

With turnkey systems for both R&D and high-volume production, LitePoint enables automotive customers to implement scalable test processes that align with design complexity and compliance requirements. LitePoint brings deep domain expertise, strong chipset vendor relationships and leading-edge test automation. By standardizing and simplifying wireless test across design and production workflows, LitePoint helps automotive manufacturers close the gap between wireless ambition and real-world performance.

Connected vehicles require wireless systems that just work. LitePoint helps ensure they do.

Your guests judge your hotel within minutes of arrival. They expect lightning-fast WiFi, seamless entertainment systems, and flawless connectivity throughout their stay. But behind every five-star digital experience lies something most hoteliers never think about: the cables hidden in your walls.

In this white paper, we discuss structured cabling, which takes a standardized approach to network infrastructure. Instead of chaotic point-to-point connections, it creates an organized foundation that supports multiple technologies through a single, unified system.

Learn about the benefits of structured cabling versus traditional piecemeal approaches, how you can implement structured cabling in your hotel, and how WorldVue and our partner Cabling Solutions LLC can help with comprehensive structured cabling solutions for both new construction and retrofit projects.

 

Download Full White Paper

Simplify Network Operations Without Sacrificing Scalability: Smarter Network Design Starts Here. RUCKUS offers a free Design My Network session where you’ll connect with an expert, talk through what’s working (and what’s not), and walk away with a personalized plan that makes sense for your environment.

If you’re responsible for keeping your organization connected—whether it’s across a campus, a hospital system, a stadium, or multiple offices—you’re likely being asked to do more with less. More users. More devices. More expectations. Meanwhile, resources and headcount aren’t exactly scaling at the same pace.

And let’s be honest – Most networks weren’t built for this kind of pressure.

If your infrastructure feels like a patchwork of short-term fixes and aging equipment, now might be the time to rethink your network design. Not by starting from scratch, but by building something smarter, more scalable, and easier to manage.

When the Network Becomes the Bottleneck

It’s easy for networks to grow in pieces—one vendor here, a new platform there—until suddenly you’re juggling more tools than time. The result? Slower performance, inconsistent coverage, and troubleshooting that eats up half your day.

Simplifying how your network is built and managed frees up your team to focus on actual priorities—not putting out fires or dealing with user complaints. And the right design doesn’t just solve problems now. It gives you a solid foundation for what’s coming next.

Why Automation Isn’t Optional Anymore

If your team is constantly buried in manual tasks, configuring access, onboarding devices, managing security — you’re not alone. But you also don’t have to live that way.

When automation is baked into the network from day one, things just work better. Devices get online faster. Policies stay consistent. Problems get flagged before they become support tickets.

Done right, automation isn’t about replacing people;, it’s about giving them back time and control. And when your team has space to be strategic instead of reactive, everything runs smoother.

Build It for How You Actually Work

There’s no one-size-fits-all when it comes to networks. What works for a hospital won’t work for a convention center. What works for your HQ might not work for your branch offices.

That’s why design matters. It should reflect how your people work, how your spaces are used, and where performance matters most. A smart design makes it easier to scale, maintain, and support without overcomplicating things.

When changes occur, as they inevitably do, you are prepared to adapt without needing to start from scratch.

See What’s Possible 

Take Dr Pepper Ballpark. RUCKUS helped them rethink their network to handle packed stadium crowds, on-site vendors, and day-to-day operations—without adding complexity for their IT team.

Better coverage, smoother performance, less stress. That’s what smart design looks like.
Read the full story if you’re curious.

Let’s Make Your Network Easier to Run

If your current setup feels clunky or like it’s barely holding on, this is a chance to reset. RUCKUS offers a free Design My Network session where you’ll connect with an expert, talk through what’s working (and what’s not), and walk away with a personalized plan that makes sense for your environment.

No pressure. Just a chance to figure out what “smarter” could look like for you.

From manufacturing to health care facilities, edge computing is a powerful technology reshaping how industries handle data and streamline operations. This is made possible through Internet of Things (IoT) devices, such as sensors, cameras and specialized processors embedded at the edge. The market value of IoT-enabled devices is projected to increase to $6.5 billion in 2030, which is a growth of over $4 billion compared to 2020.

This guide explores the key benefits of edge computing and IoT technology.

What Is Edge Computing in IoT? 

Edge computing with IoT technology involves processing data closer to where it’s generated, which is at the network’s edge. Instead of sending every bit of information online to a distant cloud server, IoT edge devices analyze and process the data locally. This localized processing helps minimize delays, improves responsiveness and reduces the burden on bandwidth, which is crucial as IoT deployments continue to scale.

IoT devices such as sensors and smart appliances gather data in real time. With edge computing capabilities, IoT devices function as nearby edge gateways that filter, analyze and respond to data instantly.

For example, when an industrial sensor detects component damage or overheating, it prompts equipment to shut down without waiting for instructions from the cloud. Edge computing improves how IoT systems operate by enabling faster, smarter and secure device interactions while keeping critical processing close to where the computing happens.

Edge Computing and Cloud Computing in IoT 

Edge-enabled IoT devices also sync with the cloud for long-term storage and coordination across systems. This hybrid approach ensures fast local action while keeping the broader IoT ecosystem connected and intelligent.

Cloud computing centralizes data processing in large-scale data centers, which is ideal for massive data storage, analytics and long-term decision-making. In contrast, edge computing processes time-sensitive data near the device itself, enabling more real-time responses. Edge computing in IoT devices complements the cloud by reducing lag, preserving bandwidth and enhancing local autonomy.

Traditional cloud-first models struggle with latency, network instability and data overload. By shifting some computing power to the edge, businesses can overcome these limitations and enhance efficiency, reliability and data security. For example, sensitive health data in hospitals can be processed locally with less risk of cyber threat than data that is routinely transmitted over networks.

How Is Artificial Intelligence Used With IoT in Edge Technology?

Artificial intelligence (AI) and machine learning (ML) have a critical role in how IoT devices process data at the edge. While traditional cloud computing relies on centralized servers for data processing, edge computing also performs AI and ML tasks directly on IoT-enabled local devices.

This decentralized approach enables real-time analysis and faster decision-making, even without internet connectivity. From security cameras to wearable health trackers, edge AI allows data to be processed where it’s generated for more immediate insights. As more industries demand timely insights, AI, ML and IoT are driving innovation in many functional and efficient ways. For example, smart traffic lights with IoT sensors or cameras analyze data and adjust traffic signals according to vehicular flow.

The Benefits of IoT Edge Computing

Edge computing offers a wealth of advantages for IoT technology, improving how businesses operate while enabling richer interactions with connected devices. The following are key benefits of employing edge-based IoT technology:

  • Reduces latency: Edge computing significantly minimizes processing delays by computing data close to IoT devices. Local data handling eliminates latency that occurs when information has to travel to and from an online cloud server. This is especially important for applications like smart surveillance systems and automated industrial equipment. With edge computing and IoT, immediate decision-making becomes possible and practical, allowing systems to respond to environmental inputs and operational changes in a timely manner.
  • Lowers energy costs: When data is processed and filtered locally, it reduces the need for constant, high-volume transmission to the cloud. It also decreases reliance on a centralized IT infrastructure and costly cloud service charges. The result is lower bandwidth consumption and less power usage across the network. Edge devices also optimize power through such techniques as sleep modes, adaptive processing and task prioritization, which are beneficial for battery-powered IoT sensors in remote areas.
  • Enables real-time tracking and analytics: Edge computing and IoT enable immediate data monitoring, making it suitable for time-sensitive applications such as predictive maintenance, asset tracking and remote monitoring. Whether it’s identifying early signs of equipment failure or adjusting environmental controls for smart buildings, decisions can be made the moment data is collected. This improves a company’s operational efficiency, safety and responsiveness.
  • Enhances data security: One of the core benefits of edge computing with IoT devices is its ability to boost data security. By processing regulated data locally, businesses reduce the risk of exposing data during cloud transmission. Edge devices also implement built-in encryption, authentication protocols and access control at the source. This layered security approach makes edge computing with IoT valuable in industries such as finance, healthcare and critical infrastructure, where data breaches can result in serious damage and regulatory penalties.
  • Leverages IoT machine learning: By integrating IoT with ML and AI, edge computing allows intelligent algorithms to operate directly at the source of the data. From smart homes that learn user preferences to industrial sensors that detect equipment defects, edge-based AI and ML process raw data and provide actionable steps. This is essential for time-sensitive smart data analysis and predictive modeling without cloud dependency.

 

  • Provides scalability solutions: As IoT networks expand, centralized processing can quickly become a bottleneck for cloud data transfers. Edge computing distributes the processing workload across multiple local nodes, making it easier to manage and scale infrastructure. This architecture allows organizations to add more devices and handle more data without compromising performance or overwhelming core systems, which is beneficial for growing edge computing capability with IoT ecosystems.
  • Boosts network reliability: Since most data processing occurs locally, the network becomes more resilient with IoT edge computing technology. This means computer systems continue to function even if cloud connectivity is lost or delayed. Network reliability is essential for mission-critical operations in industries like manufacturing, transportation and agriculture, where downtime is costly and continuous operations are a priority.

 

Eight Use Cases for Edge Computing in IoT

Edge computing with IoT technology is revolutionizing the way we interact with the world around us. Here are eight examples of how it’s being used in real-world applications:

 

1. Predictive Maintenance

Edge computing enables industrial IoT devices to continuously monitor equipment conditions at any time. By analyzing data like temperature, vibration and energy consumption at the source, businesses detect early signs of heavy equipment failure. This allows companies to schedule maintenance before breakdowns occur, reducing downtime and extending the machinery’s lifespan while streamlining operations.

2. Remote Monitoring for Jobsites

Edge computing enhances remote monitoring by enabling IoT devices to process and act on data locally. This is essential in hard-to-reach jobsites, such as oil rigs, rural cell towers and distant wind farms. Edge-enabled sensors flag discrepancies or safety risks immediately without waiting for cloud-based analysis, ensuring quicker response times and operational reliability.

3. Smart Grids

IoT in edge computing helps connect and modernize smart grids, which include power plants and substations. Smart grids enhance traditional electricity networks through the integration of digital technologies, sensors and software. This innovative approach enables precise and time-sensitive management of electricity supply and demand, resulting in reduced costs and improved grid reliability.

Edge-enabled smart meters and sensors placed throughout the grid collect real-time data, such as energy consumption, load balancing and equipment status. The data is processed locally, enabling improved decision-making for efficient energy distribution, fault detection and reduced utility costs.

4. Connected and Autonomous Vehicles

A standout application of edge computing in IoT is autonomous vehicles. Self-driving cars rely on local data processing to respond instantly to road conditions, traffic and obstacles. By analyzing input from onboard sensors such as cameras and light detection and ranging scanners, these vehicles make split-second decisions without depending on external networks. This enables efficient route optimization, improved fuel usage and enhanced safety, making edge computing instrumental in the development of smart transportation.

5. Health Care

Edge computing is transforming health care experiences through smart IoT devices. Wearables and remote monitoring systems gather data to track vital signs such as heart rate and transmit it securely for immediate analysis. Critical alerts can be generated locally on the device, allowing medical professionals to intervene right away. This setup supports telehealth services, chronic disease management and personalized care, even in clinics with limited bandwidth and connectivity.

 

6. Supply Chain Management

Edge computing in IoT delivers end-to-end visibility across supply chains. Radio frequency identification (RFID) tags, global positioning system (GPS) trackers and environmental sensors placed on goods and transport vehicles provide location and condition data. For example, it helps supply chains reroute shipments during delays or prompt alerts for temperature breaches, enabling more agile logistics and improved quality control.

7. Farming and Environmental Monitoring

IoT sensors deployed in agriculture and environmental science collect data on soil conditions, air and water quality and weather patterns. Real-time monitoring is possible without the need for constant connectivity. This leads to more efficient farming practices and proactive environmental management. Edge computing with IoT devices enables sensors to analyze areas locally and act instantly, such as activating irrigation systems or prompting air quality warnings.

8. Augmented Reality and Virtual Reality

IoT in edge computing significantly enhances augmented reality (AR) and virtual reality (VR) experiences by reducing latency and bandwidth strain. Improved responsiveness allows AR and VR tools to adapt instantly to the user’s physical environment and even function offline, providing more powerful applications that were once limited by cloud-based delays. Applications such as virtual product demos, AR-based maintenance instructions or immersive training simulations benefit from timely responsiveness when data is processed close to the user.

Five Types of Edge IoT Devices

Edge computing enhances a wide range of IoT devices for different purposes. Let’s take a closer look at some of these IoT devices:

 

 

1. Sensors

Depending on the operation, IoT sensors capture on-the-spot data such as temperature, pressure, humidity and motion. These devices gather localized data and process it at nearby edge nodes, enabling rapid decision-making without relying on a central cloud. For example, in manufacturing, vibration and thermal sensors detect early signs of equipment failure, prompting repair notifications to prevent breakdowns.

IoT-ready wireless connectivity solutions are also improving the performance and responsiveness of edge IoT devices with sensors. In smart homes, motion sensors adjust lighting dynamically, enhancing comfort and energy efficiency. This boosts applications that require timely data processing and control, including security systems and home automation.

2. Cameras

Smart cameras are evolving beyond capturing images. With edge computing in IoT, cameras process and analyze footage directly where it’s captured. This reduces latency, offloads network traffic and enables immediate action. For example, in a smart city, edge-enabled cameras detect unusual activity and trigger alerts without sending large video data to a central cloud.

In retail, cameras analyze shopping movements to optimize store layouts. At industrial sites, cameras use edge intelligence to monitor production lines and flag possible issues the moment they occur. Integrated with AI, smart cameras support facial recognition, license plate reading and crowd analytics, all with minimized data transfer.

3. Monitors

Edge-enabled IoT monitors are used to track key markers such as energy usage, air quality, fluid levels and machine performance. Whether it’s optimizing heating and cooling systems or flagging irregularities in water treatment plants, IoT monitors provide a critical layer of operational visibility.

In industrial settings, these devices combine sensor data with edge processing to deliver responsive insights that drive efficiency. For example, they enable predictive maintenance by identifying subtle signs of wear and tear. IoT monitors are also used in smart energy systems to identify consumption peaks and automatically adjust settings to minimize costs.

4. Drones

Drones integrated with edge IoT capabilities are transforming industries that require inspections, surveillance and deliveries in hard-to-reach areas. These airborne edge devices use cameras, sensors and onboard processors to collect and analyze data during flight without relying on cloud uploads.

In energy and utility sectors, drones inspect remote equipment like wind turbines or oil pipelines, relaying condition updates to technicians. In warehouses, drones assist in inventory checks and maintenance inspections. They also enable ultra-fast deliveries, bypassing traffic and reaching remote locations in emergencies.

5. Controllers

At the core of IoT in edge computing are controllers, which are smart systems that manage, automate and secure networks of connected devices. These controllers integrate sensor inputs, camera feeds and actuator outputs to make intelligent, localized decisions.

For example, a smart controller reads room temperatures from multiple sensors and instantly adjusts airflows. In industrial factories, energy management controllers monitor equipment and optimize power usage. Their ability to automate workflows and coordinate diverse devices makes them essential devices in homes, offices and industrial environments. With built-in security features and local processing power, IoT edge controllers help ensure reliability and minimize operational downtimes.

 

Connect With Synaptics Today for IoT Edge Device Design

Edge computing in IoT addresses the limitations of cloud-based IoT, enabling devices to overcome bandwidth, latency and security challenges. Partner with Synaptics for the latest IoT edge device design solutions, such as Synaptics Astra™ AI-native IoT processors powered by open software and outstanding wireless connectivity for a secure, multi-modal device edge.

Synaptics is a trusted leader in AI and edge technology, delivering reliable, high-performance solutions that make connected devices smarter and more efficient. We help industries create secure, intuitive digital experiences that transform how users engage with intelligent connected devices. We also provide customized multimedia compute solutions with a unified AI framework that rapidly deploys to edge devices. From smart homes to workplaces, we specialize in engineering exceptional experiences that drive the next wave of digital transformation.

Take advantage of the latest IoT edge computing technology. Contact us today to discover how our IoT edge device design solutions enhance your business.

Linearity of a receiver plays an important role in the overall performance of the system. It affects system-level parameters like link budget and receiver sensitivity. Designing high-performance RF receivers involves navigating the classic tug-of-war between two critical specifications: Noise Figure (NF) and input linearity, typically expressed as input third-order intercept point (IIP3). Lower NF improves sensitivity but often requires higher Gain. Conversely, high IIP3 supports better tolerance to interference but usually comes at the cost of reducing Gain.

This article explores the interaction between these specifications and how receiver linearity performance is critical to the overall system performance. We’ll also cover key concepts in intermodulation distortion, the difference in linearity optimization for transmitters versus receivers, and how tools like Error Vector Magnitude (EVM) can simplify the overall system-level trade-offs.

A Brief Review of Linearity

Linearity is the ability of the amplifier to produce an output signal that is a linear representation of the input signal. Having a good linear signal helps maintain signal integrity. Non-linearity can lead to distortion, intermodulation products and spectral regrowth – all of which can lead to a degraded signal quality in the system. Linearity performance is measured at medium signal levels, i.e., the system is not driven into compression, but intermodulation distortion is generated.

Second- and Third-Order Intercept & IMD Products

The second and third-order intercept-point (IP2 and IP3) products are widely used to benchmark the linear performance of an RF system. IIP3 refers to the hypothetical input power level where the power of third-order intermodulation products (IMD) equals the power of the fundamental output signal. To fully understand IP3, it’s important to know the difference between Input IP3 (IIP3) and Output IP3 (OIP3).

IIP3 is the signal power going into the device at the intercept point, while OIP3 is the signal power coming out of the device at that same point. Both values help us understand how an RF device performs when dealing with different input signal levels. Understanding the relationship between Input IP3 (IIP3) and Output IP3 (OIP3) can provide invaluable insights into the behavior of RF devices and systems. OIP3 can be calculated using the formula:

 

OIP3(dBm) = Gain(dB) + IIP3(dBm)
 

Before the IP3 point is reached, the device usually hits saturation. This means it can’t increase its output power in a straight linear line anymore, no matter how much you raise the input. The result is signal compression and distortion. See Figure 1.

 

 

 

Figure 1: IP2, IP3 and P1dB
 

Two-tone tests shown in Figure 2 can be performed to measure IP2 and IP3. Two closely spaced sinusoidal signals (Fundamental Tones) with equal power level input power (Pin) at frequencies f1 and f2 are applied, and the level of the second order IMD (IM2) at frequencies |f2 – f1|, 2f1 and 2f2 is then observed as Pin increases. For third-order intermodulation products, the two-tone frequencies are chosen such that the IM products fall inside the received signal band. For IP3, the IM3 components at |2f1−f2| and |2f2−f1| are observed as Pin increases.

 

 

 

Figure 2: Fundamental, IMD products.
 

The IP3 is a result of third-order IMD products, specifically |2f1 – f2| and |2f2 – f1| as shown above. The third-order intercept is a figure of merit that characterizes an RF receiver’s tolerance when subjected to multiple RF signals within the desired passband. These IMD products apply to both transmit and receive sides of the RF system. But optimizing each end of the system requires slightly different approaches.

As shown in Figure 3 below, the graph illustrates two distinct slopes that represent different regions of amplifier behavior. In the initial linear region, the output power increases proportionally with input power at a 1:1 slope, reflecting the ideal, distortion-free operation of the device where the fundamental signal dominates. This region indicates that the amplifier operates within its linear dynamic range, preserving signal integrity.

However, as the input power continues to rise, the device transitions into a non-linear region where intermodulation distortion products, particularly third-order (IMD3), begin to emerge. In this region, these distortion components increase at a much faster rate than the fundamental, with a characteristic slope of 3:1. This means that for every 1 dB increase in input power, the IMD3 components grow by 3 dB, quickly surpassing the desired signal and degrading overall performance. Understanding these two regions and their respective slopes is essential for accurately characterizing the linearity of RF components and for predicting how they will perform under real-world signal conditions.

 

 

 

Figure 3: IP3 with input power versus output power slopes.
 

 

Optimizing the Receiver via IIP3

 

Designing RF receivers with a focus on output linearity can be misleading, as it often focuses on Gain that ultimately compromises receiver performance.

While higher output Gain helps maximize output linearity (like one would do for a transmitter), it simultaneously degrades input linearity by making the system more susceptible to compression from smaller input signals, as shown in Figure 4. The data shows that optimal receiver performance requires a careful balance between input-referenced linearity and NF. Lower NF is achieved with higher front-end Gain, but this comes at the cost of reduced input linearity. Conversely, maximizing input linearity calls for lower overall Gain, which can raise the NF. Therefore, effective receiver design demands trade-offs between these parameters, so neither parameter is sacrificed excessively. Designers must avoid the transmitter-oriented mindset that more Gain is inherently better and instead select a Gain profile that optimally balances NF and IIP3 for the intended application.

 

 

 

Figure 4: Cascaded RF receiver parameter data comparison.
 

But why is the receiver linearity IIP3 so important? It is because it directly impacts the receiver’s ability to handle multiple signals and prevent IMD. A high IIP3 in a receiver design indicates the receiver is more linear and therefore can better separate designed signals from unwanted IMD products.

It is important for system designers to understand the influencing factors such as Gain, NF, OIP3 and IIP3. Balancing the trade-offs between these parameters is critical to ensuring the overall system is optimized.

EVM and the Bathtub Curve Explained

 The manifestation of Error Vector Magnitude (EVM) (a measure of the difference between an ideal signal and the actual received signal in digital communication systems) can come from many different sources. The EVM figure of merit is typically a combination of noise and distortion. As shown in Figure 5, at lower power levels, noise tends to dominate, and the EVM increases as we lower the power. At high power levels, distortion tends to dominate, and we observe increased EVM as we increase power levels. In the middle, we often see the minimum EVM, and so the overall shape resembles a bathtub curve. As such, the EVM bathtub curve becomes an essential visualization tool for system-level optimization, offering a comprehensive view of how different impairments jointly impact overall performance.

And yes, we’re keeping an eye on the next wave, too. AI-powered radar? Ultra-wideband-enabled machine vision? BAW and SAW filters that are smaller, faster and even smarter than the systems they serve? That’s all in our wheelhouse.

 

 

 

Figure 5: Bathtub EVM curve (EVM vs. Operating Power).
 

While most frequency devices exhibit low phase noise below 2 GHz, this advantage diminishes at higher frequencies and wider signal bandwidths, where integrated phase noise can increase substantially and degrade system performance. This challenge is particularly acute in millimeter wave (mmWave) systems operating above 20 GHz, where elevated phase noise directly contributes to higher EVM.

To address this, system-level design often begins with cascade analysis, using low-level performance metrics of individual components to estimate overall system behavior. In these situations, EVM serves as a highly practical system-level performance metric, enabling engineers to consolidate the effects of multiple impairments, such as noise, non-linearities and phase noise, into a single optimization target. Rather than tuning several individual parameters, designers can focus on minimizing the root-mean-square (rms) EVM value for an efficient and effective design process. This optimization is visually aided by the EVM bathtub curve.

The System Advantages of Optimized Receiver Linearity

High linearity in an RF receiver is critical for maintaining robust system performance in the presence of strong or closely spaced signals. One of the primary benefits of high receiver linearity is an expanded spurious-free dynamic range (SFDR), which quantifies the usable signal range before intermodulation products rise above the noise floor. SFDR is directly proportional to the IIP3 and inversely proportional to the noise floor (No), with the relationship defined by this formula.

SFDR = (2/3) × (IIP3 − No)
A higher IIP3 indicates better tolerance to interference, as third-order intermodulation (3IM) spurs decrease by approximately 2 dB for every 1 dB increase in IIP3. However, achieving high linearity often comes at the expense of a higher NF, presenting a well-known design trade-off. Systems that support adjustable receiver input linearity—based on real-time assessment of incoming signal strength—offer the most flexibility. In RF environments with low interference, higher receiver Gain can be used to minimize NF, even though this reduces input linearity. Conversely, in high-interference scenarios, reducing Gain improves input linearity, allowing the receiver to better tolerate strong unwanted signals. For receivers, it’s important to focus on input linearity rather than output metrics like OIP3, as high OIP3 can misleadingly suggest good receiver performance while input signal handling capability may be poor.

Conclusion

In summary, optimizing RF receiver performance demands a careful balancing act between input linearity and NF, as both parameters critically influence system sensitivity, interference tolerance and overall signal fidelity. While conventional design approaches may prioritize output metrics like OIP3 or default to maximizing Gain, this can lead to suboptimal trade-offs, particularly in dynamic RF environments. A focus on IIP3, supported by tools like EVM analysis and cascade modeling, enables more accurate predictions of receiver system behavior and better-informed design choices. Whether operating in low-interference conditions that favor higher Gain and lower NF or in high-interference scenarios that require improved linearity through Gain reduction, adaptive receiver architectures that dynamically adjust to their signal environment offer the greatest design flexibility. By embracing a system-level perspective that accounts for all impairments, including intermodulation, compression, noise and phase distortion, engineers can achieve receiver designs that are both robust and efficient across a wide range of use cases.

Additionally, you can find more interesting collateral on this subject by visiting our Qorvo Design Hub for a rich assortment of videos, technical articles, white papers, tools and more.

For technical support, please visit Qorvo.com or reach out to Technical Support.

About the Authors

Our authors bring a wealth of technical expertise in developing and optimizing high-performance RF receiver solutions and systems. With a deep understanding of customer needs and industry trends, they collaborate closely with our design teams to drive innovation and deliver cutting-edge solutions that support industry-leading products.

Thank you to our main contributors of this article: David Corman (Chief Systems Architect) and David Schnaufer (Corporate, Technical Marketing Manager) for their contributions to this blog post, ensuring our readers stay informed with expert knowledge and industry trends.

Dave’s Thoughts: Those Left Behind: Living in an Apartment Building Shouldn’t Mean Second-Rate Broadband.

Apartment complexes, large and small, are still falling behind when it comes to modern broadband connectivity. Many buildings haven’t had an Ethernet upgrade since they were first constructed. While there has been an emphasis on higher speeds to the single-family home, MDUs haven’t received the same attention.
Given the average broadband speed has exploded from a few Mbps to over 100Mbps in just the last ten years, it is safe to say most buildings are not meeting today’s demands.
This is compounded by legacy cable deployments. Today’s tenants expect connectivity in places like the laundry room, by the pool, in common areas, and throughout the property. Yet many of these locations have no infrastructure at all.

So, what are your options?

  • Squeeze more bandwidth out of your existing cable plant
    If it’s twisted pair, maybe more advanced modems can help, but they are hugely impacted by the quality of the copper and any taps, splitters or combiners on a given strand.
  • Re-wire your entire building
    Clearly this involves time, money and massive tenant disruption. And what if your floor-to-floor riser is full, or not in the right place? Drill baby drill through that cement. That means more time, more money and more permitting.
  • Deploy a wireless ethernet backbone to quickly and easily upgrade your infrastructure
    No construction. No drilling. No permits. No asbestos headaches. Just quick deployment and scalable, reliable connectivity to meet today’s demands

After Covid everyone was confronted with how much we all rely on ubiquitous high-speed connectivity and the integration of IoT into our daily lives. For single-family homes, the market responded with urgency and innovation. But in MDUs, tenants are still at the mercy of the property owner.

Now is the time for building owners and management companies to seize the opportunity to modernize their properties with wireless Ethernet backbone systems—without the cost, disruption, or delays of traditional rewiring. This is more than just an upgrade; it’s a competitive edge. Fast, reliable broadband is no longer a perk—it’s a priority. Invest in your infrastructure today to retain residents, attract new ones, and future-proof your property.