Blogs
The Quiet EV Paradox: Why Electric Cars Are Actually "Noisier" It sounds like a paradox — electric vehicles have no roaring engine, yet engineers are finding it harder than ever to achieve a truly quiet cabin. The truth is, when the low-frequency masking effect of the internal combustion engine disappears, every previously hidden noise becomes fully exposed: the high-frequency whine of the electric motor, the electromagnetic hum of the inverter, gear meshing vibrations, wind noise, road noise, even the squeak and rattle of interior trim — nothing can hide anymore. This isn't just a comfort issue. It's fundamentally redefining the automotive industry's approach to NVH (Noise, Vibration, and Harshness) testing. The global automotive NVH testing market is projected to grow from USD 3.51 billion in 2026 to USD 5.75 billion by 2034, at a CAGR of 6.4%. The core driver behind this growth? The electrification revolution. What New Noise Challenges Do EVs Bring? A Fundamental Shift in Frequency Range Traditional ICE vehicle NVH work focuses on the 20–2,000 Hz low-frequency range — engine firing, exhaust systems, crankshaft vibrations. Electric vehicles are fundamentally different: Noise Source Typical Frequency Range Characteristics Electric motor electromagnetic noise 500–5,000 Hz Sharp tonal noise, varies linearly with speed Inverter switching noise 4,000–10,000+ Hz High-frequency hum, related to PWM frequency Gear meshing noise 800–3,000 Hz Particularly prominent in single-speed reducers Battery charger noise 8,000–20,000 Hz Near-ultrasonic range, at the edge of human perception Wind / Road noise 200–4,000 Hz Highly exposed without engine masking ICE vs EV: The fundamental shift in noise frequency characteristics Key insight: EV noise problems shift from low frequencies to mid-high frequencies (and even ultrasonic ranges). The 100Hz-5kHz range is where most critical NVH issues reside—precisely where human hearing is most sensitive. Traditional NVH testing methods and frequency ranges may no longer be sufficient. New Noise Sources, New Localization Challenges In the ICE era, the assumption that "the engine is the dominant noise source" made things relatively straightforward. In EVs, noise sources become more distributed and complex: Electric drive system: The motor + inverter + reducer form a highly coupled noise system Thermal management: Battery cooling pumps and fans become dominant noise sources at low speeds Regenerative braking: Changes in inverter operating modes during energy recovery produce transient noise Structural transmission paths: Lightweight body structures (aluminum alloy, carbon fiber) have fundamentally different sound insulation characteristics compared to traditional steel This means engineers face a core challenge: How do you quickly and accurately locate the root cause among multiple distributed, dynamically changing noise sources? Sound Quality Design: From "Reducing Noise" to "Crafting Sound" NVH engineering in the EV era is no longer just about "minimizing noise." Consumers expect a carefully designed sound experience: Acceleration should feel "high-tech" without being harsh The cabin should be quiet, but not so silent that it makes the driver uneasy Different driving modes (Sport / Comfort / Eco) should deliver differentiated acoustic feedback This demand for "Sound Design" is expanding NVH testing from pure engineering validation into subjective sound quality evaluation and brand-level acoustic identity. Why Acoustic Cameras Are Becoming Essential for EV NVH Facing these new challenges, traditional NVH testing tools — single-point microphones, accelerometers — remain important but are no longer sufficient for every scenario. Acoustic cameras are filling this gap. Core Advantages of Acoustic Cameras 1. Real-Time Noise Source Visualization Traditional methods require densely placing microphone arrays on the target object — time-consuming and labor-intensive. Acoustic cameras use beamforming technology to generate a noise source heatmap in a single capture, instantly showing "where the noise is and how loud it is." Typical scenario: An EV prototype running on a test bench, the acoustic camera aimed at the electric drive system, instantly revealing that an 800 Hz resonance originates primarily from the right side of the motor — the entire localization process takes less than 5 minutes. Engineer conducting noise source localization test Automotive NVH detection and optimization 2. Wide Frequency Coverage EV noise spans from hundreds of hertz (gear meshing) to tens of thousands of hertz (inverter switching noise) — an enormous frequency range. Critical consideration for NVH: Most EV noise issues occur in the 100Hz-5kHz range—gear meshing, motor electromagnetic noise, wind leaks, HVAC systems. Traditional acoustic imaging cameras (limited to frequencies above 5 kHz) cannot capture these noise sources. Take the CRYSOUND SonoCam Pi (CRY8500 Series) as the ideal example: its 208 MEMS microphone array provides: Beamforming frequency range: 400 Hz - 20 kHz (covers the entire NVH audible spectrum) Near-field acoustic holography range: 40 Hz - 20 kHz (captures low-frequency road noise and structural vibration) Array size: >30 cm (optimized for low-frequency spatial resolution) This makes SonoCam Pi uniquely suited for full-spectrum EV NVH testing—from low-frequency road noise to high-frequency motor whine, all in a single handheld device. 3. Non-Contact Measurement EV electric drive systems are highly integrated and spatially compact. The non-contact measurement approach of acoustic cameras means: No disassembly of any components required No interference with the operating state of the system under test Rapid quality inspection directly on the production line 4. Portability Modern handheld acoustic cameras like the SonoCam Pi can be taken directly to proving grounds, production lines, or customer sites, no complex setup required. Typical Application Scenarios in EV NVH Scenario Application E-drive system NVH Locating order-based noise contributions from motors, inverters, and reducers Pass-by noise testing Analyzing noise source distribution as vehicles pass by Interior squeak & rattle tracking Locating noise from dashboards, doors, seats, and trim End-of-line production QC Rapid online detection of abnormal noise, replacing subjective human judgment Wind tunnel / Semi-anechoic chamber High-precision noise source localization and sound power analysis Real-World Case Study: OEM Dynamic Road Testing Client: A leading Chinese OEMLocation: An OEM test center, internal test trackObjective: Identify in-cabin noise sources during dynamic driving conditions CRY8500 Series SonoCam Pi acoustic cameras Test Setup Device:SonoCam Pi acoustic camera Measurement positions:Rear seat and front passenger seat Target areas:Left and right B-pillars (rear cabin area) Test mode:Beamforming app Frequency range:3,550 Hz - 7,550 Hz Dynamic range:5 dB Key Results SonoCam Pi successfully localized noise sources in real-time during vehicle motion, providing actionable data for OEM's NVH engineering team. The test demonstrated: Real-time localization during dynamic conditions: Unlike fixed laboratory setups, SonoCam Pi captured noise distribution while the vehicle was in motion on the test track Precise frequency-band analysis: By focusing on the 3,550-7,550 Hz range (critical for perceived cabin noise), engineers pinpointed specific contributors rather than measuring overall SPL Rapid testing workflow: Complete B-pillar area scan in minutes, not hours Noise Source Localization Results Key Insight: Traditional microphone arrays would require the vehicle to be stationary in a semi-anechoic chamber. SonoCam Pi enabled on-track diagnostics, dramatically reducing testing time and enabling rapid iteration during vehicle development. Future Trends — What's Next for EV NVH Testing? AI-Driven Noise Classification Machine learning is being integrated into NVH testing workflows: automatically identifying noise types, determining whether anomalies exist, and predicting potential quality issues. The high-dimensional data captured by acoustic cameras is naturally suited for AI analysis. Digital Twins and Simulation-Test Integration Simulation (CAE) predicts noise performance → Acoustic camera validates through physical measurement → Data feeds back to optimize the simulation model. This closed-loop approach is becoming the standard workflow for major OEMs. New Challenges in the Solid-State Battery Era Solid-state batteries have different mechanical properties compared to liquid lithium-ion batteries. Their vibration transmission characteristics and thermal management approaches will introduce new NVH challenges. Stricter Regulations Pass-by noise testing is the fastest-growing NVH sub-segment (CAGR 7.11%), with UNECE pushing for stricter standardized testing requirements, including indoor pass-by testing protocols. Conclusion: The Value of Acoustic Testing, Redefined for the EV Era Electrification hasn't made cars quieter — it has made noise challenges more complex, more nuanced, and more valuable to solve. For automotive OEMs, Tier 1 suppliers, and testing service providers, investing in the right NVH testing equipment is no longer a "nice-to-have" — it's foundational infrastructure for competitiveness. Acoustic cameras—especially those capable of capturing the critical 100Hz-5kHz NVH frequency range—are evolving from "useful auxiliary tools" to "indispensable standard equipment." The CRYSOUND SonoCam Pi stands out as the only handheld acoustic camera that combines: Low-frequency capability (400 Hz beamforming, 40 Hz holography) High spatial resolution (208 microphones, >30 cm array) Near-field + far-field measurements in a single system Portability (handheld, <3 kg, production-ready) Learn more: CRYSOUND SonoCam Pi (CRY8500 Series) → Contact us for NVH testing solutions →
In many practical applications, data acquisition is not performed in an “ideal laboratory” environment. The device under test may be connected to mains power, distribution cabinets, frequency converters, or large electromechanical systems, while the acquisition card on the other side is connected via USB or Ethernet to a computer—sometimes operated directly by a person. These two sides are often not at the same electrical potential. If there is no effective electrical isolation inside the data acquisition card, this potential difference may propagate through signal lines, shields, or ground paths to the system side, leading to measurement distortion, interface malfunction, or even safety hazards. This is the fundamental reason why isolation exists in data acquisition systems. What Is the Isolation Rating of a Data Acquisition Card? In a data acquisition system, the isolation rating is not a simple voltage number, nor is it equivalent to “the voltage that the input can directly withstand.” It describes whether there is a reliable electrical isolation barrier between the measurement side (connected to sensors and the device under test) and the system side (connected to the host computer, communication interfaces, and power supply), and under what level of voltage stress this isolation remains valid. Isolation principle You can think of isolation as a bridge between two islands: The bridge allows information to pass—measurement data, digital communication, control signals. But it blocks dangerous currents—fault currents, ground-loop currents, and energy that could carry high potential to the host side. For this reason, isolation in data acquisition systems typically addresses both safety and measurement stability at the same time. Why Is Isolation Often More Important Than Accuracy Specifications? In many field applications, engineers do not encounter problems such as “insufficient resolution,” but instead: The same system works well in the lab, but noise increases dramatically on site. Once multiple devices are connected together, the data begins to drift. Replacing the computer or using a different power outlet suddenly makes the problem disappear. The common root cause behind these phenomena is often not algorithms or ADC performance, but rather improper handling of electrical potential relationships within the acquisition system. The value of isolation lies precisely here: by breaking unnecessary current loops and limiting the propagation paths of common-mode voltage and fault energy, isolation allows the acquisition system to behave in a controlled and predictable manner even in complex electrical environments. In industry discussions, the core values of isolation usually fall into three categories: signal integrity, safety, and instrument protection. Signal Integrity: Breaking Ground Loops and Improving Common-Mode Rejection Many cases of “inaccurate measurement” are not caused by ADC resolution, but by unwanted currents flowing through ground wires or shields. When the device under test and the host computer, enclosure, or other equipment are at different ground potentials, connecting them via signal cables may form ground loops. Power-line interference and electromagnetic noise then appear as “baseline noise” or ripple in the waveform. Isolation improves this by breaking the current loop paths. Safety: Confine High Potential and Fault Energy to the Measurement Side When measurement points are located near mains power, distribution cabinets, or frequency converters, the real risk is not merely “high voltage,” but where abnormal voltage or fault energy may propagate. If there is no clear electrical isolation between the measurement side and the host side, this energy may travel through signal or ground connections into the computer or communication interfaces, causing equipment damage or safety hazards. Isolation establishes a clear internal safety boundary: high potential and uncertain electrical environments are confined to the measurement side, while the system side—where the host computer and operator reside—remains within a controlled and safe potential range. If an abnormal condition occurs, the problem is contained on the measurement side and does not propagate further. Instrument Protection: A Larger Measurable Window Under High Common-Mode Voltage A non-isolated acquisition system effectively binds the measurement reference ground to system ground or earth. As a result, the measurable input range is centered around earth potential. If the entire signal shifts to a high common-mode potential, the front-end amplifier or ADC may exceed its allowable range or even be damaged. An isolated system allows the measurement reference to “float,” enabling the input measurement window to be centered around the isolated local ground. This permits operation under much higher common-mode voltages, with the ultimate limits determined by the isolation barrier and input protection circuitry together. Commonly Confused Isolation-Related Terms Isolation is often misunderstood because a single term—“isolation voltage”—is used to answer very different questions. The following clarifies these related but distinct concepts. Common-Mode Voltage Common-mode voltage refers to the voltage that is simultaneously applied to both measurement inputs relative to the acquisition system reference ground. It is not the signal of interest. The measurement signal concerns the difference between two input terminals, whereas common-mode voltage describes how high the two terminals are elevated together relative to ground. For example, in battery stacks or floating power systems, the signal itself may be only a few volts, but the entire source may be elevated tens or hundreds of volts above the acquisition card ground. In industrial environments, ground noise or electromagnetic interference may also impose time-varying AC voltage on both measurement leads. These “collectively elevated or oscillating voltages” constitute common-mode voltage. Working Voltage Working voltage is the voltage that can be continuously applied to a device over long periods. It is typically understood as the combination of measured voltage and common-mode voltage, and represents the condition under which the device can operate reliably over time. Withstand Voltage Withstand voltage refers to whether the isolation barrier can survive a very high voltage applied for a short duration without breakdown or damage. To verify this, a dielectric withstand (hipot) test is typically performed. During such a test, a voltage significantly higher than normal operating conditions is applied across the isolation barrier for approximately one minute. If no breakdown, abnormal leakage, or functional damage occurs, the isolation barrier is considered electrically robust. It is critical to note that withstand voltage does not indicate that the device can operate continuously at that voltage. It is a safety and quality verification metric, demonstrating that the insulation will not fail immediately under abnormal or extreme conditions. Input Overvoltage Protection Input overvoltage protection specifies the maximum allowable differential voltage between the positive and negative terminals of the same input channel. Exceeding this limit may damage the input circuitry. This is fundamentally different from isolation withstand voltage: Isolation withstand voltage applies between the measurement side and the system side. Overvoltage protection applies between the positive and negative terminals of the same channel. Measurement Category (CAT) Measurement category defines the severity of transient overvoltage that a measurement system may encounter in its electrical environment. Categories increase from CAT I to CAT IV: CAT I: Low-energy electronic circuits. CAT II: Household appliances and receptacle outlets, typically protected by indoor distribution panels. CAT III: Industrial distribution cabinets and environments with large motors, pumps, or compressors, subject to switching transients and inductive load surges. CAT IV: Outdoor power distribution points exposed to surges and lightning strikes. Pollution Degree Pollution degree describes environmental factors such as dust, moisture, and condensation that affect insulation surfaces. Higher pollution degrees reduce effective insulation performance, requiring higher baseline insulation strength. What Does "1000 V Isolation" Actually Mean? When a specification states “1000 V isolation,” three immediate questions must be asked, otherwise the number has no real comparability: Is it AC or DC? Is it Vrms, Vpk, or Vdc? Is it withstand voltage (short-term) or working voltage (long-term)? What exactly is isolated? Channel-to-ground? Channel-to-channel? Measurement side to USB/host side? The most important takeaway is this: “1000 V isolation withstand” does not automatically mean the system can continuously operate at 1000 V common-mode voltage, nor does it mean that 1000 V can be directly applied to the input. Continuous capability depends on working voltage, measurement category, input overvoltage protection, and the entire system chain including sensors, cables, and terminals. How Isolation Is Implemented: Isolation Barriers and Signal Transfer Methods Isolation is not simply “air separation,” but a combination of structure, materials, and signal-coupling mechanisms. Common isolation signal-transfer methods include: Inductive / Transformer-Based Isolation Inductive isolation transmits energy or information via magnetic fields rather than direct electrical conduction, fundamentally based on Faraday’s law of electromagnetic induction. Inductive isolation chip block Inside the chip, planar coils are fabricated on silicon or within the package, forming transformer-like structures. Transmitter side: current → coil → alternating magnetic field Receiver side: magnetic field variation → induced voltage → signal recovery Advantages include very high common-mode transient immunity (CMTI), high speed, low jitter, long-term stability, and excellent channel consistency. Disadvantages include higher power consumption and cost compared with capacitive isolation. Capacitive Coupling Capacitive isolation uses the “DC-blocking, AC-passing” property of capacitors to achieve voltage isolation, relying on electric-field variation within the dielectric. Capacitive isolation chip block Signal variation → electric-field variation → displacement current coupling Advantages include low power consumption, small die area, high integration, lower cost, and high speed. Disadvantages include higher sensitivity to common-mode dv/dt, stricter PCB symmetry requirements, and higher dependence on reference-ground layout. Optical Isolation Optical isolation uses light as the isolation medium, with air or transparent insulation providing physical separation. The principle is photoelectric conversion plus spatial isolation. Optical isolation chip block Electrical signal → LED emission → photosensitive device → electrical signal Advantages include simple structure, extremely high withstand voltage, good performance for low-frequency and switching signals, and strong EMC characteristics. Disadvantages include slower speed due to device latency, higher variability, and unsuitability for high-precision synchronous systems. Comparison of Isolation Technologies ItemInductiveCapacitiveOpticalWithstand voltage★★★★☆★★★☆★★★★★Transmission speed★★★★☆★★★★★★★Common-mode immunity★★★★★★★★☆★★EMI immunity★★★☆★★★★★★★★★Stability★★★★★★★★★★★Low power★★★★★★★★★★Suitable for DAQRecommendedRecommendedNot recommended A frequently overlooked but critical metric here is CMTI. In high dv/dt environments such as inverters, SiC/GaN power supplies, and motor drives, the issue is often not how high the static common-mode voltage is, but how fast it changes. Rapid high-voltage transients may couple through parasitic capacitances across the isolation barrier, disrupting or corrupting data transmission. Therefore, isolation must withstand not only voltage magnitude, but also voltage transition speed. Common Isolation Topologies in Data Acquisition Before asking whether a DAQ card is isolated, a more important question should be asked: where is the isolation applied? Different products may use entirely different isolation domains, resulting in very different capability boundaries and application suitability. Common DAQ isolation topologies include: Channel-to-system-ground isolation Bank (group) isolation Channel-to-channel isolation Channel-to-System-Ground Isolation Definition: Each channel (or group of analog front ends) is isolated from system ground and host ground, while channels typically share a common reference ground. Channel-to-system-ground This topology can: Break ground loops between the measurement side and the host side. Prevent high potential or fault energy from reaching the computer, USB, or network interface. Significantly improve stability when measurement and host grounds differ. The entire DAQ effectively “floats” with the device under test, while the host remains on the safe side. Suitable scenarios include industrial field measurements where all channels share the same potential. Bank Isolation Definition: Channels are divided into groups (banks). Each bank has its own isolation domain, with isolation between banks and between each bank and system ground. Bank isolation This topology allows multiple independent systems to be measured simultaneously while preserving multi-channel synchronization within each bank, balancing cost, size, and isolation capability. Channel-to-Channel Isolation Definition: Each channel has a fully independent isolation domain and reference ground. Channel-to-channel isolation Each channel effectively functions as an independent isolated acquisition system, suitable for battery stacks, distributed measurements, and scenarios with large inter-channel potential differences, at the expense of higher cost, size, and system complexity. Isolation Selection: From Parameters to Practical Judgment After understanding isolation concepts, topologies, and voltage ratings, the key question becomes: does a given isolation design truly fit the application? Many misjudgments arise from focusing on a single number such as “1000 V isolation” without clarifying where isolation is applied, for how long, and what additional protections are required. What Is Being Isolated, and Where Does the Isolation Occur? If all measurement objects belong to the same system and there is no potential difference between them, a Channel-to-System Ground Isolation data acquisition card should be selected. If the measurement objects belong to multiple different systems, but the measurement points within each system share the same ground reference, a Bank Isolation (group isolation) architecture should be selected. In this case, measurement points from different systems must not be connected to the same bank of the acquisition card. If all measurement objects belong to the same system but there are significant potential differences between them, a Channel-to-Channel Isolation data acquisition card should be selected. This is the prerequisite for evaluating all isolation-related parameters.If the isolation location is unclear, other voltage specifications are almost meaningless for comparison. Isolation Withstand Voltage of a Data Acquisition System At a minimum, the following information must be clearly specified:whether the voltage is AC or DC, the duration (typically a 1-minute withstand test), and only then the voltage value itself. If a data acquisition card specifies an AC isolation voltage of 1000 V, it means that an AC voltage with a peak value of ±1414 V is applied between the circuit grounds on both sides of the isolation barrier, and after 1 minute the leakage current remains below 0.1 mA. If a data acquisition card specifies a DC isolation voltage of 1000 V, it means that a +1000 V or −1000 V DC voltage is applied between the circuit grounds on both sides of the isolation barrier, and after 1 minute the leakage current remains below 0.1 mA.However, one must not assume that ±1000 V AC can be applied in this case—the two are not equivalent, because different devices have different withstand capabilities for AC and DC voltages. It should be emphasized that the withstand voltages discussed above are short-term withstand ratings. They do not mean that the device can operate continuously at a 1000 V common-mode voltage. They only indicate that the device will not be damaged under those conditions, not that normal operation is guaranteed. Maximum Common-Mode Operating Voltage This is the parameter that deserves particular attention when selecting a data acquisition card. In most cases, it refers to the long-term voltage difference between the measurement side and the system ground. For example, if we want to measure the current on a 220 V mains line, the corresponding common-mode voltage is: 220 V × 1.414 = 311 V Allowing at least a 50% margin, the data acquisition card should therefore support a maximum common-mode operating voltage greater than 466 V. If a specification sheet only provides isolation withstand voltage but does not clearly specify working voltage or maximum common-mode range, extreme caution is required in practical use. Input Voltage Range The input voltage range is also referred to as differential voltage. It defines how much voltage difference the input terminals of a channel can tolerate. The key question is what happens when this limit is exceeded:is the signal clipped, is the input shut down, or is permanent damage caused? This parameter determines whether the device can protect itself under wiring errors or abnormal conditions, or whether it will fail catastrophically. If the distinction between common-mode voltage and differential voltage is still unclear at this point, the following analogy may help. Measuring Across a River In the diagram, the person cannot approach the apple directly because of the river acting as an isolation barrier, so a caliper with an extended handle is used to measure the apple on the opposite bank. The 300 cm distance across the river corresponds to the common-mode voltage in the system, while the measurement range of the caliper (20 cm) corresponds to the differential voltage range. Isolation Structure of the SonoDAQ Module (Bank Isolation Example) After distinguishing between channel-to-ground isolation, bank isolation, and channel-to-channel isolation, as well as various isolation parameters, the next question for a specific product is: where exactly is the isolation boundary drawn? The following figure shows the isolation structure of a SonoDAQ module, illustrating the division of its isolation domains. SonoDAQ Module Isolation From the module structure, it can be clearly seen that SonoDAQ Pro adopts a bank isolation architecture (see Section 6.2). Each module is isolated from the host, while the four channels on each module are not isolated from each other. The module divides functionality and electrical domains into three parts: Measurement Side: Located on the left side of the module, directly connected to sensors and the device under test. This belongs to the measurement-side electrical domain and may be at a high or uncertain common-mode potential. Bank Isolation Domain: Located in the middle of the module, this is the primary isolation barrier between the measurement side and the system side. Multiple channels within the same bank share a common measurement-side reference ground and are collectively isolated from the system side through this isolation domain. As shown in the diagram, two types of isolation circuits are used: capacitive isolation for digital communication and magnetic (transformer-based) isolation for power. System Side: Located on the right side of the module, communicating with the host through the backplane. This side operates under system ground reference and connects to processors, communication interfaces, and the host computer. From Concept to Verification: Isolation Must Be Proven, Not Assumed Through the previous discussion, we have distinguished between differential and common-mode voltages and understood the respective roles of isolation withstand voltage, working voltage, and common-mode capability. While these concepts are not complex in specifications or standards, a more critical question remains in real engineering practice: Do these isolation boundaries actually hold under real-world conditions as the parameters suggest? For example, when the device under test operates at a high common-mode potential, the acquisition system must run online for extended periods, and the host computer and operators must always remain on the safe side. Simply “trusting a specification value” is far from sufficient. Rather than staying at the conceptual level, it is better to return to engineering practice. The following two experiments are not intended to demonstrate extreme parameter limits, but to address a more practical question. For this purpose, SonoDAQ Pro was selected as the test platform—not because of exceptionally high specifications, but because its isolation structure is clear and its boundaries are well defined, making it suitable for engineering-level isolation verification. The experiments are conducted from two perspectives: withstand voltage testing (hipot) and mains-powered incandescent lamp current measurement. Withstand Voltage Test (Hipot) Test objective: To verify that the isolation barrier can withstand high voltage under specified conditions without breakdown, providing an intuitive engineering verification result The general industry definition of dielectric withstand testing is to apply an elevated voltage across an insulation barrier for approximately 1 minute. Passing the test indicates that the insulation system has sufficient electrical strength under those conditions, while also clarifying the purpose and limitations of the test to avoid misinterpretation. Test equipment: WB2671 hipot tester Test conditions: 1000 V DC, duration 1 minute, leakage current threshold 0.1 mA Withstand Voltage Test Test Results 1.02 kV DC, duration 1 minute, leakage current = 0.03 mA, with no breakdown, flashover, or arcing observed. Explanation: SonoDAQ Pro adopts a bank isolation architecture, where the six slots are isolated from each other. Therefore, during testing, the hipot voltage was applied between Channel 1 of two adjacent modules. 220 V Mains Incandescent Lamp Current Measurement Experiment Test objective: To demonstrate how the data acquisition card can measure signals in a high-voltage system under real mains conditions, and to verify measurement correctness. Why an incandescent lamp? Its steady-state behavior closely resembles a resistive load, making current waveforms intuitive and easy to interpret. The cold filament has low resistance, producing a clear inrush current at power-on, which is suitable for demonstrating transient capture and trigger recording capability. 220 V Mains Incandescent Lamp Current Measurement Wiring In the diagram, the left side is the high-voltage area directly connected to the 220 V AC source. After all wiring is completed, the power plug is inserted. The right side contains the isolated data acquisition card, forming the low-voltage area. The computer and operator remain entirely on the safe side. The experiment used SonoDAQ Pro hardware with OpenTest software. The incandescent lamp was rated at 220 V / 60 W. The following photos show the setup before power-on (left) and after power-on (right). 220 V Mains Incandescent Lamp Current Measurement Test configuration: sampling rate 192 kSa/s, AC coupling for the input signal. The acquisition card directly measured the voltage across a 1.4 Ω shunt resistor. Using the “Record” function in OpenTest, the entire power-on and power-off process was recorded. Steady-State Current Waveform Steady-state current:Vrms = 386 mV → Irms = 386 / 1.4 = 275.7 mAFrequency f = 49.962 Hz Startup Transient Current Startup current:Vpeak = 2.868 V → Ipeak = 2.868 / 1.4 = 2.05 A Crest factor calculation:CF = Ipeak / Irms = 2.05 / 0.2757 = 7.44 Incandescent lamp power calculation:P = 220 V × 0.2757 A = 60.65 W Conclusion SonoDAQ Pro can accurately measure the operating current of an incandescent lamp connected directly to the mains without using a current transformer. This experiment does not merely verify whether mains signals can be measured; it verifies whether isolation can simultaneously ensure system safety and measurement accuracy when the device under test operates at a high common-mode potential over extended periods. Isolation Is Not a Parameter, but a Boundary Isolation is not “a single voltage value,” but rather defines where risk is confined and whether signals can still pass reliably. A reliable isolation solution is the result of structure, parameters, topology, and application scenario all being valid at the same time. To see more imformation about the SonoDAQ, please fill in the form below, and we will recommend the best solution to address your needs.
A2DP (Advanced Audio Distribution Profile) is the core Classic Bluetooth profile for high-quality audio streaming. This article provides an overview of how A2DP transmits music, explains its position in the Bluetooth protocol stack, and introduces a practical A2DP testing workflow using the CRY578 Bluetooth LE Audio Interface. How Does A2DP Transmit Music? A2DP is the core profile in Classic Bluetooth for the unidirectional transmission of high-quality audio streams. It primarily defines two roles: the audio Source and the audio Sink. A2DP and the Bluetooth Protocol Stack Thinking of A2DP as a high-speed logistics channel that "delivers" music from one device to another, the diagram above illustrates the division of responsibilities from the moment audio is generated to the point it is transmitted wirelessly. Figure 1 A2DP System Block Diagram At the top of the stack, the Application / Audio Source (or Audio Sink) layer acts as the "content factory" and "player". On the transmitting side, it obtains PCM audio data from the system and encodes it into Bluetooth-supported formats such as SBC or AAC. On the receiving side, it decodes the bitstream back into audio for playback. This layer directly determines the perceived audio quality—akin to the quality of raw materials and finished products—which users experience most intuitively. Below this is the A2DP Profile layer, which functions as a "cooperation agreement". It defines which device acts as the Source and which as the Sink, along with the supported codecs, sampling rates, and other parameters. The profile itself does not carry audio data; instead, it ensures both sides agree on "what format to use and how to transmit" before streaming begins. The next layer down is AVDTP, the "transport and scheduling control center". AVDTP is responsible for establishing and managing audio streams. It translates user actions—such as play, pause, and stop—into explicit protocol procedures and sends the encoded audio data over the media channel. The smooth operation of A2DP in practice largely depends on this layer. Below AVDTP is L2CAP, which acts as a standardized "containerized transport system". Both audio data and control information are segmented, encapsulated, reassembled, and multiplexed here. They are then delivered in an orderly fashion to the lower layers, ensuring stable and reliable transmission over a single Bluetooth link. At the bottom, the LMP, Baseband, and RF layers form the system’s “roads, vehicles, and radio infrastructure.” They handle device pairing, link management, and the actual wireless transmission, converting all upper-layer data into bitstreams over the Bluetooth air interface. Viewed from top to bottom, the A2DP protocol stack exhibits a clear downward flow: the upper layers focus on the audio content itself, while the lower layers handle wireless data delivery. This strict separation of responsibilities is what allows us to enjoy stable and continuous music playback through Bluetooth headphones. How to Test A2DP Functionality with CRY578? The CRY578 Bluetooth LE Audio Interface is CRYSOUND's latest test interface dedicated to Bluetooth audio and user-interface testing. Based on Bluetooth v5.4, the CRY578 supports both Classic Bluetooth and Bluetooth Low Energy audio simultaneously, making it suitable for use in both R&D laboratories and production-line testing. Building an A2DP Test Environment CRYSOUND provides a complete Bluetooth audio test solution, including both hardware and software, to support A2DP testing. In the CRYSOUND Bluetooth audio test system, the components are as follows: CRY578 acts as the Bluetooth Source, responsible for device discovery, connection, and audio transmission. DUT (Device Under Test) acts as the Bluetooth Sink, receiving, decoding, and playing the audio stream. B&K HATS simulates human acoustic characteristics, captures audio signals, and converts them into analog signals for the acquisition system. SonoDAQ + OpenTest (https://opentest.com) perform data acquisition and analysis, evaluating DUT performance based on the test results. Figure 2 Test System Block Diagram In this setup, the CRY578 can be controlled either via its PC software (Bluetooth LE Audio Interface) or through serial commands to scan for nearby Bluetooth devices and establish connections. Standard test signals—such as sweeps, noise, and distortion signals—are played from the PC. The acoustic output from the DUT is captured and analyzed by OpenTest to evaluate performance metrics such as frequency response, distortion, and signal-to-noise ratio. The CRY578 also supports switching to high-quality codecs such as AAC and LDAC, as well as multiple sampling rates, for comprehensive testing. A2DP Test Procedure Establish the Bluetooth Connection At the beginning of the test, a Bluetooth connection must be established between the CRY578 (acting as the A2DP Source) and the DUT (acting as the A2DP Sink). Figure 3 inquiry and connect The connection process includes device discovery and pairing, ACL link establishment, A2DP profile setup, and codec capability negotiation. Test Signal Generation from the Host PC Audio test software, such as OpenTest or SonoLab, generates standard signals like single-tone sine waves or sweeps. These signals are sent as PCM data to the CRY578 via a USB Audio Class (UAC) link. Figure 4 Test Scenario Audio Transmission via Bluetooth by CRY578 The continuous PCM audio stream is first segmented into fixed-size frames, which are then passed to an encoder (e.g., SBC or AAC) for compression, producing encoded frames. These frames are encapsulated into AVDTP media PDUs according to the A2DP specification. The PDUs are segmented and multiplexed by L2CAP, passed through the HCI interface to the Bluetooth controller, packaged as ACL packets at the baseband layer, and finally transmitted over the Bluetooth RF link. Decoding and Playback by the DUT The DUT performs the reverse process of the CRY578's transmission chain. The Bluetooth packets are decoded back into PCM data, which is then converted to analog signals by a DAC and output through the speaker. Acoustic Capture by B&K HATS The high-precision microphones built into B&K HATS capture the sound produced by the DUT and convert it into analog signals. Data Processing and Analysis with SonoDAQ + OpenTest SonoDAQ digitizes the analog signals and sends them to OpenTest. OpenTest then applies its internal algorithms to analyze the audio data and generate results—such as frequency response and distortion measurements. These results are then used to determine if the DUT meets the performance requirements. The Value of Bluetooth Protocol Analyzers in Testing During testing, audio data undergoes multiple digital-to-analog conversions, RF transmission, and acoustic-to-electrical conversion. An issue at any stage can affect the final test results. Once problems in the analog and digital signal paths have been ruled out, the root cause often lies in the Bluetooth RF transmission. In such cases, a Bluetooth protocol analyzer becomes an effective tool for pinpointing the exact issue. Figure 5 Capture Bluetooth packets using Ellisys If you are interested in Bluetooth audio testing, please visit CRY578 Bluetooth LE Audio Interface to learn more or fill out the Get in touch form below and we'll reach out shortly.
Sound is everywhere in our daily life: birdsong, street noise, engine roar, even the faint airflow from an air conditioner. For people, sound is not only about whether we can hear it, but whether it feels comfortable, is disturbing, or poses a risk. The same 70 dB can feel completely different; and when something feels "noisy", the cause may come from the source itself, the propagation direction, or reflections from the environment. When we turn this "perception" into quantifiable engineering data, the three most easily confused concepts are sound pressure, sound intensity, and sound power. They answer: Sound pressure: how loud it is at a specific point; Sound intensity: how much sound energy is propagating in a particular direction; Sound power: how loud the source is in terms of its total acoustic emission; This article explains sound pressure, sound intensity, and sound power in an intuitive way, so you can better understand sound. Sound Waves In engineering acoustics, sound pressure, sound intensity, and sound power are three fundamental and important physical quantities. Before introducing them in detail, we need the concept of a sound wave. A vibrating source sets the surrounding air particles into vibration. The particles move away from their equilibrium position, drive adjacent particles, and those adjacent particles generate a restoring force that pushes the particles back toward equilibrium. This near-to-far propagation of particle motion through the medium is what we call a sound wave. Figure 1. Propagation of a Sound Wave in Air Sound Pressure When there is no sound wave in space, the atmospheric pressure is the static pressure p0. When a sound wave is present, a pressure fluctuation is superimposed on p0, producing a pressure fluctuation p1. Here p1 is the sound pressure (unit: Pa). Therefore, sound pressure is the instantaneous deviation of the air static pressure caused by the sound wave. The human brain does not respond to the instantaneous amplitude of sound pressure, but it does respond to the root-mean-square (RMS) value of a time-varying pressure. Therefore, the sound pressure p can be expressed as: In practical engineering applications, the sound pressure level Lp: where Pref = 2 × 10-5 Pa is the reference sound pressure. In practice, we usually use sound pressure level (dB) to characterize sound pressure, rather than using pressure in pascals. Why? Figure 2 answers this well. From a library to the entrance of a high-speed rail station, sound pressure may increase by a factor of 100, while sound pressure level increases by only 40 dB. This reflects the difference between a linear scale and a logarithmic scale. From an engineering perspective, using sound pressure directly leads to large numeric variations that are inconvenient for evaluation. Moreover, the human auditory system is closer to a logarithmic response, so sound pressure level better matches hearing. Figure 2. Sound Pressure and Sound Pressure Level Sound Intensity Sound intensity describes the transfer of acoustic energy. It is the acoustic power passing through a unit area per unit time. It is a vector quantity that is directional, with units of W/m2, defined as the time average of the product of sound pressure and particle velocity: where v(t) denotes the particle velocity vector. Under the ideal plane progressive-wave approximation, sound pressure and particle velocity approximately satisfy: where ρ is the air density, c is the speed of sound. Therefore, the magnitude of sound intensity along the propagation direction can be written as: Similarly, sound intensity has a corresponding intensity level LI: where I0 = 10-12 W/m2 is the reference sound intensity. Compared with sound pressure level measurements, sound intensity measurements have the following characteristics: Directional:it can distinguish whether acoustic energy is propagating outward or flowing back, so under typical field conditions it is often less sensitive to reflections and background noise; Source localization:intensity scanning can directly reveal the main radiation regions and leakage points, making remediation more targeted; Higher system complexity:it typically requires an intensity probe, with higher overall cost and more setup and calibration effort; Figure 3. Sound Intensity Testing A key advantage of sound intensity measurement in engineering applications is that it characterizes both the direction and magnitude of acoustic energy flow. It can separate the contributions of outward radiation from the source and reflected backflow from the environment, so under non-ideal field conditions it tends to be less affected by reflections and background noise. In addition, the sound intensity method can obtain sound power directly by spatially integrating the normal component of intensity over an enclosing surface. Combined with surface scanning, it can identify dominant source regions and locate leakage points. Therefore, it is highly practical and interpretable for noise diagnosis, verification of noise-control measures, and sound power evaluation. The key instrument for sound intensity testing is the sound intensity probe. Unlike a single microphone, an intensity probe is not used merely to measure “how large the pressure is”; it must provide the basic quantities required for calculating intensity (sound pressure and particle velocity). Therefore, the probe typically outputs two synchronous channels and, together with a two-channel data-acquisition front end and dedicated algorithms, yields intensity results. In engineering practice, the probe often includes interchangeable spacers, positioning fixtures, and windshields. Channel amplitude/phase matching, phase calibration capability, and airflow-interference mitigation directly determine the credibility and usable frequency range of intensity measurements. Two types of sound intensity probes are commonly used: P-U probes (pressure-particle-velocity) and P-P probes (pressure-pressure). A P-U probe consists of a microphone and a velocity sensor, measuring sound pressure p(t) and particle velocity v(t) simultaneously. The principle is more direct, but particle-velocity sensors are often more sensitive to airflow, contamination, and environmental conditions, requiring more protection and maintenance in the field and usually costing more. Figure 4. P-U Sound Intensity Probe (Microflown) A P-P probe uses two matched microphones aligned on the same axis. It uses the two pressure signals p1(t) and p2(t) to estimate the particle-velocity component v(t). However, it is sensitive to inter-channel phase matching and the choice of microphone spacing - the spacing determines the effective frequency range: a larger spacing benefits low frequencies, but high frequencies suffer from spatial sampling error; a smaller spacing benefits high frequencies, but low frequencies become more susceptible to phase mismatch and noise. Figure 5. P-P Sound Intensity Probe (GRAS) P-U probes are relatively niche, mainly because it is difficult to make them both stable and inexpensive, and they generally have poorer resistance to airflow. P-P probes, thanks to their good field robustness and the ability to adjust bandwidth flexibly via microphone spacing, are currently the mainstream choice in engineering applications. Sound Power Sound power W is the rate at which a source radiates acoustic energy, with units of watts (W). For any closed measurement surface S enclosing the source, the sound power equals the integral of the normal component of sound intensity over that surface: where n is the unit normal vector pointing outward from the measurement surface. Sound power level Lw is defined as: where W0 = 10-12 W is the reference sound power. Figure 6. Sound Power Measurement Sound power characterizes a source's inherent acoustic emission capability: the total acoustic energy it radiates per unit time. It has little to do with measurement distance or microphone position, and ideally does not depend on how "loud" it is at a particular point in a room. This is fundamentally different from sound pressure and sound intensity. To better understand sound pressure, sound intensity, and sound power, you can imagine noise as water flow. Sound pressure is like the "water pressure" you feel when you put your hand at a certain location (it changes with distance to the nozzle, direction, and the shape of the basin). Sound intensity is like the instantaneous "direction and rate of flow" (it has direction and can even be reflected by walls, creating backflow). Sound power is like "how much water the nozzle sprays per second" - it is a property of the nozzle itself. In measurement, it is obtained by integrating the outward normal flow over a surface surrounding the device. Figure 7. Analogy of Sound Pressure, Sound Intensity, and Sound Power In real projects, the algorithms for sound pressure, sound intensity, and sound power are relatively mature. The hardest part is acquiring the signals accurately and obtaining results quickly. In particular, tasks such as multi-channel microphone arrays, sound intensity, and sound power impose three hard requirements on the data-acquisition front end: low noise and wide dynamic range, strict synchronization and phase consistency, and stable on-site connections and power. SonoDAQ + OpenTest is positioned to provide a "front-end acquisition + synchronous analysis" foundation for engineering acoustics, allowing engineers to focus more on operating-condition control and data interpretation. It delivers the most value in the following types of projects: Sound intensity diagnostics: dual-channel synchronous sampling plus better amplitude/phase consistency management provide a more stable data basis for P-P intensity probes and intensity scanning. Microphone array systems: better aligned with engineering deployment needs in channel scalability, synchronization, and cabling, making it suitable for building expandable distributed test platforms. Sound power and standardized testing: helps engineers quickly lay out measurement points, covering multiple international sound power test standards. With guided configuration, one-click testing, and automatic report export, it saves substantial time and effort for engineers. Figure 8. SonoDAQ + OpenTest To see more clearly how SonoDAQ is connected and configured, typical application cases (such as equipment noise evaluation, sound source localization, and sound power testing), and commonly used BOM lists, please fill in the form below, and we will recommend the best solution to address your needs.
In audio and vibration testing, FFT analysis (Fast Fourier Transform) is one of the tools almost every engineer uses sooner or later: Loudspeaker frequency response Headphone distortion NVH diagnostics Structural resonance troubleshooting Production noise and “mysterious tone” hunting A lot of practical questions are actually asking the same few things: Where is the energy concentrated in frequency? Is it dominated by one tone or a bunch of harmonics? How high is the noise floor? Are there any resonance peaks? FFT is the most universal entry point to answer these questions. This article will help you clarify three things from an engineering perspective: What FFT analysis is How FFT works conceptually How to use FFT correctly and efficiently in practice What Is FFT? In the time domain, a signal is just a waveform changing over time – all components “stacked together” in one trace. You can see it, but it’s hard to tell which frequencies are inside. FFT (Fast Fourier Transform) decomposes a time-domain signal into a sum of sinusoids at different frequencies. In the frequency domain, the signal is represented by frequency + amplitude + phase. In simple terms: Time domain: how the signal moves over time Frequency domain: what frequency components it contains, which are strongest, and how they relate to each other Historically, Fourier’s key idea (early 19th century) was that a complex periodic function can be expressed as a sum of sines and cosines. This evolved into the continuous-time Fourier transform, mapping signals onto a continuous frequency axis. In the computer age, things changed: engineers work with sampled data and typically only have a finite-length record of N samples. That leads to the DFT (Discrete Fourier Transform), which maps N time samples to N discrete frequency bins. FFT (Fast Fourier Transform) is not a different transform. It is a family of algorithms that compute the exact same DFT much more efficiently: Direct DFT: complexity ~ O(N²) FFT: complexity ~ O(N log N) The output X[k] is identical to the DFT result – FFT just gets there far faster by exploiting symmetry and divide-and-conquer. What FFT Is Good at – and What It Isn’t FFT is very good at: Finding deterministic narrowband components Fundamental tones, harmonics, switching frequencies, whistle tones, speed-related lines Looking at broadband distributions Noise floor, 1/f slopes, in-band power, SNR Characterizing system behavior Transfer functions, resonances / anti-resonances, coherence, delay estimation Serving as the foundation of time–frequency analysis STFT, spectrograms, etc. FFT is not good at (or not sufficient on its own for): Strongly non-stationary signals and “instantaneous frequency” For chirps and rapidly changing content, you need STFT, wavelets, or other time–frequency methods, not a single FFT on a long record Separating two extremely close tones below your frequency resolution If the spacing is smaller than your bin resolution (set by N), no algorithm will magically resolve them Turning short data into “long measurements” Zero padding only interpolates the spectrum visually; it does not add new information Before Using FFT: Key Concepts to Get Right To use FFT well, you need to be confident about a few fundamentals: Sampling rate DFT and its interpretation What you actually plot (magnitude, amplitude, power, PSD) Windowing and spectral leakage Averaging Sampling Rate: How High in Frequency You Can See Before FFT, you already made one crucial decision: sampling. A continuous-time signal x(t) is turned into a discrete sequence x[n]=x(n/fs). The sampling rate fsf_sfs determines the highest frequency you can observe without aliasing: the Nyquist frequency, fs/2. If the analog signal contains energy above fs/2, it does not disappear – it folds back into the band below Nyquist as aliasing. Once aliasing happens, FFT cannot “undo” it; the information is irretrievably mixed. In practice, you must use an anti-alias filter before the ADC (or before any resampling) to suppress components above Nyquist. Example: A 900 Hz sine sampled at fs=1 kHz will appear at 100 Hz in the discrete spectrum – a classic aliasing artifact. DFT Computation and Interpretation Given N samples x[0]..x[N−1], the DFT is defined as: The inverse transform (IDFT) reconstructs the time signal: Intuitively, X[k] tells you how strongly the signal correlates with a complex exponential at that bin’s frequency. The magnitude X[k] indicates “how much” of that frequency component exists The phase encodes time alignment relative to other components What Are You Plotting? Magnitude, Amplitude, Power, PSD From one set of FFT results X[k], you can create many different “spectra” that look similar but represent different physical quantities. This is where confusion between tools and platforms often arises. Common variants include: Magnitude spectrum |X[k]| Units depend on normalization (e.g., “V·samples”) Useful for locating peaks, harmonics, and general spectral shape Amplitude spectrum Properly scaled magnitude, in physical units (e.g. V) Appropriate for reading off sinusoid amplitudes and doing calibrated measurements Power spectrum |X[k]|² Again, scaling dependent; often used for power/energy comparisons when conventions are fixed Power Spectral Density (PSD) Sxx(f) Units like V²/Hz or Pa²/Hz Used for noise analysis, band power, and comparisons across different FFT lengths If you want to compare noise levels across different FFT sizes, windows, or tools, use PSD (or amplitude spectral density). Raw |X| or |X|² values are rarely directly comparable. A Concrete Example: Two Tones in Time and Frequency Imagine a signal consisting of two sinusoids at different frequencies. In the time domain, their sum may look like a “wobbly” waveform. In the frequency domain (FFT/PSD), you will see two distinct narrow peaks at the corresponding frequencies. In OpenTest’s FFT analysis, you can visualise both the spectrum and PSD/ASD side by side, making it easy to: Identify tonal components Inspect noise distribution Compare different operating conditions on the same frequency grid Try it yourself: Download the free OpenTest edition and run an FFT on a simple two-tone signal to see both peaks clearly separated. Window Functions and Spectral Leakage: Cleaning Up Spectra In theory, FFT assumes the sampled block contains an integer number of periods and is then repeated periodically. In reality, the record almost never lines up perfectly with an integer number of cycles. When you repeat that block, you get discontinuities at the boundaries, which causes energy to spread into neighboring bins — this is spectral leakage. To reduce leakage, we typically apply a window function to the time record before doing FFT. A window simultaneously affects: Main lobe width Wider main lobe = peaks get broader → it’s harder to separate close tones Side lobe height Lower side lobes = easier to see small peaks near a large one (better dynamic range) Amplitude/energy scaling Windows change the relationship between a pure tone’s true amplitude and the observed peak, as well as the noise floor level Some practical guidelines: Rectangular window Only use when you can ensure coherent sampling (an integer number of periods in the record) and you want the narrowest possible main lobe Hanning (Hann) window A very robust default choice for general acoustics and vibration work Widely used with Welch/PSD methods Hamming Similar to Hann, with slightly different side-lobe behavior, common in communications Blackman / Blackman–Harris Lower side lobes, useful when you need to see small peaks next to big ones, at the cost of a wider main lobe In OpenTest, you can switch between different window functions in the FFT analysis module and immediately see the impact on peak width, side lobes, and noise floor. Averaging: Making Spectra More Stable For noisy or non-stationary signals, a single FFT can look very “spiky” or unstable. By averaging multiple spectra, you obtain a smoother, more repeatable result. Common averaging types include: Linear averaging A simple arithmetic mean of several FFT results Exponential averaging Recent data gets more weight; good for live monitoring when the spectrum should react but not jump wildly Energy (power) averaging Based on power; ensures power-related quantities remain consistent A good averaging configuration strikes a balance between suppressing random fluctuations and preserving genuine changes in the signal. Where Do We Use FFT in Practice? Audio and Acoustics Typical applications include: Finding feedback frequencies, harmonic distortion, and device noise floors Frequency response (transfer function) measurement Room modes / resonance analysis Spectrograms of speech, music, and equipment noise In audio/acoustics, you must be clear about units and conventions: dB SPL, A-weighting, 1/3-octave bands, etc. FFT is the engine; the reporting convention (reference, weighting, bandwidth) must be clearly defined. Vibration and Rotating Machinery Identifying speed-related peaks (1X, 2X, gear mesh frequencies) Structural resonances and mode behavior under different operating conditions Bearing diagnostics, gear whine, imbalance, misalignment For bearing and gearbox analysis, envelope detection/demodulation is often used: Band-pass filter the signal Demodulate and then perform FFT on the envelope to reveal fault frequencies If the rotational speed is changing, a simple FFT will “smear” peaks. In that case, order tracking or synchronous resampling is more appropriate, turning the axis from “frequency” into “order”. Power Electronics and Power Quality Line frequency harmonics (50/60 Hz and multiples), THD, ripple, switching spikes Pre-compliance EMI checks: spectral lines, noise floor, in-band power In power systems, non-coherent sampling is a common issue: if the record length is not an integer number of mains cycles, leakage affects harmonic accuracy. Solutions include synchronous sampling, integer-cycle windows, or specialized harmonic analyzers. RF and Communications (Baseband View) Modulated signal spectra and spectral masks OFDM and multi-carrier spectral analysis, adjacent channel leakage Here, consistency is paramount: Same units Same bandwidth (RBW) Same window, detector, and averaging style FFT itself is straightforward; turning it into comparable power measurements requires tightly defined settings. Imaging and 2D Filtering 2D FFT extends the same idea to images: Edges correspond to high spatial frequencies; smooth areas to low frequencies Low-pass / high-pass filtering, removal of periodic noise, convolution acceleration in the frequency domain The same periodic extension assumption now applies in 2D: discontinuities at image borders produce strong artifacts in the frequency domain. Padding, mirrored borders, or 2D windows are common ways to mitigate this. Turning FFT into an Everyday Engineering Tool From a mathematical standpoint, FFT is not particularly “lightweight”. But in engineering use, the goal is actually simple: See what’s hidden inside the signal more clearly and much faster. When you understand: What FFT really computes How sampling, windowing, scaling, and averaging affect the result When to use spectra vs PSD, and which settings matter for your use case …then FFT stops being an abstract math topic and becomes a practical, everyday tool for acoustics and vibration work – from R&D and validation all the way to production testing. Download and get started now -> or fill out the form below ↓ to schedule a live demo. Explore more features and application stories at www.opentest.com.
Octave-band analysis can be implemented in two fundamentally different ways: FFT binning (integrating PSD/FFT bins into 1/1- and 1/3-octave bands) and a true octave filter bank (standards-oriented bandpass filters + RMS/Leq averaging). In this post, we compare how the two methods work, where their results match, where they diverge (scaling, window ENBW, band-edge weighting, latency, transient response), and how OpenTest supports both for acoustics, NVH, and compliance measurement. For a detailed explanation of the concepts, read this → Octave-Band Analysis: The Mathematical and Engineering Rationale Octave-band filter banks (true octave / CPB filter bank) Parallel bandpass filters + energy detector + time averaging A filter-bank (true octave) analyzer typically: Design a bandpass filter H_b(z) (or H_b(s)) for each band center frequency. Run filters in parallel to obtain band signals y_b(t). Compute band mean-square/power and apply time averaging to output band levels. To be comparable across instruments, filter magnitude responses must satisfy IEC/ANSI tolerance masks (class) for the specified filter set. [1][3] IIR vs FIR: why IIR (cascaded biquads) is common in practice IIR advantages: lower order for a given roll-off, lower compute, good for real-time/embedded; stable when implemented as SOS/biquads. FIR advantages: linear phase is possible (useful when waveform shape matters); design/verification can be more straightforward. For band-level outputs, phase is usually not the primary concern, so IIR filter banks are common. Multirate processing: the “secret weapon” of CPB filter banks Low-frequency CPB bands are very narrow. Implementing them at the full sampling rate is inefficient. A common strategy is to group bands by octave and downsample for low-frequency groups: Low-pass then decimate (e.g., by 2 per octave) for lower-frequency groups. Implement the corresponding bandpass filters at the reduced sampling rate. Ensure adequate anti-aliasing before decimation. Time averaging / time weighting: band levels are statistics, not instantaneous values Band levels typically require time averaging. Common options include block RMS, exponential averaging, or Leq (energy-equivalent level). In sound level meter contexts, IEC 61672-1 defines Fast/Slow time weightings (Fast ~125 ms, Slow ~1 s). [5][6] Engineering implication: different time constants produce different readings, so time weighting must be stated in reports. How to validate that a filter bank behaves “like the standard” Sine sweep: verify passband behavior and adjacent-band isolation; observe time delay effects. Pink/white noise: verify average band levels and variance/stabilization time; check effective bandwidth behavior. Impulse/step: examine ringing and time response (critical for transient use). Cross-check against a known compliant reference instrument/implementation. From band definitions to compliant digital filters: an end-to-end workflow (conceptual) Choose the band system: base-10/base-2, the fraction 1/b (commonly b=3), generate exact fm and f1/f2. Choose performance target: which standard edition and which class/mask tolerance? Choose filter structure: IIR SOS for real-time; FIR or forward-backward filtering if phase/zero-phase is required. Design each bandpass: map f1/f2 into the digital domain correctly (e.g., pre-warp for bilinear transform). Implement multirate if needed: decimate for low-frequency groups with sufficient anti-alias filtering. Verify: magnitude response vs mask; noise tests for effective bandwidth; sweep/impulse tests for time response. Calibrate and report: units and reference quantities, averaging/time weighting, method details. Time response explained: group delay, ringing, and averaging all shape readings A band-level analyzer is a time-domain system (filter → energy detector → smoother), so readings are governed by multiple time scales: Filter group delay: how late events appear in each band. Filter ringing/decay: how long a short pulse “rings” within a band. Energy averaging/time weighting: the time resolution vs fluctuation of the output level. Thus, for transients (impacts, start/stop events, sweeps), different compliant implementations can yield different peak levels and time tracks—consistent with ANSI’s caution. [3] Rule of thumb: for steady-state contributions, use longer averaging for stability; for transient localization, shorten averaging but accept higher variability and lock down algorithm details. Common real-time pitfalls Forgetting anti-aliasing in the decimation chain: low-frequency bands become contaminated by aliasing. Numerical instability of high-Q low-frequency IIR sections: use SOS/biquads and sufficient precision. Averaging in dB: always average in energy/mean-square, then convert to dB. Assuming band energies must sum exactly to total energy: standard filters are not necessarily power-complementary; verify using standard-consistent criteria instead. Octave-Band Filter Bank Analysis in OpenTest OpenTest supports octave-band analysis using a filter-bank approach:1) Connect the device, such as SonoDAQ Pro2) Select the channels and adjust the parameter settings. For an external microphone, enable IEPE and switch to acoustic signal measurement.3) In the Octave-Band Analysis section under Measurement Mode, choose the IEC 61260-1 algorithm. It supports real-time analysis, linear averaging, exponential averaging, and peak hold.4) After configuring the parameters, click the Test button to start the measurement.5) A single recording can be analyzed simultaneously in 1/1-octave, 1/3-octave, 1/6-octave, 1/12-octave, 1/24-octave, and 1/24-octave bands. Figure 1: Octave-Band Filter Bank Analysis in OpenTest FFT binning and FFT synthesis FFT binning: convert a narrowband spectrum into CPB band integrals Estimate spectrum (single FFT, Welch PSD, or STFT). Integrate/sum within each octave/fractional-octave band to obtain band power. This is common in software/offline work because a single FFT provides high-resolution spectrum that can be re-binned into any band system (1/1, 1/3, 1/12, …). Key challenge #1: FFT scaling and window corrections After an FFT, scaling depends on your definitions: 1/N normalization, amplitude vs power vs PSD, one-sided vs two-sided spectrum, and windowing. For noise measurements, ENBW is crucial; ignoring it can introduce systematic offsets. [7] A practical PSD normalization (periodogram form) # convert to one-sided PSD: multiply by 2 except DC (and Nyquist if present) This yields PSD in units of (input unit)²/Hz and supports energy consistency checks by integrating PSD over frequency. Two quick self-checks for scaling White noise check: generate noise with known variance σ²; integrate one-sided PSD over 0..fs/2 and recover ≈σ² (accounting for the ×2 rule). Pure tone check: generate a sine with amplitude A (RMS=A/√2); integrating spectral energy should recover ≈A²/2 (subject to leakage and window choice). If both checks pass, your FFT scaling is likely correct; then partial-bin weighting and octave binning become meaningful. Key challenge #2: band edges rarely align to bins → partial-bin weighting Hard include/exclude decisions at band edges cause step-like errors, especially at low frequency where bands are narrow. Use overlap-based weighting (Section 4.2.4) for the boundary bins. Does zero-padding solve edge misalignment? (common misconception) Zero-padding interpolates the displayed spectrum but does not improve true frequency resolution (which is set by the original window length). It can reduce visual stair-stepping but cannot turn 1–2-bin low-frequency bands into reliable band-level estimates. Fundamental fixes are longer windows or multirate processing/filter banks. Key challenge #3: time–frequency trade-off (window length sets low-frequency accuracy and delay) FFT resolution is Δf = fs/N. Low-frequency 1/3-octave bands can be only a few Hz wide, so achieving enough bins per band requires very large N, increasing latency and smoothing transients. Root cause: 1/3 octave is constant-Q, but STFT uses constant-Δf bins In CPB, band width scales with frequency (Δf_band ∝ f, constant-Q). In STFT, bin spacing is constant (Δf_bin constant). Therefore low-frequency CPB needs extremely fine Δf_bin (long windows), while high frequency is over-resolved. Solution routes: long-window STFT vs multirate STFT vs CQT/wavelets Long-window STFT: simplest, but high latency and transient smearing. Multirate STFT: downsample low-frequency content and FFT at lower fs, similar in spirit to multirate filter banks. Constant-Q transform (CQT) / wavelets: naturally logarithmic resolution, but matching IEC/ANSI masks requires extra calibration/validation. [4] For compliance measurements, standards-oriented filter banks are preferred; for research/feature extraction, CQT/wavelets can be attractive. FFT synthesis: constructing per-band filtering in the frequency domain FFT synthesis pushes the FFT approach closer to a filter bank: Define a frequency-domain weight W_b[k] per band (brick-wall or smooth/mask-like). Compute Y_b[k] = X[k]·W_b[k] and IFFT to get y_b[n]. Compute band RMS/averages from y_b[n]. It can easily implement zero-phase (non-causal) filtering. For strict IEC/ANSI matching, W_b and normalization must be carefully designed and validated. Making FFT synthesis stream-like: OLA, dual windows, and amplitude normalization To output continuous time signals per band, use overlap-add (OLA): frame, window, FFT, apply W_b, IFFT, synthesis window, and OLA. Choose analysis/synthesis windows to satisfy COLA (constant overlap-add) conditions (e.g., Hann with 50% overlap) to avoid periodic level modulation. If the goal is to match standard filters, how should W_b be chosen? W_b[k] depends on what you want to match: Match brick-wall integration: W_b is hard 0/1 within [f1,f2]. Match IEC/ANSI filter behavior: |W_b(f)| approximates the standard mask and effective bandwidth (matches ∫|W_b|²). Match energy complementarity for reconstruction: design Σ_b |W_b(f)|² ≈ 1 (Section 7.6). You typically cannot satisfy all three perfectly at once; define your priority (compliance vs decomposition/reconstruction) up front. Energy-conserving frequency-domain filter banks: why Σ|W_b|² matters If you want band energies to sum to total energy (within numerical error), a common design aims for approximate power complementarity: IEC/ANSI masks do not necessarily enforce strict complementarity, so don’t assume exact additivity in compliance contexts. Welch/averaging strategies: how to make FFT band levels stable Use Welch averaging (segment, window, overlap, average power spectra). Average in the power domain (|X|² or PSD), then convert to dB. For non-stationary signals, consider STFT to obtain time–band matrices. Report window type, overlap, averaging count, and ENBW/CG treatment. FFT-Binning Analysis in OpenTest OpenTest supports octave-band analysis based on FFT binning:1) Connect the device, such asSonoDAQ Pro2) Select the channels and adjust the parameter settings. For an external microphone, enable IEPE and switch to acoustic signal measurement.3) In the Octave-Band Analysis section under Measurement Mode, choose the FFT-based algorithm.4) A single recording can be analyzed simultaneously in 1/1-octave, 1/3-octave, 1/6-octave, 1/12-octave, and 1/24-octave bands. Figure 2: FFT-Binning Octave-Band Analysis in OpenTest Filter-bank vs FFT/FFT synthesis: differences, equivalence conditions, and trade-offs A comparison table DimensionFilter-bank (True Octave / CPB)FFT binning / FFT synthesisStandards complianceEasier to match IEC/ANSI magnitude masks; mainstream for hardware instruments. [1][3]Hard binning behaves like band integration; matching masks requires extra weighting or standard-compliant digital filters.Real-time / latencyCausal real-time possible; latency set by filter order and averaging.Block processing adds at least one window length of delay; low-frequency resolution often forces longer windows.Transient responseContinuous output but affected by group delay/ringing; different compliant implementations may differ. [3]Set by STFT windowing; transients are smeared by windows and sensitive to window type/length.Leakage & correctionsControlled via filter design; leakage can be managed.Strongly depends on window and ENBW/scaling; edge-bin misalignment needs partial weighting. [7]InterpretabilityRMS after bandpass filtering—aligned with sound level meters and analyzers.Spectrum estimation + binning—more statistical; interpretation depends on window/averaging settings.ComputationMany filters in parallel; multirate can reduce cost.One FFT can serve all bands; efficient for offline/batch.Phase & reconstructionIIR is typically nonlinear phase (fine for levels).Frequency weights can be zero-phase; reconstruction needs attention to complementarity and transitions. When do both methods give (almost) the same answers? Band-averaged results typically agree closely when: You compare averaged band levels (not transient peak tracks). The signal is approximately stationary and the observation time is long enough. FFT resolution is fine enough that each band contains enough bins (especially at the lowest band). FFT scaling is correct (one-sided handling, Δf, window U, ENBW/CG where needed). Partial-bin weighting is used at band edges. Why differences grow for transients and short events Differences are driven by mismatched time scales: filter banks have band-dependent group delay and ringing but continuous output; STFT uses a fixed window that sets both frequency resolution and time smoothing. If event duration is comparable to the window length or filter impulse response, results depend strongly on implementation details. Error budget: where mismatches usually come from (and how to locate them quickly) Wrong averaging/combination in dB: must average and sum in the energy domain. Inconsistent FFT scaling: 1/N conventions, one-sided vs two-sided, Δf, window normalization U. Missing window corrections: ENBW for noise; coherent gain/leakage for tones. Using nominal frequencies to compute edges instead of exact definitions. No partial-bin weighting at band boundaries (especially harmful at low frequency). Multirate/anti-alias issues in filter banks. Different averaging time constants/windows between methods. True method differences: brick-wall binning vs standard filter skirts/roll-off imply systematic offsets. A strong debugging approach: first match total mean-square using white noise (scaling/ENBW/partial-bin), then validate band centers and adjacent-band isolation using swept sines or tones. Engineering checklist: make 1/3-octave analysis correct, stable, and reproducible Choose a method: compliance → filter bank; offline statistics → FFT binning For regulations/type testing/instrument comparability: prefer IEC/ANSI-compliant filter banks and report standard edition and class. [1][3] For offline processing, large datasets, or flexible band definitions: FFT binning can be efficient, but scaling and boundary weighting must be rigorous. If you need per-band time-domain signals (modulation, envelope, etc.): consider FFT synthesis or explicit filter banks. Selecting FFT parameters from the lowest band (example) Example: fs=48 kHz, lowest band of interest is 20 Hz (1/3 octave). Its bandwidth is only a few Hz. If you want at least M=10 bins per band, you may need Δf_bin ≤ bandwidth/10, implying a very large N (e.g., ~100k points; 2^17=131072). This illustrates why real-time compliance often favors filter banks. Typical mistakes that prevent results from matching Summing magnitude |X| instead of power |X|² or PSD. Averaging in dB instead of in linear power/mean-square. Ignoring ENBW/window scaling for noise. [7] Computing band edges from nominal frequencies. Not stating time weighting/averaging conventions (Fast/Slow/Leq). [5][6] Recommended validation flow (regardless of implementation) Tone-at-center test (or sweep): verify that energy peaks in the correct band and adjacent-band rejection behaves as expected. White/pink noise: verify expected spectral shape in band levels and assess stability/averaging time. Cross-implementation comparison: compare your implementation with a known reference on identical signals; isolate scaling vs definition vs filter-skirt differences. Record and freeze parameters (band definition, windowing, averaging) in the test report. Reproducibility checklist: include these in reports so others can recompute your levels Band definition: base-10 or base-2? b in 1/b? exact vs nominal used for computation? reference frequency fr? Implementation: standard filter bank (IIR/FIR, multirate) vs FFT binning/synthesis; software/library versions. Sampling/preprocessing: fs, detrending/DC removal, anti-alias filtering, resampling. Time averaging: Leq / block RMS / exponential; time constants, block size, overlap, averaging frames; Fast/Slow context if relevant. FFT details (if used): window type, N, hop, zero-padding, PSD normalization, one-sided handling, ENBW/CG, partial-bin weighting. Calibration/units: input units and reference quantities (e.g., 20 µPa), sensor calibration factors and dates. Output definition: RMS vs peak vs band power; 10log vs 20log conventions; any band aggregation steps. If you remember one line: document “band definition + time averaging + FFT scaling/window treatment (if any)”. Most disputes disappear. Quick formulas and numeric example (ready for code/report) Base-10 one-third-octave constants G = 10^(3/10) ≈ 1.995262 r = 10^(1/10) ≈ 1.258925 # adjacent center-frequency ratio k = 10^(1/20) ≈ 1.122018 # edge multiplier about center f1 = fm / k f2 = fm * k Example: the 1 kHz one-third-octave band fm = 1000 Hz f1 = 1000 / 1.122018 ≈ 891.25 Hz f2 = 1000 * 1.122018 ≈ 1122.02 Hz Δf ≈ 230.77 Hz Q ≈ 4.33 OpenTest integrates both methods. Download and get started now -> or fill out the form below ↓ to schedule a live demo. Explore more features and application stories at www.opentest.com. References [1] IEC 61260-1:2014 PDF sample (iTeh): https://cdn.standards.iteh.ai/samples/13383/3c4ae3e762b540cc8111744cb8f0ae8e/IEC-61260-1-2014.pdf [3] ANSI S1.11-2004 preview PDF (ASA/ANSI): https://webstore.ansi.org/preview-pages/ASA/preview_ANSI%2BS1.11-2004.pdf [4] HEAD acoustics Application Note: FFT - 1/n-Octave Analysis - Wavelet (filter bank description): https://cdn.head-acoustics.com/fileadmin/data/global/Application-Notes/SVP/FFT-nthOctave-Wavelet_e.pdf [5] IEC 61672-1:2013 (IEC page): https://webstore.iec.ch/en/publication/5708 [6] NTi Audio Know-how: Fast/Slow time weighting (IEC 61672-1 context): https://www.nti-audio.com/en/support/know-how/fast-slow-impulse-time-weighting-what-do-they-mean [7] MathWorks: ENBW definition example: https://www.mathworks.com/help/signal/ref/enbw.html
Octave-band analysis converts detailed spectra into standardized 1/1- and 1/3-octave bands using constant-percentage bandwidth on a logarithmic frequency axis. In this post, we explain the mathematical basis of CPB, why IEC 61260-1 and ANSI S1.11 define octave bands the way they do, and how band levels are computed in practice (FFT binning vs. filter-bank RMS). The goal: repeatable, comparable results for acoustics, NVH, and compliance measurements. What is octave-band analysis, and what problem does it solve? Octave-band analysis is a family of spectrum analysis methods that partition the frequency axis on a logarithmic scale into band-pass bands. Each band has a constant ratio between its upper and lower cut-off frequencies (constant percentage bandwidth, CPB). Within each band we ignore fine line-spectrum details and focus on total energy / RMS (or power) in that band. In other words, it is not “what happens at every 1 Hz,” but “how energy is distributed across equal relative bandwidths.” This representation naturally matches human hearing and many engineering systems, whose frequency resolution is often closer to a relative (log) scale than a fixed-Hz scale. It is a common reporting format required by many standards: room acoustics parameters, sound insulation ratings, environmental noise, machinery noise, wind/road noise, etc., often use 1/3-octave bands. From linear Hz to log frequency: why CPB looks more like an engineering language Using equal-width frequency bins (e.g., every 10 Hz) to accumulate energy leads to inconsistent behavior across the spectrum: At low frequencies, a 10 Hz bin may be too wide and can smear details. At high frequencies, a 10 Hz bin may be too narrow, giving higher variance and less stable estimates for random noise. In contrast, CPB bandwidth grows with frequency (Δf ∝ f). Each band covers a similar relative change, improving stability and repeatability—important for standardized testing. A visual intuition: bandwidth increases on a linear axis, but is uniform on a log axis Figure 1: the same 1/3-octave bands plotted on a linear frequency axis—bandwidth appears larger at high frequencies Each horizontal segment represents a 1/3-octave band [f1, f2]; the short vertical mark is the band center frequency fm. On a linear axis, higher-frequency bands look wider. Figure 2: the same bands on a logarithmic frequency axis—bands become evenly spaced (the essence of CPB) Once the horizontal axis is logarithmic, these bands appear equal-width/equal-spacing; this is exactly what “constant percentage bandwidth” means. These two figures capture the core idea: octave-band analysis uses equal steps on a log-frequency scale, not equal steps in Hz. Standards and terminology: what do IEC/ANSI/ISO systems actually specify? In practice, “doing 1/3-octave analysis” is constrained by more than just band edges. Standards specify (or strongly imply): how center frequencies are defined (exact vs nominal), the octave ratio definition (base-10 vs base-2), filter tolerances/classes, and even the measurement/averaging conventions used to form band levels. IEC 61260-1:2014 highlights: base-10 ratio, reference frequency, and center-frequency formulas IEC 61260-1:2014 is a key specification for octave-band and fractional-octave-band filters. It adopts a base-10 design: the octave frequency ratio is G = 10^(3/10) ≈ 1.99526 (very close to 2, but not exactly 2). The reference frequency is fr = 1000 Hz. It provides formulas for the exact mid-band (center) frequencies and specifies that the geometric mean of band-edge frequencies equals the center frequency. [1] Key formulas (rearranged from the standard): [1] If the fractional denominator b is odd (e.g., 1, 3, 5, ...): If b is even (e.g., 2, 4, 6, ...): And always: Why does the even-b case look “half-step shifted”? Intuitively, the center-frequency grid is evenly spaced on log(f). When b is even, IEC chooses a half-step offset relative to fr so that band edges align more neatly in common reporting conventions. In practice, a robust implementation is to generate the exact fm sequence using the standard’s formula, then compute edges via f1 = fm / G^(1/(2b)) and f2 = fm * G^(1/(2b)), and only then label bands by the usual nominal frequencies. View the data with OpenTest (IEC 61260-1 Octave-Band Analysis) -> Band edges, center frequency, and the bandwidth designator b Standards commonly use 1/b as the “bandwidth designator”: 1/1 is one octave, 1/3 is one-third octave, etc. [1] Once (G, b, fr) are chosen, the entire band set (centers and edges) is fixed mathematically. Exact vs nominal: why two “center frequencies” appear for the same band “Exact” center frequencies are used for mathematically consistent definitions and filter design; “nominal” values are used for labeling and reporting. [1] ISO 266:1997 defines preferred frequencies for acoustics measurements based on ISO 3 preferred-number series (R10), referenced to 1000 Hz. [2] As a result, the exact geometric sequence is typically labeled with familiar nominal values such as: 20, 25, 31.5, 40, 50, 63, 80, 100, 125, 160, …, 1k, 1.25k, 1.6k, 2k, 2.5k, 3.15k, …, 20k. Implementation tip: compute edges from exact frequencies; only round/display as nominal. This avoids drifting away from the standard. Base-10 vs base-2: why standards don’t insist on an exact 2:1 octave Although “octave” is often thought of as 2:1, IEC 61260-1 specifies base-10 (G=10^(3/10)) rather than G=2. Key motivations include: Alignment with decimal preferred-number series (ISO 266 is tied to R10). [2] International consistency: IEC 61260-1:2014 specifies base-10 and notes that base-2 designs are less likely to remain compliant far from the reference frequency. [1] In base-10, one-third octave corresponds to 10^(1/10) ≈ 1.258925 (also interpretable as 1/10 decade), which yields a clean mapping: 10 one-third-octave bands per decade. “10 one-third-octave bands = 1 decade”: why this matters With base-10 one-third-octave spacing, each step multiplies frequency by r = 10^(1/10). Therefore: 10 consecutive 1/3-octave bands multiply frequency by exactly 10 (one decade). This matches ISO 266/R10 conventions and simplifies tables, plotting, and communication. Standardization values readability and consistency as much as raw mathematical purity. Figure 3: Base-10 one-third-octave spacing—10 equal ratio steps per decade (×10 in frequency) ANSI S1.11 / ANSI/ASA S1.11: tolerance classes and a transient-signal caution ANSI S1.11 (and later ANSI/ASA adoptions aligned with IEC 61260-1) specify performance requirements for filter sets and analyzers, including tolerance classes (often class 0/1/2 depending on edition). [3][4] A practical caution in ANSI documents: for transient signals, different compliant implementations can produce different results. [3] This highlights that time response (group delay, ringing, averaging time constants) matters for transient analysis. What do class/mask/effective bandwidth actually control? “I used 1/3-octave bands” is not just about nominal band edges. Standards aim to ensure different instruments/algorithms yield comparable results by constraining: Frequency spacing: center-frequency sequence and edge definitions (base-10, exact/nominal, f1/f2). Magnitude response tolerance (mask): allowable ripple near passband and required attenuation away from center. Energy consistency for broadband noise: constraints on effective bandwidth so band levels are comparable across implementations. Effective bandwidth matters because real filters are not ideal brick walls. For broadband noise, the output energy depends on ∫|H(f)|^2 S(f)df. Differences in passband ripple, skirts, and roll-off can cause systematic offsets. Standards constrain effective bandwidth to keep such offsets within acceptable limits. [1][3][4] The transient caution is not a contradiction: masks mainly constrain steady-state frequency-domain behavior, while transients depend on phase/group delay, ringing, and time averaging. [3] Mathematics: band definitions, bandwidth, Q, and band indexing CPB and equal spacing on a log axis CPB is equivalent to equal-width spacing in log-frequency. If u = log(f), then every band spans a fixed Δu. Many spectra (e.g., 1/f-type) look smoother and statistically more stable in log frequency. Band-edge formulas from the geometric-mean definition (general 1/b form) IEC defines the center frequency as the geometric mean of the edges: fm = sqrt(f1 f2). [1] For 1/b octave bands, the edge ratio is typically f2/f1 = G^(1/b), where G is the octave ratio. Then: For base-10 one-third octave (b=3): G=10^(3/10). Adjacent center ratio is r = G^(1/3) = 10^(1/10) ≈ 1.258925; edge multiplier is k = 10^(1/20) ≈ 1.122018. Q-factor and resolution: octave analysis is constant-Q analysis Define Q = fm / (f2 − f1). For CPB bands, Δf = f2 − f1 scales with fm, so Q depends only on b and G (not on frequency). Quick reference (base-10, fr=1000 Hz): Fractional-octaveBand ratio f2/f1Relative bandwidth Δf/fmQ = fm/Δf1/11.9952620.7045921.4191/21.4125380.3471072.8811/31.2589250.2307684.3331/61.1220180.1151938.6811/121.0592540.05757317.369 Interpretation: for 1/3 octave, Q≈4.33 and each band is about 23% wide relative to its center. Finer bands (1/6, 1/12) give higher resolution but higher variance for random noise and typically require longer averaging. Band numbering (integer index) and formulaic enumeration Implementations often use an integer band index x. In IEC, x appears directly in the center-frequency formula: fm = fr * G^(x/b). [1] This provides a stable way to enumerate all bands covering a target frequency range and ensures contiguous, standard-consistent edges. For base-10: so and you can invert as Figure 4: Q factor for common fractional-octave bandwidths (base-10 definition) Two meanings of “1/3 octave”: base-2 vs base-10—do not mix them Some literature uses base-2: adjacent centers are 2^(1/3). IEC 61260-1 and much modern acoustics practice use base-10: adjacent centers are 10^(1/10). A quick check: if nominal centers look like 1.0k → 1.25k → 1.6k → 2.0k (R10 style), it is likely base-10. Mathematical definition of band levels: from PSD integration to dB reporting Continuous-frequency view: integrate PSD within the band Octave-band level is essentially the integral of power spectral density over a frequency band. For sound pressure p(t): For vibration (velocity/acceleration), the same logic applies with different units and reference quantities. Key point: because dB is logarithmic, any summation or averaging must be performed in the linear power/mean-square domain first. Two discrete implementations: filter-bank RMS vs FFT/PSD binning Filter-bank method: y_b(t)=BandPass_b{x(t)}, then compute mean(y_b^2) as band mean-square (optionally with time averaging). FFT/PSD binning method: estimate S_pp(f) (e.g., via periodogram/Welch), then numerically integrate/sum bins within [f1,f2]. For long, stationary signals, averaged results can be very close. For transients, sweeps, and short events, they often differ. Be explicit about what spectrum you have: magnitude, power, PSD (and dB/Hz) Magnitude spectrum |X(f)|: amplitude units (e.g., Pa), useful for tones/harmonics. Power spectrum |X(f)|²: mean-square units (Pa²). Power spectral density (PSD): mean-square per Hz (Pa²/Hz), most common for noise. Because octave-band levels represent band mean-square/power, you must end up integrating/summing in Pa² (or analogous) regardless of starting representation. Frequency resolution and one-sided spectra: Δf, 0..fs/2, and the “×2” rule FFT bin spacing is Δf = fs/N. A typical discrete approximation is: If you use a one-sided spectrum (0..fs/2), to conserve energy you typically multiply all non-DC and non-Nyquist bins by 2 (because negative-frequency power is folded into the positive side). Different software handles these conventions differently, so align definitions before comparing results. Window corrections: coherent gain (tones) vs ENBW (noise) are different Windowing reduces spectral leakage but changes scaling: For tone amplitude: correct by coherent gain (CG), often CG = sum(w)/N. For broadband noise/PSD: correct by equivalent noise bandwidth (ENBW), e.g., ENBW = fs·sum(w²)/(sum(w))². [9] CG controls peak amplitude; ENBW controls average noise-floor area. Octave-band levels are energy statistics and are more sensitive to ENBW. WindowCoherent Gain (CG)ENBW (bins)Rectangular1.0001.000Hann0.5001.500Hamming0.5401.363Blackman0.4201.727 Partial-bin weighting: what to do when band edges do not align to FFT bins Band edges rarely land exactly on bin frequencies. Treat PSD as approximately constant within each bin of width Δf, and weight boundary bins by their overlap fraction: This produces smoother, more physically consistent band levels when N or band edges change. Figure 5: Partial-bin weighting schematic when band edges do not align with FFT bins A unifying formula: both methods compute ∫|H_b(f)|² S_xx(f) df Both filter-bank and PSD binning can be written as: Brick-wall binning corresponds to |H_b|² being 1 inside [f1,f2] and 0 outside. A true standards-compliant filter has a roll-off and ripple, which is why standards constrain masks and effective bandwidth. Band aggregation: composing 1-octave from 1/3-octave, and forming total levels Under ideal partitioning and energy accounting: Three adjacent 1/3-octave bands can be combined to approximate one full octave band. Summing all band energies over a covered range yields the total energy. Always combine in the energy domain. If L_i are band levels in dB, energies are E_i = 10^(L_i/10). Then: IEC 61260-1 notes that fractional-octave results can be combined to form wider-band levels. [1] Effective bandwidth: why standards specify it Real filters are not ideal rectangles. For white noise (constant PSD S0), output mean-square is: For non-white spectra such as pink noise (PSD ~ 1/f), standards may define normalized effective bandwidth with weighting to maintain comparability across typical engineering noise spectra. [1] Practical implication: FFT “hard-binning” implicitly assumes a brick-wall filter with B_eff = (f2 − f1). A compliant octave filter has skirts, so B_eff can differ slightly (and by class). To match results, either approximate the standard’s |H(f)|² in the frequency domain or document the methodological difference. Why 1/3 octave is favored (math + perception + engineering trade-offs) Information density is “just right”: finer than 1 octave, steadier than very fine fractions A single octave band can be too coarse and hide spectral shape; very fine fractions (e.g., 1/12, 1/24) can be unstable and expensive: Higher estimator variance for random noise (each band captures less energy). More computation and higher reporting burden. Often more detail than regulations or rating schemes need. One-third octave is the classic compromise: enough resolution for engineering insight, stable enough for standardized measurements, and broadly supported by instruments and software. Psychoacoustics: critical bands in mid-frequencies are close to 1/3 octave Many psychoacoustics references describe ~24 critical bands across the audible range, and in the mid-frequency region the critical-bandwidth is often similar to a 1/3-octave bandwidth. [7][8] This makes 1/3 octave a natural intermediate representation for problems tied to perceived sound, while still being more standardized than Bark/ERB scales. Direct standards/application pull: many workflows mandate 1/3 octave I/O Once major standards define inputs/outputs in 1/3 octave, ecosystems (instruments, software, reporting templates) converge around it. Examples: Building acoustics ratings: ISO 717-1 references one-third-octave bands for single-number quantity calculations. [5] Room acoustics parameters (e.g., reverberation time) are commonly reported in octave/one-third-octave bands (ISO 3382 series). [6] Extra base-10 benefits: R10 tables, 10 bands/decade, readability 10 bands per decade: multiplying frequency by 10 corresponds to exactly 10 one-third-octave steps (very clean for log plots). R10 preferred numbers: 1.00, 1.25, 1.60, 2.00, 2.50, 3.15, 4.00, 5.00, 6.30, 8.00 (×10^n) are widely recognized and easy to communicate. Compared with base-2, decimal labeling is less awkward and cross-standard ambiguity is reduced. Octave-band analysis is typically implemented using either FFT binning or a filter bank. Keep reading -> Octave-Band Analysis Guide: FFT Binning vs. Filter Bank OpenTest integrates both methods. Download and get started now -> or fill out the form below ↓ to schedule a live demo. Explore more features and application stories at www.opentest.com. References [1] IEC 61260-1:2014 PDF sample (iTeh): https://cdn.standards.iteh.ai/samples/13383/3c4ae3e762b540cc8111744cb8f0ae8e/IEC-61260-1-2014.pdf [2] ISO 266:1997, Acoustics - Preferred frequencies (ISO): https://www.iso.org/obp/ui/ [3] ANSI S1.11-2004 preview PDF (ASA/ANSI): https://webstore.ansi.org/preview-pages/ASA/preview_ANSI%2BS1.11-2004.pdf [4] ANSI/ASA S1.11-2014/Part 1 / IEC 61260-1:2014 preview: https://webstore.ansi.org/preview-pages/ASA/preview_ANSI%2BASA%2BS1.11-2014%2BPart%2B1%2BIEC%2B61260-1-2014%2B%28R2019%29.pdf [5] ISO 717-1:2020 abstract (mentions one-third-octave usage): https://www.iso.org/standard/77435.html [6] ISO 3382-2:2008 abstract (room acoustics parameters): https://www.iso.org/standard/36201.html [7] Ansys Help: Bark scale and critical bands (mentions midrange close to third octave): https://ansyshelp.ansys.com/public/Views/Secured/corp/v252/en/Sound_SAS_UG/Sound/UG_SAS/bark_scale_and_critical_bands_179506.html [8] Simon Fraser University Sonic Studio Handbook: Critical Band and Critical Bandwidth: https://www.sfu.ca/sonic-studio-webdav/cmns/Handbook5/handbook/Critical_Band.html [9] MathWorks: ENBW definition example: https://www.mathworks.com/help/signal/ref/enbw.html
From the outside, a measurement microphone looks deceptively simple. But in real-world engineering, its interface options are surprisingly diverse: Lemo, BNC, Microdot, 10-32 UNF, M5, SMB… Many newcomers to acoustics ask questions like: Why can’t microphone interfaces be standardized? Why are cables often not interchangeable between microphones? What power and signal schemes are hidden behind different connectors? This article provides a structured overview of common measurement microphone interfaces, looking at physical connectors, powering methods, cable characteristics, and typical application-driven selection. Main Physical Interfaces for Measurement Microphones Below is a connector-by-connector summary, including the typical powering approach for each. Lemo (5-pin, 7-pin): The Classic Solution for Externally Polarized Microphones Lemo is a precision circular multi-pin connector and is the most common choice for externally polarized measurement microphones. The Lemo B series is widely used (e.g., 0B, 1B, 2B), and most standard measurement microphones adopt the Lemo 1B interface. Key Characteristics A multi-pin connector can carry multiple signals simultaneously, such as: Microphone output (analog signal) External polarization high voltage (typically 200 V) Preamplifier power supply Calibration/identification signals Additional benefits: Very reliable mechanical locking Well-suited for lab environments, metrology, and semi-anechoic chamber measurements where stability and traceability matter Notes on External Polarization Common polarization voltage is 200 V; some systems support switching between 0 V / 200 V Polarization voltage stability affects microphone sensitivity; in engineering practice, sensitivity variation is often treated as approximately proportional to voltage variation The preamplifier is typically powered separately (up to 120 V) but delivered via the same multi-pin connector Maximum output voltage can reach 50 Vp Includes pins for charge injection methods Separate output and ground paths help achieve lower noise In metrology labs, type testing, acoustic calibration, and high-precision semi-anechoic chamber work, the combination of “externally polarized microphone + Lemo multi-pin connector” is essentially a standard configuration. When not to use Lemo: Harsh environments with heavy contamination, oil exposure, and salt spray High costs of cables and connectors require careful trade-offs in field engineering applications BNC: The Most Common External Connector for IEPE Microphones Names like IEPE / ICP / CCP refer to the same general technology route: constant-current powering, where power and signal are transmitted on the same line (Constant Current Powering). In this system, the most common physical connector is the coaxial BNC. Interface and Powering Characteristics Coaxial structure, ideal for analog voltage transmission Bayonet lock (quick and reliable plug/unplug) Supports longer cable runs with good noise immunity Low cost and highly universal Typical IEPE Powering Parameters Constant current: 2–20 mA (common settings include 2 mA, 4 mA, 8 mA, etc.) Compliance voltage (supply capability): typically 18–24 V Maximum output voltage: generally around 8 Vp If the constant current is too low or the compliance voltage is insufficient, the maximum output signal swing is limited—directly affecting the maximum measurable SPL and the linear measurement range. In everyday testing such as engineering noise measurements, NVH, and environmental noise work, “IEPE microphone + BNC” has become the de facto standard. When not to use BNC: Applications requiring long-distance transmission of high-frequency signals, where signal attenuation becomes significant Applications involving frequent plugging and unplugging, to avoid an increased risk of poor electrical contact Microdot (10-32 UNF / M5): Lightweight Connectivity for Small Microphones Microdot is a threaded miniature coax connector widely used for small sensors (compact measurement microphones, accelerometers, etc.). It commonly uses a 10-32 UNF thread. What “10-32 UNF” Really Means This is simply an imperial fine-thread standard: Nominal diameter: 0.19 inch ≈ 4.826 mm Pitch: 1/32 inch ≈ 0.7938 mm Because 10-32 UNF is the typical thread used on Microdot connectors, the term “10-32 UNF” is often used informally to refer to the Microdot interface itself. What about M5? M5 is a metric thread standard: Nominal diameter: 5 mm Pitch: 0.8 mm Its dimensions are close to 10-32 UNF, and when tolerances are not extremely strict it can serve as a substitute—commonly seen in accelerometers or vibration microphones. Interface Characteristics Very compact; ideal for lightweight setups Threaded locking provides strong mechanical stability Commonly paired with IEPE powering Best for short runs and high-speed signal transmission When microphones must be placed in tight spaces, or where sensor mass/size is critical, Microdot is a common choice for compact, high-density installations. When not to use Microdot: Applications requiring quick connect/disconnect or frequent sensor replacement Use in systems with low constraints on installation space and requiring large-size connectors or high-power transmission, to avoid increased connection complexity and cost SMB (SubMiniature B): For High-Density Multi-Channel or Internal Connections SMB is a small “push-on” coaxial connector. Interface Characteristics Compact size supports high channel density Push-on structure enables fast connection Better high-frequency performance than BNC More suitable for semi-permanent internal wiring SMB is often best viewed as an engineering connector used inside equipment, rather than a field-plugging standard. When not to use Microdot: Applications involving frequent plugging and unplugging or repeated mechanical stress Use as a front-end connection interface for external devices, to avoid structural damage and reduced reliability Extended Interface Function: TEDS and Smart Identification In multi-channel and integrated systems, TEDS (Transducer Electronic Data Sheet) is increasingly common. By integrating a small memory chip into the sensor or cable, TEDS can store: Model and serial number Sensitivity Calibration date and other parameters Compatible front-end hardware or acquisition software can automatically read TEDS to: Identify the sensor type on each channel Load sensitivity and calibration coefficients automatically Reduce manual entry errors Save calibration time and labor At the connector level, TEDS is typically implemented by using certain pins in multi-pin Lemo connectors, or via overlay methods in specific BNC-based solutions. When planning an interface system, it’s wise to consider early on whether TEDS support is required. Why Are There So Many Interfaces? Connector diversity is best explained from three perspectives: Different Polarization and Powering Schemes Externally polarized microphones (≈ 200 V polarization) → better suited to multi-pin connectors like Lemo Prepolarized + IEPE systems → better suited to coaxial connectors like BNC / Microdot / SMB Different Scenarios and Priorities Laboratory / metrology: high stability, multiple signals in one cable, secure locking → Lemo Field engineering / environmental measurement: convenient wiring, strong universality → BNC + IEPE Miniaturization / high-density arrays: size and channel density first → Microdot / SMB Long Product Lifecycles and Backward Compatibility Measurement systems often have lifecycles of 10–20 years or more To avoid forcing users to replace large numbers of cables and front-end systems, manufacturers typically continue existing interface ecosystems Under long lifecycle constraints, “full unification” is often impractical and offers limited engineering return Typical Application Mapping (Quick Reference) Engineering noise, NVH, vibration/noise tests: BNC / MicrodotEasy wiring, many channels, low maintenance cost Precision lab measurement, type testing, metrology calibration: Lemo 7-pin / 5-pinSupports polarization HV and multiple signals; suitable for traceable high-precision measurement Acoustic arrays, multi-channel acquisition card systems: Microdot / SMBHigh channel density, compact wiring, easier system integration Long-term environmental noise monitoring systems: BNC / customized protected connectorsFocus on weather resistance, waterproofing, salt fog resistance, and stable long-distance transmission Conclusion The variety of measurement microphone interfaces is mainly the result of trade-offs between technology routes, application requirements, and historical compatibility—not simply a “lack of standards”. Taking NVH testing as an example: if an existing system uses BNC connectors to connect accelerometers, high-frequency signal attenuation and intermittent contact issues may occur in multi-channel array measurements. To improve connection reliability and signal quality, LEMO connectors with locking mechanisms and superior vibration resistance should be selected. After replacement, signal transmission stability is significantly improved, noise interference is reduced, and the consistency of test data is enhanced. You are welcome to learn more about microphone functions and hardware solutions on our website and use the “Get in touch” form to contact the CRYSOUND team.