www.eejournal.com
Open in
urlscan Pro
68.168.102.15
Public Scan
Submitted URL: https://cmszf04.na1.hubspotlinks.com/Ctc/W1+113/cMsZf04/VWx9bv1VpRfnW1xhnVF4VSGdhW3S9mTt50V_QLMkNllX3q90JV1-WJV7CgK93N3p1y-cgtzdpW4wg...
Effective URL: https://www.eejournal.com/article/ai-augmented-earbuds-that-read-your-mind/?utm_campaign=One%20More%20Epoch&utm_medium=ema...
Submission: On July 14 via api from GB — Scanned from GB
Effective URL: https://www.eejournal.com/article/ai-augmented-earbuds-that-read-your-mind/?utm_campaign=One%20More%20Epoch&utm_medium=ema...
Submission: On July 14 via api from GB — Scanned from GB
Form analysis
2 forms found in the DOMGET https://www.eejournal.com
<form role="search" method="get" class="search-form" action="https://www.eejournal.com">
<label>
<span class="screen-reader-text">Search for:</span>
<input type="search" class="search-field" placeholder="Search …" value="" name="s">
</label>
<button type="submit" class="search-submit"><span class="screen-reader-text">Search</span></button>
</form>
POST https://techfocus.activehosted.com/proc.php
<form method="POST" action="https://techfocus.activehosted.com/proc.php" id="_form_64B15A9756F05_" class="_form _form_1000 _inline-form _dark" novalidate="" data-styles-version="0">
<input type="hidden" name="u" value="64B15A9756F05" data-name="u">
<input type="hidden" name="f" value="1000" data-name="f">
<input type="hidden" name="s" data-name="s">
<input type="hidden" name="c" value="0" data-name="c">
<input type="hidden" name="m" value="0" data-name="m">
<input type="hidden" name="act" value="sub" data-name="act">
<input type="hidden" name="v" value="2" data-name="v">
<input type="hidden" name="or" value="0ae685f04901f28cf76f2a3212a46d1a" data-name="or">
<div class="_form-content">
<div class="_form_element _x95353959 _full_width ">
<label for="email" class="_form-label">Email*</label>
<div class="_field-wrapper">
<input type="text" id="email" name="email" placeholder="" required="" data-name="email">
</div>
</div>
<div class="_form_element _x95484604 _full_width ">
<label for="firstname" class="_form-label">First Name</label>
<div class="_field-wrapper">
<input type="text" id="firstname" name="firstname" placeholder="Type your first name" data-name="firstname">
</div>
</div>
<div class="_form_element _x88705123 _full_width ">
<label for="lastname" class="_form-label">Last Name</label>
<div class="_field-wrapper">
<input type="text" id="lastname" name="lastname" placeholder="Type your last name" data-name="lastname">
</div>
</div>
<div class="_form_element _x23199688 _full_width ">
<label for="ls" class="_form-label">Enter the following to confirm your subscription*</label>
<div class="g-recaptcha" data-sitekey="6LcwIw8TAAAAACP1ysM08EhCgzd6q5JAOUR1a0Go" id="recaptcha_0">
<div style="width: 304px; height: 78px;">
<div><iframe title="reCAPTCHA"
src="https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LcwIw8TAAAAACP1ysM08EhCgzd6q5JAOUR1a0Go&co=aHR0cHM6Ly93d3cuZWVqb3VybmFsLmNvbTo0NDM.&hl=en&v=khH7Ei3klcvfRI74FvDcfuOo&size=normal&cb=l6iq4r3c0thf"
width="304" height="78" role="presentation" name="a-6w9v1ks9pvr5" frameborder="0" scrolling="no"
sandbox="allow-forms allow-popups allow-same-origin allow-scripts allow-top-navigation allow-modals allow-popups-to-escape-sandbox"></iframe></div><textarea id="g-recaptcha-response" name="g-recaptcha-response"
class="g-recaptcha-response" style="width: 250px; height: 40px; border: 1px solid rgb(193, 193, 193); margin: 10px 25px; padding: 0px; resize: none; display: none;"></textarea>
</div><iframe style="display: none;"></iframe>
</div>
</div>
<div class="_button-wrapper _full_width"><button id="_form_1000_submit" class="_submit" type="submit">Subscribe</button></div>
<div class="_clear-element"></div>
</div>
<div class="_form-thank-you" style="display:none;"></div>
<input type="hidden" name="pum_form_popup_id" value="74267">
</form>
Text Content
Search for: Search Menu * Articles * News * Fish Fry * Chalk Talks * Blogs * Categoriesexpand child menu * Category List * Media Kit Skip to content feature article June 22, 2023 AI-AUGMENTED EARBUDS THAT READ YOUR MIND! by Max Maxfield I was 16 or so years old when I built my first brainwave amplifier circa 1973. This was prior to the widespread availability of microcontrollers. As far as I know, it was also before anyone had even coined the term DSP (digital signal processing), because—to the best of my knowledge—all signal processing at that time was ASP (analog signal processing), which was performed using analog components and techniques. My humble brainwave amplifier artifact required a cornucopia of electrodes to be clamped around my cranium, buffered with conductive paste, and connected to my noggin by surprisingly sticky sticky tape. Why is it that such tape is always at its stickiest when you don’t want it to be while managing to be at its least sticky when you need it the most? Suffice it to say that once the sensors were firmly attached, I didn’t feel a great urge to take them off again in a hurry. The idea was to detect and filter out theta and alpha brain waves, which increase when you enter a meditative state, amplify them, and use them to stimulate a pink noise generator into outputting a relaxing “chuff, chuff, chuff…” sound a bit like an old-fashioned steam locomotive (see also Colors of Noise). Theoretically, after practicing for only a few weeks, it should have been possible for me to out-meditate a Shaolin monk. In practice… well, that’s a tale that remains to be told. Suffice it to say that my brainwave amplifier did not distinguish itself on either the “small” or “lightweight” fronts. So, you can only imagine my surprise to discover that various groups are currently working on creating AI-augmented earbuds that can monitor your brainwaves. “Why on earth would anyone want to do this?” I hear you cry. Well, I shall expound, explicate, and elucidate shortly, but first… I was just chatting with Martin Croome from GreenWaves Technologies, which is a fabless chip manufacturer that was founded in 2014 and is based in Grenoble, France. Martin dons a strange double hat in the company—on the one hand, he’s the VP of marketing; on the other hand, he’s the chief architect for the company’s neural network development tools (and I thought my life was confusing). The core mission for the folks at GreenWaves is to design, develop, and market extreme processors for energy-constrained devices. In this context, an energy-constrained device may be something tiny, like an earbud that must keep running for a day or more on a single charge, or it may be something like an IoT sensor node that has larger batteries but that must truck along for years or even decades. Their initial product offering, the GAP8, has been in production since 2020. As one of the very first AI-enabled RISC-V-based microcontrollers on the market, the GAP8 is currently deployed in a wide variety of products. In the case of their second-generation product, the GAP9, which is billed as an ultra-low-power AI-and-DSP-enabled RISC-V-based microcontroller, they’ve been sampling since the beginning of last year, they currently have production wafers in their hands, and their customers will be launching GAP9-based products later this year. Just to set the scene, the GAP8 boasts 22.65 giga operations per second (GOPS) at 4.24mW/GOP, while the GAP9 flaunts 150.8 GOPS at 0.33mW/GOP. An example GAP8-based application that has been shipping for some time is shown below. Ceiling-mounted people detector (Source: GreenWaves) Using a GreenWaves’ GAP8 as its brains, this little beauty—a ceiling mounted infrared (IR) people detecting sensor—was created by a company called Kontakt.io (there’s little wonder I have so many problems with spelling). The sensor wakes up once a minute to count the number of people in the room and convey this information “upstairs.” Each sensor is also a Bluetooth mesh node. The sensors pass their data from node to node until it reaches a gateway (e.g., Wi-Fi or cellular) that can communicate the data to a local fog or a global cloud. There are many applications for this sort of thing, such as meeting room utilization. For example, is this meeting room really getting used? And, if it is being used, how many people typically use it at any particular time? (If it’s a 10-person room but 99% of the time there are no more than 5 people in it, then maybe it should be divided into two meeting rooms.) Other applications include things like cafeteria utilization (“Hey, this would be a good time for you to go to the cafeteria because there aren’t many people in it”) and smart cleaning (“No one has used these toilets since they were last cleaned, so there’s no point in cleaning them again”), and so forth. As I pen these words, GreenWaves’ partners are feverishly developing products based on GAP9 samples. The folks at GreenWaves say they can’t say (no pun intended) who their GAP9 customers are, but they can say that they’ve already had 10+ design wins, and this is for a device that’s not yet in production, which is rather exciting. The GAP9 next to a Euro cent (Source: GreenWaves) Let’s take a slightly deeper dive into this little rascal. In a crunchy nutshell, the GAP9 essentially embodies three different processing capabilities. The first is real-time streamed autonomous time domain digital signal processing, which is predominantly focused on audio. This employs an in-house developed IP called the smart filtering unit (SFU), which we will discuss in more detail a little later. The second processing capability is general-purpose digital signal processing (DSP) like codecs, FFTs, frequency-based filtering, frequency domain filtering… that kind of stuff. This is performed in what they call a multi-core RISC-V computational cluster. Last, but certainly not least, the third processing capability involves the combination of the aforementioned computational cluster with a neural network accelerator called the NE16. In line with our earlier discussions, all these processing capabilities are focused on delivering extreme performance whilst consuming very, very low power for energy-constrained devices. Martin told me that, in addition to offering lower latency and lower power consumption to pretty much anything else on the market, a key differentiator is that “GAP processors are easy to program, as compared to most of the alternatives.” Martin also says, “GAP processors are very flexible, which allows them to cope with accelerating the types of things that are happening today, not the things which happened last year.” By this he means that these processors can implement all the latest and greatest types of neural networks, such as recurrent, convolutional, transformer, etc. So, what does the GAP9 look like inside? Well, as we see in the diagram below, there are fundamentally two parts to the chip that we might call “the left side” and “the right side” (stop me if I’m getting too technical). High level block diagram of the GAP9 (Source: GreenWaves) The left side looks fundamentally like a regular MCU. What they call the “Fabric Controller” (for historical or hysterical reasons, I forget Martin’s exact wording) is basically a RISC-V core. They’ve exploited the fact that you can extend the instruction set architecture (ISA) of the RISC-V, and they’ve employed this capability to add their own custom instructions for things like DSP bit manipulation, lightweight vectorization in the cores, and what they call transprecisional floating-point, which means they can do 32-bit floating point along with two different 16-bit floating point representations in the form of IEEE FP16 and BF16 (these formats were discussed in more detail in my Mysteries of the Ancients: Binary Coded Decimal (BCD) column). Meanwhile, on the right-hand side we find an additional nine RISC-V cores (this is the “computational cluster” we were waffling about earlier). This means that there are 10 RISC-V cores on the chip, all identical, not only in terms of their instruction sets, but also in terms of their memory map. In turn, this means you can write a function, compile it once, and run it anywhere on the chip. Adjacent to the computational cluster on one side is the NE16 AI accelerator. On the other side, we find a hardware synchronization block, which implements all kinds of synchronization primitives—forks, joins, barriers, mutexes, so on and so forth—in hardware. This facilitates extremely fine-grained parallelism, allowing multiple cores to be working on different facets of the same task (like an FFT) or in a pipelined fashion, or on completely different tasks. In addition to clock gating all over the device (there are multiple clock and voltage domains across the chip), the GAP9 also supports dynamic voltage and frequency scaling (DVFS), with all this power control taking place inside the chip. Also, there’s an abundance of external interfaces, including a MIPI interface for a camera, three different serial audio interfaces with time-division multiplexing (TDM) support for up to 16 channels on each. Three pulse-density modulation (PDM) inputs and one PDM output, which can support a total of nine PDM microphones in and three PDM sources or syncs out. And then there’s the smart filtering unit (SFU) that we mentioned earlier. Located on the left-hand-side of the device, this is a completely autonomous unit that can process from interface-to-interface, or interface-to-memory, or memory-to-interface, or memory-to-memory (it can also be working on multiple streams simultaneously). Block-ish diagram of the GAP9’s smart filtering unit (Source: GreenWaves) One key aspect to this block is its interface-to-interface capability. This comes into play when we want to do something like active noise cancellation (ANC), which demands extremely low latency. One of the big markets for GAP processors is hearables, and one of the key areas for hearables is ANC. Why would you want to put ANC in a chip that can implement neural networks? Well, because you can use the neural networks to steer the ANC. To do this, the primary path of the ANC must be fast enough that—between the time the signal takes to go from the error microphone into the user’s ear—we can generate the anti-noise to cancel it. Of course, this technology can be used for lots of other things, including heavy-duty filtering, spatialization, sound effects, etc. So, what are the results of all this? Well, I’m just about to show you. MLCommons is an open engineering consortium with a mission to benefit society by accelerating innovation in machine learning (ML). As part of this, MLCommons publishes the MLPerf Tiny benchmarks for embedded devices (these benchmarks are available from GitHub). The results from two of these benchmarks are shown below. GAP9 vs. competitor benchmark results (Source: GreenWaves) The results were that the GAP9 consumed ~2 to 3X less energy than a highly specialized analog neuromorphic chip and ~2 to 4X less energy than a specialized neural network accelerator chip, all while providing the lowest latency in all the tests. The image above reflects only two of the four benchmarks. Martin notes that the GAP9 provided the best latency and best energy out of all the contenders on all four benchmarks except the anomaly detection benchmark. However, the company whose device reported the best latency on this benchmark didn’t report energy, which must be viewed as being a tad suspicious. So, if we limit ourselves to those companies that reported both energy and latency values, the GAP9 was the best in latency and energy on all benchmarks across the board. All of which brings us back to thinking what can be achieved with GAP9 processors being used to power true wireless stereo (TWS) earbuds. GAP9 deployed in TWS earbuds (Source: GreenWaves) In turn, this brings us back to my meandering mutterings at the start of this column when I made mention of the fact that various groups are currently working on creating AI-augmented earbuds that can monitor your (well, our) brainwaves. I just had a quick Google while no one was looking and immediately found a couple of somewhat related articles: Google Spinoff Working on Earbuds That Spy on Your Brain Signals and Earbuds That Read Your Mind. Martin tells me that the guys and gals at GreenWaves are working with another company to create earbuds for the 21st century. In addition to acting as regular TWS earbuds and/or hearing aids, these will have sensors involving flexible contacts on the rubber that you put in your ear, thereby allowing them to act like an electroencephalogram (EEG) to pick up your brainwaves. There are obvious medical applications for this sort of thing, like epilepsy detection. But the thing that piqued my interest came in the context of hearing aids that can address the “cocktail party problem,” which involves trying to pick out one voice when multiple people are speaking at the same time, like at a party, for example. I know several people who are hard of hearing. Can you imagine earbuds with processing power sufficient to disassemble the sound space into individual voices? Now, take this one step further and imagine using AI to monitor the brainwave signals to determine which of these voices is of particular interest to the earbud owner (well, wearer), and to selectively boost that voice whilst fading back any extraneous sounds including any other speakers. All I can say is “Wowsers!” (and I mean that most sincerely). The more I think about this, the more applications come to mind, like gaming, for example (your earbuds could detect changes in your brainwave patterns and signal information to your game controller—I’m not sure where I’m going with this, but it’s both exhilarating and scary at the same time). Now I’m thinking about my recent The End of the Beginning of the End of Civilization as We Know It? column. Supposing a generative AI like ChatGPT were using the signals from your EEG earbuds to monitor your brainwaves and… I’ll let your imagination run wild from here. As always, please do share any thoughts you have on anything you’ve read here in the comments below (as quickly as you can before ChatGPT gets to hear about it—again, no pun intended). * * * * * RELATED AGILE ANALOG MAKES THE ANALOG PARTS OF SOC DESIGN EASY PEASY LEMON SQUEEZY I sometimes joke that I’d bend over backwards to be flexible and that I’d give my right arm to be ambidextrous (I didn’t say they were good jokes). Now, following a recent excursion into wibbly-wobbly analog space (where no one can hear you scream), I’m tempted to say that I’d… SAY HOWDY TO HIGH FIDELITY SOFTWARE-DEFINED SENSING FOR THE 21ST CENTURY There are many things I don’t know much (if anything) about. So many, in fact, that I could write a book about them… or not, as the case might be. An example of one of the things I know very little about would be touchscreens, which makes it somewhat paradoxical… NEXT-GEN AI VISION PROCESSORS TARGET EDGE APPLICATIONS It’s been a long time coming, but I have to say that artificial intelligence (AI) is finally starting to make my life easier in meaningful ways. I know, right? I’m as surprised to hear myself saying this as I’m sure you are to be reading it. As a brief aside,… Share LEAVE A REPLY CANCEL REPLY You must be logged in to post a comment. * * * * * featured blogs Joules RTL Design Studio: Accelerating Fully Optimized RTL Jul 14, 2023 Cadence announced Joules RTL Design Studio today at CadenceLIVE Japan 2023, a new product that expands upon Cadence's existing Joules RTL Power Solution. The solution will address all aspects of physical design by adding visibility beyond just power into performance, area, an... More from Cadence... Synopsys & Georgia Tech Team Up on Reinforcement Learning-Driven Research—Win DAC 2023 Best Paper Award Jul 13, 2023 We're experimenting with deep reinforcement learning in chip design alongside Georgia Tech ECE; see why our research won the DAC 2023 Best Paper Award.The post Synopsys & Georgia Tech Team Up on Reinforcement Learning-Driven Research—Win DAC 2023 Best Paper Award ... More from Synopsys... Self-Solving Levitating Rubik’s Cube Jul 11, 2023 Just when I thought I'd seen everything, I ran across a video of a Rubik's cube that solves itself while levitating above the table.... More from Max's Cool Beans... FEATURED VIDEO SYNOPSYS 224G ETHERNET PHY IP INTEROP AT TSMC SYMPOSIUM 2023 Sponsored by Synopsys At TSMC Symposium 2023 we showcased a successful 224G Ethernet PHY IP interop demonstration with backplane channels. Watch the various plots, ADC histogram and excellent eye diagrams results. Learn More FEATURED PAPER SHIFT LEFT WITH CALIBRE TO OPTIMIZE IC DESIGN FLOW PRODUCTIVITY, DESIGN QUALITY, AND TIME TO MARKET Sponsored by Siemens Digital Industries Software Every IC designer strives to create a “clean,” or error-free, cell, block, chiplet, SoC, or 3DIC assembly before passing their work downstream for full signoff verification. However, waiting until signoff verification to find out how well you did is probably the least efficient approach to achieving production-ready layouts, impacting engineer productivity, project schedules, and hardware resources. Click to read more FEATURED CHALK TALK High-Voltage Isolation for Robust and Reliable System Operation Sponsored by Mouser Electronics and Texas Instruments In this episode of Chalk Talk, Amelia Dalton and Luke Trowbridge from Texas Instruments examine the benefits of isolation in high voltage systems. They also explore the benefits of TI’s integrated transformer technology and how TI’s high voltage isolations can help you streamline your design process, reduce your bill of materials, and ensure reliable and robust system operation. Click here for more information about Texas Instruments ISOW774x/ISOW774x-Q1 Quad-Ch Digital Isolators Apr 27, 2023 9,651 views All material on this site copyright © 2003 - 2023 techfocus media, inc. All rights reserved. Privacy Policy Media Kit Contact Us Log In Subscribe to EE Journal Daily Email* First Name Last Name Enter the following to confirm your subscription* Subscribe × We'd like to show you notifications for the latest news and updates. AllowCancel