Removing Background Noise from a Bird Recording

When recording sound out in the real world, you can often run into some unwanted ambient noise in the background. That can mean less clarity for your desired sound. Sometimes the background noise doesn’t seem like much, but when added to other sounds in the mix, it can add up to something really messy sounding. Luckily, these days there are more and more tools to help cut through the noise and get to your desired sound.

Enter: iZotope RX. For those of you who have worked in audio editing and mixing for a while, you are probably already familiar or have heard of iZotope’s vast array of tools for “audio repair.” From eliminating (or at least lessening) unwanted background noise from a dialogue recording, to repairing clipped and distorted audio, iZotope RX has been the industries go-to program for rescuing bad recordings from the depths of audio hell. Although, as mentioned, there are more and more tools that have hit the market recently, such as dxRevive, Acon Extract Dialogue, and various others that work similarly.

In this case, it isn’t dialogue that we are trying to revive — it’s bird noises. This video is a quick demonstration of how Derek was able to take an old recording that Ryan had made over ten years ago on a cheap recorder and make it more usable. As you will see in the video, the bird sounds are plagued by some broadband rumble, hiss, and distant noise from the neighbor in which they were recorded. And possibly some self-noise from the recording device itself. The goal here was to try and eliminate as much of the unwanted background noise as possible and isolate the bird sounds so that they could then be placed into any ambient sound effect you want. For example, a forest with a river, a neighborhood, or even a city with traffic going by. Having a clean, isolated track means that there won’t be any of the unwanted noise to muddy-up your final mix.

For the purposes of a quick demonstration, Derek used some fairly extreme settings to isolate the birds. It is often said with iZotope, “less is more,” and that doing several de-noise passes at more conservative settings usually yields better results. With this case however, we found that actually having those higher settings seemed to work the best, and for the purposes of making the video shorter, it worked just fine.

Although ideally a sound effect is recorded in a controlled environment with high quality equipment, we don’t live in an ideal world. This video demonstrates what this technology is becoming capable of and that some audio issues can be, dare we say… fixed in post.

What We Know About AI Audio (So Far)

 

Yes, we made this image using AI

 

You really can’t talk about media or production these days without someone bringing up artificial intelligence (AI). This year, the news and social media has exploded with talk of programs like ChatGPT, Midjourney, Adobe, DALL-E and countless others that creatives can use to generate images, text, voices, and even pop songs. With AI being one of the key sticking points in both the WGA and SAG/AFTRA stikes, those of us in production and post-production have a lot of questions still yet to be answered. How will we use AI in our creative process in the future? What does this mean for job-security?

Recently, Allyson and I had the chance to virtually attend the AI Creative Summit, put together in partnership with NAB. At the summit, various experts discussed the latest advancements in AI, as well as demonstrated some of the capabilities of this technology within the creative production realm. Our main area of interest, obviously, is how this technology affects audio production. We know that generating a voice over from text has gotten better and better. But one of the latest advancements that has really piqued everyone’s interest, has been something called voice cloning. This is where you take pre-recorded audio of a particular person’s voice that a computer is able to learn and then simulate to read whatever script you want. Most famously, voice cloning was recently used to de-age actor Mark Hamill’s voice in order to recreate a young Luke Skywalker for both “The Mandalorian” and “The Book of Boba Fett” TV series. LucasFilm used a Ukrainian company called Respeecher to recreate the famous actor’s voice by using archived ADR, audio book, and radio play recordings to train their software.

AI was also used to recreate Val Kilmer’s voice for both “Top Gun: Maverick” and “VAL,” a documentary on the actor’s life. Kilmer lost much of his voice years ago after undergoing treatment for throat cancer.

Allyson and I were able to see a few demonstrations of similar voice cloning technology at the summit. One company called Instreamatic can generate text-to-speech advertisements, complete with music, as well as personalized versions for whatever platform the audience is listening on. Although the built-in library of voices that this company can generate sounded pretty good, it was when they demonstrated their voice cloning tool where I thought the effect started to fall apart a bit. The demonstrator uploaded a recording of Will Ferrell playing the character “Ron Burgandy.” The resulting voice (to me), sounded more like Will Ferrell doing a less-than-perfect impression of his character. Perhaps with a longer recording to learn from, it may have sounded better? Hard to say from this short demonstration.

Another company in the generative AI voice realm is one called ElevenLabs. In this demonstration, the host showed the various built-in voices that users can use to read a script - complete with various (although somewhat ambiguously-named) controls that you can adjust to change the style of performance. This program was also able to clone a voice pretty well, however what we noticed was just how difficult it would be to customize the performance. Although there are different settings you can adjust, such as “Stability” and “Style Exaggeration,” these are at best a shot-in-the-dark, guess-and-check method of “refining” the performance to get something that you like.

Screen shot of ElevenLabs user interface

For those of us in the advertising world, where reading a script to an exact time is imperative, ElevenLabs has yet to come out with a control for that. Near as we can tell, you would still need to bring in your resulting audio file into an editing program, chop it up to take out extra long pauses, and time compress things to get them to fit properly. There is also no controls for intonation or inflection. One thing we have learned over years of directing voice talent, is that sometimes the way you read something can completely change the meaning of the message. Listening back to the resulting voice, you occasionally get odd sounding phrasing, or an emphasis on the wrong word, etc. I would be very curious to see a text-to-voice program like this generate a VO that required sarcasm or any sort of play-on-words within a script. In these instances, a professional voice talent’s ability to “wink at the camera” (or in this case a microphone) cannot be understated.

We did notice that the style of the voice that is generated is largely at the mercy of the style of the voice that is put in. Meaning, if you input a slower, more relaxed read for the AI to learn - that’s what the program will generate. If you input a faster, more upbeat read, that’s more likely what you will get when the new recording is generated. Unfortunately, this particular demonstration didn’t include that, although the presenter did mention that he wanted to try this in the future. The presenter also pointed out that if your original recording has small mistakes in it, such as plosives (popping “P” sounds), the program will actually include that in the cloned voice. Therefore it is important to get a clean and high-quality recording to start with.

So what practical applications does a program like ElevenLabs have? Considering the need to have advertising copy fit in a very specific amount of time, while including very specific messaging, we are dubious that this would be a perfect tool for that. However, for corporate videos, or scientific explainer videos, where reading with style or reading to exact time isn’t much of a factor, we could see this as a useful application.

It should be mentioned, however, that we have heard instances of peoples’ voices being illegally uploaded and cloned on several of these types of sites, including ElevenLabs. And according to a voice actor we have spoken with, ElevenLabs has done very little to combat this. We won’t cover the ethical or legal implications of voice cloning technology in this post. That will be a separate post for the future, because there is A LOT to cover there. For the time being, if you are interested in how the voice acting community is trying to work with AI and get a framework in place to help protect artists' rights, we encourage you to visit the website of the National Association of Voice Actors (or NAVA).

As for the other big questions, such as “How will us creatives use AI in our workflows?”— we can reiterate what you have probably already heard others say: “AI is a tool.” When it comes to editing audio, there are AI programs to reduce background noise, edit music to certain lengths, create voices, create volume automation for mixing and much more. All of these are tools for a person to control and make creative decisions when working towards a final, polished product. They are tools to help us work faster and smarter, and when used properly (i.e. to enhance an already original product), simply adds to a professional’s “bag of tricks.”

As for the job security question, that may be a bit more up in the air. For years voice talent and recording studios have been competing with cut-rate internet VO products, where those on a budget, or those who don’t really care how their scripts are read, can get two-takes of an un-directed read from an anonymous speaker with fairly quick turnaround. With the advancements in AI voices, this only adds to the pile of competition. Will studios have to pivot yet again and become more “audio producer” than “engineer?” Will it be our job as experts to be the ones who input prompts and data into AI programs in order to yield the best possible result? Or will studios simply become a novelty for those who like the idea of using old-school methods for creativity? Time will tell, and at the rate AI is advancing, we may find out sooner rather than later. For now, we will try to stay up on the latest and greatest to continue to do what we have already been doing for years: deliver a damn good product.

A Day in the Life of an Audio Engineer

Did you catch our day in the life of an audio engineer post on social media? If not- don’t worry- you can check it out here: a busy day time compressed into 45 seconds!

You probably already know that audio engineers record, mix and do sound design to make your projects stand out. But as the video shows, did you know that here at ADS we go above and beyond to help our clients? Whether that means helping you print and organize scripts so your session runs as efficiently as possible, to timing your spots ahead of recording to flag any issues that might come up- we are here for you. Long time clients have even left sessions entirely up to us, including recording and directing when their schedules get too busy.

We are here to help- just let us know what you need. Head to our contact page to reach out.

 
 
 

Sticky-Shed Syndrome (Or: How To Save Your Old Reel-To-Reel Tape By Baking It)

Sticky-shed syndrome — it’s not just a cool name for a punk band, it’s a real condition that affects old magnetic audio tape and prevents it from being played back properly on a reel-to reel. Without getting overly technical (because let’s be honest, we probably barely understand it ourselves), this is a condition where the binder that holds the magnetic iron oxide coating to the plastic carrier begins to absorb moisture and break down, causing it to separate. This can cause friction, meaning the tape speed will slow down when played back, as well as causing a build-up of residue on the playback heads and other parts of the machine that the tape passes through. This affects older tapes (mostly from the 70s-80s) when manufacturers produced tapes with this particular type of binder, before realizing the mistake and switching to a more stable formula.

Thankfully, some genius found a pretty simple solution, which Ryan recently had the chance to demonstrate in the short video above. Essentially, you bake the tape in an oven at a very low temperature (we have read the range should be between 120-150 degrees Fahrenheit) for anywhere from an hour to two hours. This drives out the moisture, allowing the tape to be played back without the oxide rubbing off and causing too much friction. As Ryan shows us in the video, he was able to do this using a small toaster oven, as well as a digital thermometer to make sure the temperature stays in a good range. We should note that the temperature of the oven does fluctuate up and down. This should be okay, so long as you don’t get it too hot where you begin to melt any plastic parts.

Some people report being able to play back their baked tape several times, or that a tape will remain playable for a couple weeks before it begins to absorb moisture again. However, we recommend transferring the tape to digital right away as to not take any chances. If all goes as planned, you should be able to get a decent enough playback to re-archive a tape to digital.

So if you have any old magnetic audio tape from this era that you are wondering if you can salvage, check out the video to see how it’s done. And for those of you who have been around the Pacific-Northwest for a long time, you may even recognize the old Fred Meyer jingle from the early 1980s.

What Does Your Audio Editor Do For You?

What exactly do we audio editors and mixers do for our clients?  As you may already know, it's a little more than just recording and setting levels.  The above video shows a time-lapse of a typical edit, cleanup, and mix of a simple 30 second radio spot.  In it you can see us edit together various phrases from different takes, adjust levels with volume dynamics, remove breaths, and mix with the music.  This is all standard work to be done when preparing audio for an ad.  However, editing can sometimes get a bit more complex.

The great thing about working in a program like Pro Tools is how detailed you can be in your editing.  If you don't like the sound of certain syllable or consonant - replace it!  If mouth clicks and pops can't be edited out, there's a special tool to get rid of them (in the video we used the pencil tool to actually redraw the waveform).  It's this fine-tuning that helps ensure that the voice over sounds as smooth and clean as possible.

Sometimes script changes call for an emergency re-edit.  For example, a certain word should have been plural, and the talent is unavailable to re-record.  We can find an "s" from somewhere else in the read (or in extreme cases, go into the booth and record it ourselves) and after a bit of finessing, make it sound perfectly natural.

All of this is just a snapshot of the basic tasks we perform on a daily basis.  There is of course a bit more that goes into it, but we won't bore you with technical details.  So the next time you run into an audio engineer, say "thank you!"  But make sure to say it three times in a row so they have choices to edit from.  ;)

The Foley Farm

When it comes to doing Foley sound effects, we're "all ears" (Get it? Because corn?). Sometimes when our effects library just doesn't have the right sound, we get to record our own. Just like we did in this video. Here you will see us creating these farm-fresh sounds and even get a short preview of how it all sounds when it's mixed in (minus the VO and final music). All to make this commercial sound "a-maize-ing." Okay, no more puns, here's the video and more details below:

Throughout this process we discovered, in spite of using the real product, sometimes the sound of shucking corn, doesn't actually sound like corn being shucked. During one of our early takes, we found that grabbing a handful of leaves and pulling them backed created a very squeaky rubbery sound. After some experimenting, we found that pulling just one leaf or even squeezing the loose leaves in our hand yielded a much more gentle and natural sound that you would expect to hear. Suffice it to say that sometimes the actual sound of the action on screen isn't always the best fit for your video.

In terms of mixing -- this was for a commercial, so these sounds will act as added texture underneath voice over and music. The goal here is to be subtle enough to not compete with the message of the spot, but to add a nice layer to the picture that really brings it to life.

Cheers to Twenty-One Years!

One of the many great contributions to culture that Portland, Oregon is famous for (besides lumberjack fashion and vegan-themed gentlemen's clubs) is its beer.  So it only makes sense that when your local PDX audio studio turns twenty-one, you celebrate with a pint of the hoppy good stuff from a local Oregon brewery -- and even better that we get to do so in our new custom glasses!

On April 1st of this year, ADS Recording reached legal drinking age, so we poured ourselves this Red Chair NWPA from Deschutes Brewery.  And although we can't officially condone drinking and operating audio equipment, we've been told this beer pairs well with a relaxed VO read over a bright and upbeat piece of music that reminds us that summer is just around the corner.

Cheers!

"His Master's Voice" And The Dog That Became A Legend

Many of you probably recognize the image above, or have seen a version of it somewhere. Some of you may even know the original artwork was titled: "His Master's Voice." But only a few may actually know the name of the dog - "Nipper."

A painter by the name of Francis Barraud originally depicted Nipper after he saw him listening to a recording of his master's voice. And after being used in a Victrola advertisement (and many other brands and products that followed), Nipper became an icon of the audio world.

Nipper has also become something of an unofficial mascot here at ADS Recording, and over the years our office has collected myriad versions of this audio idol. Everything from printed ads and reproductions, to sculptures and figurines. There are even many hidden Nippers in the pictures throughout our website (can anyone find them all?). We even have a framed illustrated explanation of the history of Nipper (see below). Although is it just us, or does Nipper look a little leaner in this version?

Special Visitor from Los Angeles

Today our regular voice talent Debi Mae West (pictured middle) happened to be in Portland, so she stopped by to record some promos for Nickelodeon and our weekly Fred Meyer session. Debi has voiced just about everything from video games and cartoon shows, to commercials and promos, and has been one of the voices of Fred Meyer for over 15 years.

ADS Recording 20th Anniversary

This past April, ADS Recording hit a major milestone- being in business for 20 years! We are very proud of this accomplishment and are thrilled that we have been able to provide quality audio services to our clients for so long. 

We decided to commemorate this event with a video- interviewing our "founding father" to find out how he got into this business, and what he feels is the best part of his job.

Check it out and let us know what you think!  And here's to another 20+ years of serving both our local, and national community.