How to fix mismatched audio and video frame rates
Choosing different frame rates for your audio recording versus your video recording is a really stupid mistake. I would never be so thoughtless, and you, dear reader, certainly would never do such a thing. But for the sake of argument, let’s pretend that hypothetically I happened to have recorded some audio at 29.97fps for a recent project, but the matching video was at shot at 30fps. With no possibility of a reshoot or overdub, I really needed to get the audio and video frame rates to match. Again, I would never make this mistake, but if I had, this is what I would do to fix my screw up.
The Fix for a Purely Hypothetical Scenario
- Record some audio at the wrong FPS. Way to go!
- Fire up the application Izotope RX. I used version 7 for this example and cannot recommend it enough. This is not an advertisement. I’m simply a fan of this software suite. It has saved and improved countless recordings for me.
- Open the Preferences for RX and select the Misc tab.
- Set the “Time scale frame rate” to your destination frame rate (the frame rate of your video).
- Click OK to close Preferences.
- Open your audio with RX.
- Make any edits you desire.
- Save or Export your audio.
- Import your audio with the corrected frame rate into your video editing software and time align it with your video.
- Wipe your brow and breathe a sigh of relief.
Let me know if this worked for your friend or co-worker, because, again, like me, you would never make this mistake.
I’m posting this article because when I try searching for solutions to this problem the typical results are mostly professionals on forums with their stance: “THE RIGHT WAY TO DO IT IS TO RECORD IT RIGHT THE FIRST TIME, NOOB.” Yeah? Well you know what? No.
No Comments >
Check out this new release!
The debut EP from Focus Fox was released today. My brother-in-law Daniel Nelson is the brain child behind this five track modern folk rock/alt-country gem.
Dan’s songwriting has found a strong footing here. This short album seems to continue on from where Jeff Buckley abruptly left off. His intricate and lush guitar work accompanies his clear and sometimes vibrato-shaken voice. Lyrics are delivered directly, pulling no punches, but never feeling forced.
I had the honor of laying down some BGVs on track 2.
Get your ears on the album via iTunes, Amazon, Google Play, or Spotify. Follow the band on Instagram.
Focus Fox – EP Tracklist
Better Than Me
It Must Be Hard
If You Would Try
We’re Gonna Fall
Way Aback When
No Comments >
GarageBand likes to keep MIDI data hidden and inaccessible. Here’s how to extract it anyway.
Apple’s GarageBand makes it relatively easy to sketch out an audio demo, but it does have some severe, intentionally-crippled limitations.
One of the biggest drawbacks is the lack of built-in support for exporting MIDI data.
Performances are stored inside the session file in some sort of MIDI fashion, but Apple doesn’t give users an easy way to get that information out. Major bummer. *looks west towards Cupertino, squints eyes, shakes fist in air, mutters under breath*
However, a nice guy named Lars Kobbe has put together a workaround/hack that extracts MIDI data from the reluctant clutches of GarageBand. You can download his GB2MIDI Apple droplet script from his site: MIDI-Export in Apples Garageband. Here’s the direct download: GB2MIDI.ZIP If that link doesn’t work, I’m providing the file hosted on my site here: GB2MIDI.ZIP
The article is in German, but instructions in English are found near the bottom of the article (just before the comments section). Getting the MIDI data out involves several steps. Here’s my summary of the process.
How to Extract MIDI Data from GarageBand
- Join (Command-J) regions of a track you want to export
- Convert that region to a loop via Edit > Add to Loop Library (NOTE: In GarageBand 10.1.0 this menu item is now located under File > Add Region to Loop Library )
- Find the newly created loop file (an .AIF with MIDI data hidden inside it) in the folder:
Macintosh HD (or whatever your system drive is named)/Users/(your home folder)/Library/Audio/Apple Loops/User Loops/SingleFiles/
or the abbreviated:
~/Library/Audio/Apple Loops/User Loops/SingleFiles/
- Drop that .AIF file on Lars’ GB2MIDI droplet
- Grab the freshly extracted .MID file, which should appear in the same folder where the .AIF loop was. If not, see the comment section below.
- Import the .MID file into a respectable DAW (basically almost anything other than GarageBand).
- Make next hit record.
That last step is optional, but I say go for it. 😉 Let me know if this helped you.
Locating The Files
If you’re having trouble locating the loop file, it may be because your Library and/or Users folders are hidden, as later OS X versions have been wont to do.
To unhide the Library folder, open the Terminal application, which is found in the /Applications/Utilities/ folder. At the prompt type the following:
chflags nohidden ~/Library/
To unhide the Users folder, type this into Terminal:
sudo chflags nohidden /Users
Then enter your administrator password.
Look for the newly unhidden Users folder in your hard drive’s root folder. It should look something like this:
After running “sudo chflags no hidden /Users” you should see the Users folder (highlighted in red in the image above) appear under the root folder of your hard drive (often named “Macintosh HD” by default).
For more on the hidden Users folder issue check this article from The Mac Observer. It seems the problem was introduced with iTunes 11.2 when Find My Mac is enabled. Another blog suggests that updating to iTunes 11.2.1 fixes the issue.
This GarageBand MIDI article has regularly been one of the most popular posts on my site. That means there are a lot of people using GarageBand and discovering its unfortunate MIDI limitations. The best bit of advice I can give to any musician or audio engineer still using GarageBand is STOP. I know that may sound harsh, but GarageBand is intentionally made to be consumer-grade software. If you’re serious about recording, take the time to investigate other DAWs. Find an alternative solution. There are many to choose from and nearly every one of them is less limited than GarageBand. They range from super affordable to “professionally priced.” Here’s a list to get you started. (Some links are affiliated.)
Pick any of the DAWs above (or find another — this list is by no means exhaustive) and you’ll find it much easier to work with MIDI. Let me know what software you chose.
If you are on OS X 10.15 Catalina or greater on your Mac, then you can only run 64-bit apps. As of the time of this update (May 2020) the app is not 64-bit compatible. This is a known issue. I am not the developer of GB2MIDI, but thankfully the developer Lars Kobbe maintains his app on Github. Here is the link to an open GitHub request for updating GB2MIDI to 64-bit.
112 Comments >
This might be the solution.
An unexpected authorization error 14051 occurred.
ID: ePAY : 14051 / Dngl : 1595
I got this error a few days ago. It’s a new one for me. What caused this? Good question. I have no idea. Pro Tools wouldn’t really start after this.
As usual the Avid forums weren’t very helpful. Which led to this tweet…
For the record, at the time the error occurred I was running OS 10.8.4 and Pro Tools 9.0.6 on a Mac Book Pro with an iLok 2.
I had to force quit Pro Tools. Then I unplugged my iLok 2 and plugged it into a different USB jack. Presto. Working again. Not sure what caused it, nor if switching USB jacks was actually the fix, but I did get it working again after doing so. Hope this helps somebody.
I confirmed again that switching which USB jack the iLok 2 was plugged into made the difference. I would think that this is a problem with that particular USB jack, but all other USB devices work just fine plugged in there. Hmm…
5 Comments >
or… How to release a new version of your product/service and bring an entire industry to a screeching halt.
Shown: original iLok on top, iLok 2 on bottom
PACE has changed how their customers interface with their infamous iLok. The iLok is a DRM dongle, that many software manufacturers use to manage licensing. Formerly, all licenses were managed (mostly just fine) through the ilok.com website, which is now an insufferable “support” site. The new, prematurely launched system PACE requires users to install the iLok License Manager application on their computer.
Ok, no big deal, right?
I recently purchased several plugins to use in my audio production. I’d love to use these great new plugins, but I can’t because the PACE application is horrible.
In order to use the plugins, I need an iLok 2, which has to have the licenses on it, which must be loaded onto the iLok only by using the iLok License Manager, which won’t even allow me to sign in. This is the error I get.
The session you were using is no longer valid. Press OK to establish a new session.
Pressing OK makes the error go away, but it comes right back when I click “Sign In.” The iLok support site doesn’t list this problem as a issue I can submit a support ticket for. So that’s it. I can’t sign in.
Maybe it is just as well. Even if I could sign in, the advice on the the “street” is don’t try to sync your licenses, ’cause you might lose them.
PACE has acknowledged there are issues, but has been otherwise silent.
If this were a football game, PACE fumbled at kickoff, bungled the whole first half, refuse to answer any questions at half time, and amazingly the fumbled ball is still loose in the second half.
I think this screen grab from the iLok.com website says perfectly what many digital audio workers are thinking.
A funny thing happened with some of the content on this page. I can’t tell the story just yet, but I bet it’s going to be a good laugh when it’s all over. Interweb lulz.
As promised…a funny story. After poking around my site stats and hits, I discovered someone was hot linking me.
If you’re not familiar with hot linking, it’s like stealing cable TV from a neighbor, except it hurts the neighbor instead of the cable company. I had a bandwidth leech!
Anyway, a very popular, well-respected pro audio plug-in development company (who will remain unnamed, because it ended well) was using an image from my site on their support page. It was the photograph I took of two iLoks, which is featured at the top of this very blog post.
I knew I could do something funny with the hot link and maybe get a free plug-in out of it. So I created this new image to replace the one they were linking to on my server.
The names of people and plug-ins are blurred out to protect both the guilty and the innocent.
This meant that the above image would now show up on their site. Zing!
I had formatted it to look nearly identical to their artist endorsements in hopes that it might ride under the radar, remaining visible on their support page for as long as possible. For a short while this unofficial endorsement was live on their site.
Long story short…I uploaded the image and went to bed.
Surprisingly, less than 12 hours later I received an email from one of the company’s developers. He basically said, “well played,” thanked me for not goatse-ing them (If you don’t know what that is, don’t Google it.), and let me pick out a free plug-in. Woohoo!
Moral of the story: Hot linking costs everyone something.
Side note: The very same iLok 2 that’s in the picture featured in this debacle must have a desire to make me famous/infamous. It is the very same iLok I photographed to use in the satirical movie poster THE SNOWDEN ULTIMATUM, which was featured in Forbes and lots of other places. There’s something strange about that iLok.
19 Comments >
A little history plus a free download of plug-in settings for Haas Effect panning.
A smart guy named Helmut Haas discovered a bunch of cool things about the way our human brains decode the sounds we hear to determine the direction of where those sounds originate.
Back in 1949, Mr. Haas found that early reflections of sounds help our brains decipher where the sounds came from. We can tell a noise came from the left not simply because we hear it in our left ear, but also because the sound bounces off a wall to our right and hits our right ear a very short time after it hit our left ear. Almost instantaneously, the brain detects the short time between the two signals and tells us, “Hey, that sound you just heard came from your left. Better turn your head to see what it was!” This happens so quickly that we don’t really even think about it. We just “know” it came from the left.
Haas also recognized that early reflections are basically copies of the initial sound that are delayed slightly. He started messing with people’s heads. He pointed speakers at them and firing sounds with very short delay differences. Then he asked the test subjects which direction the sound seemed to come from.
His conclusion: Not only is it fun to play with sounds, but also 40 ms (milliseconds) is some kind of magic point for our brains. If an echo is more than 40 ms after the initial sound, then we hear the sounds as separate instances. But if the delays happen within 40 ms or less of each other, then we perceive them together as merely directionality cues of a single sound.
For example, if a sound hits our right ear and the same sound hits our left ear 0.3 ms later, we don’t hear two sounds, we only hear one sound coming from approximately our 1 o’clock position.
And so the Haas effect was named after him.
Engineers have implemented the Haas effect as an alternative to panning. Most of the time panning works just fine, but it does have limits.
Sometimes panning leaves the location of the audio feeling indeterminate, smeared, mono, or one dimensional. This is why a lot of engineers skip the pan knob altogether and mix LCR.
To effectively localize a track in a stereo field using the Haas effect, engineers have to do a couple things. They duplicate the track, pan the two tracks hard left and right, and then apply a delay to only one of the sides. The delay is applied to the side opposite of the side from which the sound is intended to perceived as originating.
Typical delay times for this technique are increments of 0.1 ms from 0.1 to 0.7 ms. This yields linear movement across the stereo field. You can think of it like this chart shows.
Example: Want the sound to come from 9 o’clock on the left? Delay the right side by about 0.4 or 0.5 ms.
After researching the Haas Effect, I decided I wanted to try it out in a mix. Since the settings must be very exact, setting it up correctly can be a bit confusing. Presets to the rescue!
I made these presets for the stock Digidesign Mod Delay II plug-in. These presets only work for this specific plug-in and Pro Tools. If there’s interest, maybe I’ll make more presets for other DAWs in the future.
Download this ZIP file, unzip it, and drop the folder and included presets in the Mod Delay II folder in the Plug-in Settings folder. On a Mac it’s probably located at Library / Application Support / Digidesign / Plug-In Settings / Mod Delay II, but may be in a different location on your system.
Setting up the tracks
Insert an instance of the Mod Delay II (mono/stereo) plug-in on the mono track you want to Haas-ify. Select the preset you want. No need to duplicate tracks. Bingo.
Haas Effect Panning
Understanding how to use the Haas effect properly means you need to understand and pay attention to things like stereo-to-mono compatibility and comb filtering, as well as other stereo field mixing techniques. As with all effects, have fun but be careful not to over do it. Experiment and do your homework. Then let me know if you find learn or discover anything cool.
3 Comments >
Here’s a cool video that got me thinking about the Haas effect. This video no longer available.
Here is how I fix this Pro Tools error.
Really? A typo in the error? Grrrrr…
Ever get this error? Can’t open your session, right? Not only is it a major workflow stopper, but the double punctuation typo at the end is annoying as well.
Luckily, the solution is quite simple.
This is the quick fix that works for me and my particular setup of hardware/software. Your mileage may vary.
- Quit Pro Tools
- Restart Pro Tools
- Open the session that wouldn’t open before
- Get back to work
The IT mantra “Have you tried turning if off and on again?” waves the problem away like a magic wand, but why is this problem happening in the first place?
The last time this error occurred for me, I noticed that it was after I had ejected my audio hard drive, removed my iLok, and left Pro Tools open, but put my machine to sleep before Pro Tools could issue the panic message: “Hey! Where’s your iLok, buddy?! That’s it! We’re shutting this whole thing down.” Then when I went to reopen the last session I was working on, boom, the error in question occured.
I’m guessing that between the time I ejected everything and the time I plugged it all back in and tried to fire it up again, Pro Tools had switched its default sample rate from whatever my Mbox 2 Pro says it was to whatever my MacBook Pro thinks it should be. Then when I try to open a session with a particular sample rate that doesn’t jive with what the current rate is, Pro Tools freaks out because it thought it knew what was right, but doesn’t even know anymore, man.
Disclaimer: I don’t actually know how or why the error is occurring. These are just my slightly educated stabs in the dark. If you know anything more about this error, why it happens, and, most importantly, why there’s a typo in it, please leave your thoughts in the comments section below.
No Comments >
Is every new technological development just a deeper dream state?
Sound is basically waves of pressure changes. The exact definition is more complicated, but essentially we perceive sound because our ears decode the frequencies of oscillating movement of particles in gases, liquids, and solids. There are many ways to generate sound waves, such as plucking guitar strings so they vibrate, or hitting a membrane like a drum head.
A long time ago, people discovered that sound could also be made by blowing air through a pipe with a opening on the side, thus inventing the whistle. They also found that a range of tones could be produced by assembling a group of whistles with varying lengths and diameters. Then they attached a controller (called a keyboard or manual) so that one person could “play” this collection of pipes. Their invention is what we now know as the pipe organ.
At the start, pipe organs had only one timbre â€“ a basic whistle sound, but over the next several hundred years, smart inventors and musicians made improvements in the technology. They found ways to emulate lots of other instruments, like brass, woodwinds, percussion, and even human voices. Their hope was to fully replicate those real life instruments.
As features were added, pipe organs evolved into enormous, elaborate, and expensive installations, increasingly more complicated to play and maintain. While these pipe organs were truly amazing inventions, capable of creating complex and beautiful music, they were actually quite poor emulations of the real life instruments they were intended to replace.
Still, we humans are adaptable and we fell in love with the sound of pipe organs, learning to appreciate the instrument for what it was, not what it wasn’t.
Eventually, we discovered electricity and began to harness its power to create electromechanical instruments. Creative minds developed things like vacuum tubes, tone wheels, and transistors. Companies like Hammond and Wurlitzer implemented tone wheels to generate sounds approximating a pipe organ.
However, similar to the pipe organ, this new technology was a brilliant invention that poorly emulated its predecessor. These new organs were affordable alternatives to pipe organs, so in spite of being a bad imitation they became popular with smaller houses of worship. Traveling musicians took advantage of the portability of these smaller organs too, making their sound common in popular jazz, blues, and rock music.
Once again, our ears grew accustomed to the sound of the imitation, developing an affinity for the quirks of its particular aesthetic.
As the march of progress continued, electronics became smaller and more powerful. Engineers found ways to replace the delicate mechanical parts in electric organs, which were subject to wear and tear, with completely electronic sound generators. Lightweight, all electronic keyboard synthesizers used a variety of methods in attempts to replicate the sounds of their heavier electromechanical ancestors.
But just like before, history would repeat itself. The new emulators were incredible technological achievements that fell short of their goal of replacing the old technology. Though they lacked the ability to fully replicate the previous generation, they possessed attributes that eventually found an audience of connoisseurs that valued them not just in spite of their glitches, but because of their unique properties.
Today, we synthesize the sounds of the old technologies with computers and keyboard MIDI controllers. While initially computers could only crudely imitate the old masters, DSP technology is progressing rapidly. CPU speed and available RAM are no longer the main limitation factors. As the computational power ceiling continues to rise higher and higher, software programmers are able to provide increasingly nuanced emulators that can easily fool the listener into believing that the software is actually the real thing.
At this point, if you’re still reading, then you probably can see how this history correlates to the plot of the film Inception. Each new technological breakthrough has been like a deeper dream state, where the simulation moves further and further away from reality.
→ Pipe organs
→ → Electric organs
→ → → Keyboards
→ → → → Software
However, just like in the film, while each level becomes more strange and abstract, the deepest level — Limbo — actually approaches something most like the real thing or maybe even better. Today’s emulators delve into such detail and are able to control even the most minute aspects of the sound, that it won’t be long before they easily eclipse the believability of the old technology. In fact, we may already be there.
A few years ago (when the emulators weren’t half as good as they are now), a friend of mine (who has very good ears) dropped by the studio to hear a song I was working on. When the B3 organ kicked in during the chorus, he declared, “That organ sounds great. There’s nothing like the real thing!” Muwhahaha! The smoke and mirrors of software emulation had worked.
Inspiration for This Article
This idea of how keyboard technology relates to Inception came about through a discussion with my friend Hoss. Over the weekend we were working on the keyboard parts for our band Rudisill’s next album Take To Flight. In between takes of an organ part we marveled at the realization that the software he was using was an emulation of an emulation of an emulation — a truly strange scenario.
Follow Rudisill to hear about the new album when it is released later.
No Comments >
A response to the question â€œAfter Analog vs. Digital, what will we fight about in the future?â€
As part of their “#DJChat,” German audio equipment manufacturer Behringer asked this question on Twitter:
…Analog vs. Digital is a debate that will always continue. But in the future, what technology will we move on to AFTER digital? 😀 #DJChat
It’s an interesting concept. The wars between analog and digital rage on because they are systems separated by technologies that both have pros and cons. As technology progresses, what new pros and cons will we have to debate against older systems? Initially I answered with the following:
@BEHRINGER future: Cerebral vs. Digital. Was it made entirely “in the box (aka your head)” or did you collab with other humans and devices?
Realizing there’s much more to this debate than just a tweet, I thought I’d talk more about it here.
We Need Better Words to Describe How We’ll Make Music in the Future
In my original tweet, I used the phrase “Cerebral vs. Digital” to describe the future debate I imagine will happen. Maybe my choice of opposites wasn’t perfect. Better words can probably be found. This concept of diametrics I have in mind could be expressed in a variety of ways.
- Cerebral vs. Physical
- Solitary vs. Collaborative
- Internal vs. External
Each of those word combinations is describing the same contrast of ideas. But how to best describe it?
The New System of Mind Music
In the (maybe not so distant) future, musicians will have the ability to directly output music from their heads. Technology will be developed that will allow artists to simply think/imagine/hear the music in his/her head and output this as audio and/or notation. This cerebrally generated “audio feed” could be routed (maybe even wirelessly) to a recording device to be documented, distributed, and sold. Theoretically, this process could happen as a live performance. The signal could be routed to a sound system for a concert, to an internet connection for worldwide streaming, or even directly injected (almost telepathically) into the head of a “listener” outfitted with the proper “receiver” device.
The possibilities are fantastic. Composers could direct an entire imaginary orchestra as they hear it in their minds. Dancers could dance to their own music in real time. Musicians could play exactly what they intend to play. Singers could sing in whatever voices they can imagine. Handicapped artists suddenly would be unrestricted by their handicaps.
This is not a matter of if, it’s a matter of when. If we already can control toy helicopters with our thoughts, then it’s only a matter of time before we can output music directly from our minds. UPDATE (2011-09-23): This just in… UC Berkeley neuroscientist Professor Jack Gallant announced today that it’s possible to recreate the video from brain activity.
This technological breakthrough in music will follow a path familiar to video games. With the Wii, Nintendo brought wireless motion-sensing accelerometer action to everyday people. The developers of Guitar Hero and Rock Band banked a lot of cash by making it really easy to “play” popular music without having to learn an instrument. Microsoft’s Kinect for Xbox removed the need for a controller, allowing the person to become the controller. I don’t know who will create the first mind-controlled music technology, but somebody’s going to do it.
Cool meant something totally different back then. Don’t judge.
As with any change, it’s going to get worse before it gets better. Unfortunately, music will experience yet another Regrettable Period in which we have to learn how to use this new technology properly. I predict some gross and unsavory abuse of the technology, much like the ubiquity of terrible synthesizers in the 1980s or prevalence of auto-tuned vocals since Cher started believing in life after love. But some lucky artist is going to enjoy the honor of being known as the one that mastered this wonderful new system, thus becoming the “Grand Master Flash of whatever-this-thing-may-become-known-as.” Someone will figure out how to use it right, but it might take some time. In the meantime, wear earplugs.
Why We’ll Argue About This
At first, this newfangled gadgetry will be heralded as the end of “real” music and musicianship. The critics will say it’s too easy and not authentic music. Traditional composers and invested players will complain that no one has to learn how to write or play anymore. And much in the same way that digital was derided as a poor substitute for analog, purists will say that this cerebral form loses something in the process. Those arguments all might be right, but there may be a bigger issue lurking.
Trapped “In The Box”
When the process of making music becomes entirely internalized it will be really great because of it’s purity and singularity of thought, but will it simultaneously suffer from lack of external influences? When digital recording became popular, the question was often asked by one artist or engineer to another: “Was this all done â€˜in the box?’” â€“ meaning: was the audio signal created, mixed, and mastered on the same computer? Early on, music created entirely in this fashion lacked the beneficial effects that analog systems inherently imparted upon the audio signal. Today, the line has been blurred by better technology, so it’s harder to tell if something was recorded analog or digital. Only engineers with “golden ears” can hear the difference (even then I suspect shenanigans). At any rate, the question still remains: What benefits will be lost due to the signal remaining “in the box” of your head?
Potential Musical Influences
- People â€“ The comradery, inspiration, ideas, criticism, differing views, and friction found when people work together often makes for better music. Being alone can lead to dead ends and boring or bad music. Collaboration can make beautiful things.
- Hardware â€“ Though they are inanimate objects, the instruments and devices used to make music come with their own inspirations, challenges, rewards, frustrations to overcome, and occasional good glitches. Sometimes a piece of gear has to be conquered and relinquishes its magic upon defeat.
- Criticism â€“ The critic is the archenemy of the artist, but every good story needs a villain. Without judgement, no work is ever as best as it can be. Words are often revealed for their folly only after they’ve left the head.
- Movement â€“ Music and movement are very strongly related. When making music, movement is both part of the instigation of sound, but also a reaction to the sound being created. Performance and dance are like cousins. So if movement is not necessary for the creation of music, what effect will that have on the final product?
Good Things Will Happen
A lot of things can go wrong in this new system, but a lot of things can go right too. Eventually we’ll work out the kinks. We’ll figure out the typical pitfalls. We’ll master this medium like we have with all the others. One day amazing music will be generated using nothing but musicians’ brains. I’m hedging a bet it will be the direct output of some ridiculously young Mozart’s mind that will blow us all away. Perhaps this new interface will teach us something about how our brains work. Maybe it will allow us to communicate more precisely on ever deeper levels. What if it develops into a new universal language? Hmm.
2 Comments >
How to get a Pro Tools rig up and running when the error message â€œThe audio device buffer underflowed…â€ wonâ€™t go away.
The Error Message
The audio device buffer underflowed. If this occurs frequently, try decreasing the “H/W Buffer Size” in the Playback Engine panel or remove other devices from the audio firewire bus. (-6085)
Occasionally this error pops up in Pro Tools, usually after I return from a meal in the middle of a long recording or mixing session. The session file will only playback audio for 1 second or less and then the error message pops up. Apparently, Pro Tools 9 is a workaholic and doesn’t like taking lunch breaks, at least when running on the particular combination of MacBook Pro, Mbox 2 Pro, and Western Digital hard drive that I’m using.
Following the directions to decrease the “H/W Buffer Size” in the Playback Engine panel doesn’t seem to help. In fact, not only does decreasing the buffer size seems contrary to the suggested way to solve a buffer underrun, but it then sometimes throws this error message:
A CPU overload occured. If this happens often, try increasing the “H/W Buffer Size” in the Playback Engine Dialog, or removing some plug-ins. (-6101)
I’ve tried a lot of things and the problem seems to be related to the hard drive and firewire ports. Here’s how I fix it.
- Save and Close the session.
- Quit Pro Tools.
- Eject the hard drive used for recording audio.
- Unplug the audio hard drive and Mbox 2 Pro (or the audio interface you’re using).
- Wait 10 seconds.
- Reconnect the audio hard drive and audio interface.
- Restart Pro Tools.
- Reopen the session and press Play.
If the session plays back without stopping, then it worked. If not, then I don’t know what to tell you, which reminds me of a “Deep Thought” by Jack Handey.
If you ever crawl inside an old hollow log and go to sleep, and while you’re in there some guys come and seal up both ends and then put it on a truck and take it to another city, boy, I don’t know what to tell you.
Hopefully this solution worked for you. Let me know if you’ve had the same problem, what hardware you are running and if this solved the problem.
5 Comments >
Pro Tools hardware is either not installed or used by another program. If you thought that having Pro Tools 9 installed meant no more “Hey, Mr. Engineer Genius, where’s your fancy hardware?” errors, then this nagging error probably came as a surprise. It did for me. Since installing Pro Tools 9, my workflow has allowed […]
Pro Tools hardware is either not installed or used by another program.
If you thought that having Pro Tools 9 installed meant no more “Hey, Mr. Engineer Genius, where’s your fancy hardware?” errors, then this nagging error probably came as a surprise. It did for me. Since installing Pro Tools 9, my workflow has allowed me to jump around from my Mbox 2 Pro, Mbox 2 Micro, and MacBook Pro’s built-in sound card. This has been really handy while trying to finish up my album on the road. But, apparently, all that hardware hopping can cause the playback engine to get stuck in some funky states that don’t so work â€“if at all. See my previous post “FIX: Pro Tools could not set sample rate to specified value” for a similar issue.
Obviously, the problem has something to do with the playback engine. Since the error dialog only offers an â€˜OK’ button, which closes Pro Tools, there doesn’t seem to be a way to work around the problem. There is not even a way to know what hardware Pro Tools is expecting.
I found a simple solution via this Sweetwater forum. The answer given there details how to get Pro Tools running on a PC, but I found that it worked for Macs too and without having to install any drivers. The fix is kind of like booting Pro Tools in safe mode. Simply hold the â€˜N’ key while starting up Pro Tools. This will bypass the normal start up sequence and open up the Playback Engine window. Now you can select the correct playback engine and continue using Pro Tools.
In my situation, Pro Tools was looking for the last connected device (my Mbox 2 Pro), but since it wasn’t available it opted for the next available option: my MacBook Pro’s line input, which doesn’t make a very good playback engine.
Let me know if this fix worked for you.
10 Comments >
This problem may have been fixed in the Pro Tools 9.0.2 update that came out yesterday, though I’ve not been able look through the 9.0.2 Readme file in detail or to test this out on the updated software. I’ll update this page when I find out more.
After upgrading to the newly released Pro Tools 9, I couldn’t open sessions or create new ones. I got this error: “Could not complete the Open Session… command because Pro Tools could not set sample rate to specified value..” I hunted around on the web and various forums, but couldn’t find a solution that fit. […]
After upgrading to the newly released Pro Tools 9, I couldn’t open sessions or create new ones. I got this error: “Could not complete the Open Session… command because Pro Tools could not set sample rate to specified value..” I hunted around on the web and various forums, but couldn’t find a solution that fit. I found several items relating to Windows and Pro Tools 8, but nothing for a Mac running Pro Tools 9. After messing around a bit I figured out the problem was with my playback engine. Here’s how I solved it. Let me know if it works for you too.
Open the Playback Engine dialog under the Setup menu item.
From the menu bar select Setup > Playback Engine… to open the Playback Engine dialog window.
The problem is with the Pro Tools Aggregate I/O.
By default, my current engine was set to “Pro Tools Aggregate I/O.” It’s odd that this Pro Tools would leave it that way after an installation since AVID states that it is not supported in OS X.
Select your current playback engine.
The fix is easy. Simply select the right playback engine. Your options may differ based on your setup.
In my case, I usually would edit with my Mbox 2 Micro, but since Pro Tools 9 gives us so many more options for hardware compatibility, I selected Built-in Output. I was able to edit some vocal takes using my Macbook Pro’s speakers instead of pulling out my headphones. Nice!
32 Comments >