Support this Site
Become a Six Colors member and get access to an exclusive weekly podcast, community, newsletter and more.
By Jason Snell
November 11, 2019 5:22 PM PT
When the new Apple Pencil came out a year ago, I integrated it into my iPad editing workflow. I can edit podcasts with the Apple Pencil at a pretty impressive rate of speed, and the precision of the Pencil means that I’m more inclined to make detailed edits on the iPad than I am when I’m editing on my Mac with Logic Pro X and a trackpad. In fact, every episode of The Incomparable that I’ve edited in the past four months has been done on my iPad Pro.
I started editing podcasts on the iPad when I was traveling, since I haven’t regularly traveled with a Mac laptop in a few years now. But this summer I decided I’d rather edit The Incomparable, which I tend to do on Saturday mornings, somewhere other than at the same desk I use during the week. It’s nice to be in the same space as the rest of my family, even if we’re all doing our own thing and I’ve got headphones in while I edit the podcast.
Ferrite Recording Studio is a fantastic app that does almost everything I’d want an editing app to do, and combined with the power of the iPad Pro I can even edit podcasts with enormous panels, like our Incomparable draft episodes—though I had to rotate my iPad to fit all the tracks on screen.
This year the iPad has become much more capable at being a podcasting device than ever before. iPadOS 13 and the updated Files app give me access to audio files on USB media, which was a major hurdle before. A new update to Ferrite added support for recording on up to 8 tracks simultaneously, so I could record a multi-person session directly into my iPad if I wanted to. (In general I use a Zoom recorder for this, though—I will trust dedicated recording hardware over computer software every time.)
That leaves a couple of places where the iPad still lacks, though.
First: Recording multiple people via the Internet. On my Mac I use Audio Hijack to record my own voice as well as the audio from all the other people on a session, but you can’t run two audio apps at once in iOS. I’ve taken to recording many podcasts using Zoom Cloud Meetings, which will theoretically record the audio from participants on iOS as well as it does from those on desktop operating systems. I’ve also used RINGR, a cross-platform conferencing app, with results of varying quality. And I figured out a way to record my own audio locally onto an external recorder, so that’s an option.
But the truth is, I just record my podcasts on my Mac most of the time. On the road, I have iPad-only alternatives, but they offer enough trade-offs that I wouldn’t use them if I have a Mac handy.
Now here’s the tough one, one I don’t have a good answer for as yet. As cool as it is that I edit every episode of The Incomparable on my iPad, the fact is that all the audio files for that episode are prepped on my Mac before they get to my iPad. I sync audio tracks using a proprietary tool, then use iZotope RX to remove background noise, and finally use a compressor (currently it’s Klevgrand’s Korvpressor, but it’s the latest in a string of ones I’ve used, they’re like Spın̈al Tap drummers) to balance the volume of audio across different tracks.
Ferrite includes a compressor plug-in and a volume-leveling preprocessing feature, neither of which can I get to generate the output I desire. Korvpressor has an iOS version that I can use as a plug-in in Ferrite as I do on the Mac with Logic Pro X, but the iPad version crashes reliably, so I can’t use it. And there’s absolutely nothing I’ve found on iOS that can match the quality of noise and echo removal that iZotope provides on the desktop.
Now, I have definitely posted podcasts that were entirely processed on my iPad. I’ve made use of existing tools to make the audio sound as good as I could. And yet, when I listen back to those podcasts, I can tell that they don’t sound as good as the ones processed on my Mac.
Maybe iZotope will bring a subset of their audio tools to iOS at some point. Maybe I’ll find a plug-in that’s more stable inside Ferrite, or maybe I’ll figure out a way to use Ferrite’s leveling features more effectively. But for now, my iPad editing workflow still passes through my Mac.
Things are a lot better than they were even a year ago, but we’re not all the way there yet.
By Jason Snell
February 19, 2019 2:41 PM PT
Last week I took a trip during which I needed to record three podcasts (Liftoff, Download, Six Colors Subscriber Podcast) with guests who would be participating via Skype. I almost took my trusty old MacBook Air with me, but I decided to see if I could figure out a way to replicate the bulk of my home recording setup without requiring a Mac.
In the past, I’ve done something similar using the Audio-Technica ATR2100-USB, a microphone that can output a digital signal using USB and an analog signal via an XLR cord simultaneously. The problem is that the last time I tried to use the ATR2100-USB with my iPad Pro, it didn’t return my own voice into my ears, making me unable to judge the sound quality of my own microphone. After years of having my own voice return to me, I strongly prefer not to record unable to hear my own voice. (I use in-ear headphones that largely shut out audio from the outside world, so the experience of speaking while not hearing yourself is even more profoundly weird than it would be with leaky earbuds.)
This time I wanted it all, or at least as close to all as I’m able to get with iOS in the mix: A pristine recording of my own voice, that same high-quality microphone audio also flowing across digitally to my podcast guests via Skype, and the ability to hear both my guests and myself at the same time.
I made it work with the addition of one box to my usual iPad workflow. Here’s what I did:
First, I plugged an analog XLR microphone into my Zoom H6 recorder. That solves the “get a pristine recording of my own voice” problem. But how to get that audio out of my Zoom recorder and into my iPad Pro? If I plug my headphones into the Zoom, I’ll be able to hear myself but not my guests. If I attach the Zoom to the iPad, I can relay my audio—but the Zoom is unable to record audio when it’s being used as a USB audio interface.
Second, I need to route my microphone audio out of the Zoom to a device capable of transferring it to my iPad Pro (and also transferring the voices of my panelists from the iPad back to me). Any standard USB audio interface should be more or less capable of that, and so I used mine—the Sound Devices USBPre2. The trick was how to connect the Zoom to the USBPre2. Fortunately, the zoom has a Line Out port on its front, and the USBPre2 has a line-in port on its side, and I happened to have the right cable (minijack on one side, stereo RCA on the other) to connect the two of them in my random drawer of audio cables.
Third, I attach my USB audio interface to my iPad Pro. (I used a USB-B to USB-C audio cable for this, but an old-school cable will also work with an adapter.) I haven’t yet met a USB device that my iPad Pro is incapable of powering by itself, so the USBPre2 worked just fine. I also attached my headphones to the USBPre2, so I could hear myself and my guests.
That’s it! I could launch Skype, press record on the Zoom, and record a podcast. My guests heard my high-quality microphone audio, I could hear them, and I could hear myself (with no noticeable latency). The only thing I’m really missing is the ability to record my guests’ audio too, as a backup, but I chose to live dangerously and speak only to people who know what they’re doing when it comes to recording for a podcast.
The final step was one that I’ve described before, namely using an external Wi-Fi box to transfer my audio files back to my iPad for editing. This workaround remains until the day where Apple decides to let iPads see external storage devices directly. Then it was off to Ferrite to put the podcasts together after the participants sent me their files and I imported them into Ferrite. (As an added bonus, in a recent update, Ferrite has gained the ability to split multi-track QuickTime audio files into their component tracks. Ecamm’s Call Recorder for Skype uses this approach and until Ferrite was updated, I’d have to use a Mac to split those audio files in two. No longer.)
And that’s it! It’s not pretty, it’s two more boxes than I’d otherwise bring, and I refuse to weigh the difference in boxes and compare it to the weight of my 11-inch Air. The important thing is that I was able to travel with my iPad and no Mac and have more or less the same podcast experience that I have when I’m sitting at home at my iMac.
By Jason Snell
May 7, 2018 5:08 PM PT
Back before all the MP3 patents expired, my favorite iOS podcast editing app, Ferrite Recording Studio, couldn’t export MP3 files. Instead, I tried various alternative methods, including using the Auphonic service and various other iOS apps that didn’t seem to care so much about potential outstanding patents.
The good news is, the patents lapsed and Ferrite now supports MP3 exporting. Not only can you set it to export at various MP3 quality levels—bit rate, stereo or mono, and CBR (most compatible) or VBR, but you can enter MP3 tags and show art, and even optionally embed chapter markers with links and custom art. There’s even an automatic volume adjustment feature that will level the volume of your file so that everyone sounds like they’re speaking at the same volume.
Here! Let me show you a video.
If you want a great, low-cost, full-featured editing app for podcasts, I can’t recommend Ferrite Recording Studio enough.
By Jason Snell
December 18, 2017 4:29 PM PT
For more than a year I’ve been trying to give myself the maximum amount of travel flexibility by finding ways to record and edit podcasts on iOS, so I don’t need to bring a laptop with me just to make podcasts. Ferrite has solved my editing needs, and I’ve found a few ways to record audio locally while using iOS.
The big challenge has been iOS’s sad and continued lack of support for external storage devices. When I’m traveling with only my iPhone and iPad, I can record audio on an external device—an SD-card recorder from Zoom, usually—but how do I get those files onto my iOS device? iOS can’t see the contents of a standard SD card.
A year ago I extolled the virtues of using a Wi-Fi enabled SD card to transfer files. And while that works, the problem is that the kind of Wi-Fi that’s embeddable in a tiny SD card is slow. Painfully slow. Especially when transferring large audio files.
This year, though, I found a new device that solved my problems. It’s the Kingston MobileLite G3, a peculiar little multi-tool of a product that can charge iOS devices, act as a mobile router to convert hotel Ethernet into Wi-Fi, and more. But there’s only one feature that I really use: its onboard SD card slot.
Like the Wi-Fi-enabled SD card I previously used, you have to download a custom app in order to view the contents of the SD card and transfer it over to your iPad or iPhone. The difference is speed. The MobileLite’s Wi-Fi transfer speeds are vastly better than those from the tiny SD card.
It’s still a little bit silly that, now that iOS has a file-management app, you still can’t plug in a mass storage device via a USB adapter and copy files off of it directly. But until Apple relents—or if it never does—the MobileLite G3 gives me a fast way to transfer audio files on the road.
By Jason Snell
December 4, 2017 4:35 PM PT
Marco Arment’s Forecast is a newly released (into a public beta) Mac MP3 encoding and tagging tool for podcasters. It’s a tool that Marco built a couple of years ago to serve his own needs, and for the last 18 months or so I’ve been using it (in a private beta) to encode most of the podcasts that I create. Here’s an overview of how Forecast works and what it does.
Forecast takes input files—generally uncompressed audio exported from an audio editing app in WAV format, though it can open other file formats—and outputs MP3 files for use in a podcast feed. This is nothing remotely new. What makes Forecast interesting is the details of how it encodes and tags those files.
First off, the encoding process itself: Forecast is extremely fast at encoding MP3 files for a few different reasons. At its heart, it’s using the common (and excellent) LAME MP3 encoder, but Forecast spreads the encoding job across all of your Mac’s processor cores. The result is that files encode much, much faster (in 29 percent of the time as standard LAME, in my tests, and 80 percent of the time of the iTunes encoder)—and your Mac’s fans will probably spin up briefly, because Forecast is pushing your processor to use all its power to do the job.
There’s also a perceptual trick that Forecast uses to make encoding seem quick: When you add a file to be encoded, encoding begins immediately in the background. By the time you edit your file’s metadata, the encode may have already completed in the background. The first time I used Forecast, I thought something had gone wrong—because when I typed Command-S to save the file, it just saved. There was no wait. The file had already encoded—it was waiting for me, the slow human, to finish typing in episode titles and show descriptions.
All the rest of Forecast is about tagging files to include things like the episode title, show art, and chapter data. Just about any MP3 app can add description tags, but only a handful support MP3 chapters. (Some others are Rogue Amoeba’s Fission and Thomas Pritchard’s Podcast Chapters.)
It turns out that the WAV file format includes support for markers—specific designations of events that happen at particular time codes—and that most audio editors (including Logic and Audition) that export as WAV files will export any markers found in that particular project. This means that in order to add chapters to my podcasts, I don’t need to add a step where I laboriously write down time code for all the events in the episode and then input them one by one into Forecast.
Instead, I just click the Plus icon next to the Marker label in Logic and add a marker. When I export that project to a WAV file and import it into Forecast, the app automatically reads the markers and converts them into chapters. I don’t need to do anything.
That said, Forecast also does support the manual entry of chapter times and the editing of chapter data, including title, URL, and custom per-chapter images. (Manually entering times is a little bit buggy—frequently I need to do it twice before it displays properly. I don’t do this a lot, but it’s an annoying bug I hope Marco will fix.)
There’s also a checkbox that allows for the creation of invisible chapters that don’t display in the episode’s chapter list, but do change the displayed art or link at a particular time. There’s a lot here, depending on how much work you want to do to add a rich media layer on top of your podcast.
Forecast also tries to save you time by recognizing that similarly named source files are probably part of the same podcast, and attempting to intelligently autofill data based on that assumption. When I add a file called
theincomparable382.wav to Forecast, it realizes that this is almost certainly episode 382 of The Incomparable and automatically enters The Incomparable in the Podcast Title field, adds 382: to the Episode Title field, adds the right image to the show art, and even sets the proper MP3 output format—and all because it knows what I did when I encoded
theincomparable381.wav last week. (This autofilling extends to URLs and art in chapters, too. If I have a regular sponsor for a podcast, Forecast is smart enough to remember the URL attached to those sponsorship chapters.)
For editors of sponsored podcasts, Forecast can detect your sponsorship chapters and export those out as separate files, ready to be sent to your ad network or sponsors as “airchecks”—i.e., proof that the ad spots aired as promised. There are also quick-copy features that let you quickly put the show’s duration or file size on the clipboard—apparently this is something Marco needs for one particular podcast host, though I’ve never needed to use those features myself. There’s also a feature that warns you if your audio file contains long amounts of silence—a sign that perhaps something is wrong with your podcast, so you might want to check it before posting.
If you’re a podcaster, you should give Forecast a try. It’s free, and a whole bunch of podcasters have been using it enthusiastically for more than a year, so it’s battle tested. I recommend it highly.
By Jason Snell
April 21, 2017 4:04 PM PT
Every now and then when I complain about Skype, which most of my podcast peers and I use for our conversations, someone suggests an alternative voice-over-IP service and asks why we don’t switch.
The truth is, Skype’s terribleness may be overstated—people get cranky when they’re entirely dependent on a single product and that product isn’t reliable—and the product has gotten better recently after a few particularly rocky months.
But it’s not just about abandoning Skype. Yes, there are numerous services that will let multiple people connect over the Internet and have a voice conversation. 1 Yes, we could move to Google Hangouts or some other web-based business conferencing tool or video game chat app 2.
But here’s the thing: Everybody I know uses Skype. If I’m going to start the painful process of moving house—of getting everyone I’m on a podcast with to, over the course of many months, upgrade their software and get used to a new way of working—I want to move to something that is vastly superior to what we’re currently using. There is no point in dealing with transition costs—inevitably including many lost minutes as everyone waits for someone to install unfamiliar software and figure out how to use it—to make a lateral move. 3
Leaving aside the fact that I have no real faith that alternative option X is actually better than Skype—one person’s “I’ve never had any problems” can be another person’s “omigod it was a disaster”—I’ve decided that I’m leaving Skype only if I’m forced to or if I can find a tool that solves other problems specific to podcasters.
Right now, the biggest issues I have with Skype, beyond the occasional bout of unreliability, are related to recording audio. This isn’t Skype’s fault—it wasn’t built with recording podcasts in mind!—but it’s a necessity for podcasters. While I’m doing a podcast, I need to record my own microphone and, ideally, the rest of the conversation—and in separate files or on separate tracks. And all of my panelists need to record their own microphones, locally, at full quality. (You can read more about this in my “How I Podcast: Recording” article.)
On the Mac, this is pretty easy. I bought a bunch of copies of Ecamm Call Recorder for Skype, which is a plug-in that integrates recording right into Skype, and distributed them to my most frequent panelists. For people who don’t have Call Recorder, QuickTime can record audio fairly easily. On Windows, it’s more complicated—the podcast guest guide that I use recommends downloading the free audio app Audacity. More complexity means there are more chances to do something wrong. (This leads me to an additional feature I require: The software involved needs to be dead simple for a novice guest to set up correctly.)
And then there’s iOS, where this is just impossible. You can’t record your microphone locally while talking on Skype. This severely limits iOS podcasting.
Plus there are some things that Skype does really well, that any replacement needs to do a decent job at. Skype massages audio before it reaches you, leveling and boosting audio and removing background noise and echoes. Its servers merge audio streams together so that multi-person conversations can happen even on on low-bandwidth connections. Skype may have its issues, but it’s also got a lot of strengths that I didn’t appreciate until I began investigating alternatives.
So if I’m going to move from Skype, I need to move to something that won’t be dramatically worse than Skype in terms of stability and audio quality, and it needs to make it easier to record podcast audio across all major platforms, desktop and mobile.
This is a big ask. And it turns out, there’s basically no solution today. But there is hope.
The closest we’ve come are two web services, Cast and Zencastr. Both of these services rely on WebRTC, a browser-based set of real-time communication protocols that let browsers transfer audio and video without special plug-ins. Both services automatically record the local audio of participants and upload them to a remote server, so panelists don’t need to install or run any special software to have their high-quality audio captured for later use.
Cast costs $10/month for its basic plan. I’ve used it for several months in the recording of the TV Talk Machine podcast, and have found it to be quite reliable. It can’t handle conferences with more than four participants, including the host, which disqualifies it from my large panels on The Incomparable, but most podcasts don’t have panels with five or six people in them. (And an expansion of that limit is forthcoming.)
Zencastr has a basic free tier, but to record with more than one guest it’s $20/month. Zencastr claims it can handle “unlimited” guests, though I haven’t tested this and suspect it will bog down quickly if you have a large panel. I’ve used it a few times and found it a little less reliable than Cast—I’ve seen files cut off a few seconds too early, and the quality of the live audio connection had more artifacts than I’ve seen with Cast.
I appreciate Zencastr’s cloud-storage integration: all source files are automatically deposited in my Dropbox after a session is over. In contrast, Cast makes me wait for several minutes before I can download my files.
If you’re recording a podcast with three or four participants, Cast’s $10/month plan is a pretty good deal. If it’s just a one-on-one chat, Zencastr’s free tier is even better. For more than four participants, though, you’re back to Zencastr and you’ll pay $20/month for the privilege. Still, there’s a lot to be said for automatically recording panelist audio without any intervention.
…but then there’s mobile. The fact is, Safari doesn’t support WebRTC right now, so you can’t use either Cast or Zencastr on an iPad or iPhone. It looks like WebKit will support WebRTC at some point in the near future, but we might not see support in iOS until 2018.
In looking for a solution that would work on my iPhone or iPad, I discovered Ringr, which offers built-in microphone recording and supports WebRTC on the desktop and offers iOS and Android apps. Unfortunately, Ringr only supports one-on-one calls, so while it would work great for two-person podcasts, that’s all it supports. A recent Ringr email to customers suggests multi-user conferences are forthcoming.
For the record, business-conference-call apps with desktop and mobile versions don’t support recording of local microphone tracks. Some of them will record the entire conference call on the server, which is cool, but that’s only good for reference—for the best podcast audio, you want to record the microphone at the source.
So the end result of all this? I’ve got a close eye on Zencastr, Cast, and on the progress of implementing WebRTC in WebKit. But for now, there doesn’t seem to be a single voice-over-IP product of any kind that will work on Mac, Windows, and iOS and automatically record local audio.
Since many of my podcasts feature more than two people, two-person tools like FaceTime are not an option. ↩
The open-source gaming VoIP app Mumble offers multi-track recording and mobile clients, but recordings aren’t supported on mobile and its ease of use is what you’d expect from an open-source project. ↩
This isn’t just about Skype, but the tools people use to record their audio—if we leave Skype, often those tools have to change, too. ↩
By Jason Snell
March 15, 2017 3:34 PM PT
Last month at the Code Media conference, Apple’s Eddy Cue was interviewed by Recode’s Peter Kafka about a variety of subjects. The trailer for “Planet of the Apps” seemed to get the lion’s share of the attention immediately afterward, but I was intrigued by what he said about podcasting:
I think there’s a huge resurgence in podcasting. And it’s exactly what customers want because it’s the ability of listening to something on demand when you want. And that’s exactly what it’s about. Can we do more and will we do more? Absolutely…. We’re working on new features for podcasts. Stay tuned.
I’ll grant you, Cue didn’t say much about podcasting. He was cryptic as any other Apple executive on stage at a non-Apple event might be. But I care a lot about podcasting—it makes up a surprisingly large share of my income these days—and Apple’s place in the podcasting world has always been a strange one. It has been a prominent player for well over a decade, but a strangely passive one. So much so that last year a bunch of prominent podcasters complained to the New York Times that Apple wasn’t doing enough to help them.
Cue’s remarks at Code Media could easily be interpreted as mumbly marketing-speak by an executive who doesn’t have anything to say. But I take Cue at his word that Apple is “working on new features for podcasts,” and that the company has noted the huge resurgence of podcasting. I suspect that, after more than a decade of slumber, Apple’s about to become much more active on the podcasting front.
A decade of podcast curation
Apple holds such a prominent place in podcasting because, very early on, it embraced the medium as a way to improve what was available for the iPod. In June 2005—at the very predecessor to the Code Media conference, All Things D—Steve Jobs demonstrated podcast integration with iTunes, as a part of iTunes 4.9. Now you could subscribe to a podcast in iTunes and sync episodes directly to your iPod—a process that seems barbaric today, but was a delightful innovation 12 years ago.
Key to Apple’s strategy was its creation of a large and relatively open directory of podcasts. Three years before the App Store, Apple repurposed the iTunes store infrastructure to build a global podcast directory. Anyone could submit their podcasts to Apple’s directory and, once approved, those podcasts would remain in the directory more or less forever. And it’s been pretty much this way ever since. At some point Apple provided podcasters with some back-end tools to make publishing and promoting their podcasts in the directory a bit more hands-on; it was a scattershot process, but in recent years it’s rolled those tools out to a much broader audience of publishers.
Other than adding some podcasting-related features to GarageBand (which it stripped out of a later version), Apple hasn’t been particularly active in the realm of podcasting. There’s a small iTunes team that promotes podcasts in the iTunes interface, and those promotions can be very helpful in acquiring new listeners. Apple’s release of an iOS app for listening to podcasts, and its bundling of that app with releases of iOS, was a huge step forward in both the visibility of the format and the curation being done in iTunes.
After 12 years, Apple’s directory is the definitive directory of podcasts. You don’t have to be in iTunes to be a podcast, but most podcasts are in iTunes. Other directories exist, but iTunes is the big fish. Google’s trying to build one, sort of, with Google Play Music—but Apple has a decade head start. Even third-party podcasting apps tend to use Apple’s directory data, either as their entire directory or as a verification tool for their own homebuilt podcast databases.
Apple looms large in the world of podcasting, but in all this time, it hasn’t really changed its basic approach from what it was in 2005: A simple, open directory of podcast submissions with a set of curated pages to help people find new podcasts to listen to.
When Eddy Cue says Apple is “working on new features for podcasts,” he might not be indicating a change from Apple at all. He could simply mean that there are some new features coming to the Podcasts app that will make it better. But I suspect that Apple’s planning on some larger moves, given the increased popularity of podcasting and the leverage Apple has built up over a decade.
So, shifting into pure speculation mode, here are some things that Cue could be talking about when he promises new features for podcasts:
What Apple won’t do: Provide a lot of user data
This is, I suspect, what every podcasting startup wants. Unlike the web, where user behavior can be closely measured and quantified, podcasting is a bit of a mystery. In general, we know that you downloaded a file—and that’s it. To know more, you need to be inside of the apps that people use to listen to podcasts.
Apple’s Podcasts app might be the most popular single piece of podcast-listening software out there today; if Apple were to measure how its users listen to podcasts and then shared that data with the publishers of podcasts, it could be revolutionary to our understanding of how podcasts work. Podcasters (and podcast advertisers) could know how many downloads lead to plays, how deep most listeners get into any given episode, and whether people listen to or skip the ads. It would be a flood of data, and most modern digital publishers say they love data.
Consider me skeptical. While I’m frustrated by the lack of detail and consistency about podcast listenership—I’ve got a podcast that regularly hits 30,000 downloads by one measurement and 20,000 by a different one—I’ve see what the flood of user data has done to the world of web publishing. Most web data is used to justify reducing ad rates and increasing the invasiveness of advertising.
Besides, if an advertiser is happy with the result it receives on a podcast that claims 20,000 listeners, doesn’t that mean the advertiser is paying the right price? If it turned out that same podcast only had 10,000 listeners for a regular episode, it wouldn’t change the result. In fact, you could argue it shows that podcasts are that much more effective at connecting with an audience. But someone else might use that data to argue for a 50 percent rate cut for the podcast instead. More data doesn’t generally improve the quality or price of advertising.
I’m also dubious about what anyone would do with that level of data. For more than a decade I’ve been flooded with page-view data, and I have ignored most of it and focused on using my judgment to make good stuff. In aggregate, it could be useful to find out when people tune out podcasts, whether certain podcast topics or lengths are more or less successful, and what makes podcast advertising successful versus unsuccessful. But the day to day drone of stats? On the web, you just let it fade into the background, because there’s too much data and a lot of it is conflicting.
I doubt Apple will do anything that increases individual surveillance on the habits of its users, and then shares that with third parties, because that’s not what Apple does. While this is the progress that many commercial podcasters say they want, I don’t think it’s likely to happen.
What Apple might do: Support paid podcast subscriptions
There is a technical barrier to making money from podcasts: They have to be free. The podcast and RSS format make it essentially impossible to charge for podcasts and protect them with passwords, as you can to subscription websites. You can make your podcast feeds secure-ish via obscurity, but a dedicated person can find their way to the crown jewels. Right now if a podcaster wants to wall off content—whether it’s new podcasts, back episodes, or everything—the only real choice is to use a separate app. I listen to “Presidents are People Too” in the Audible app, and “Offices and Bosses” in the Stitcher Premium app for these reasons.
But I know a company with a whole lot of credit-card numbers and a great facility at taking payments on the internet, including subscription payments. Apple could potentially build a paid podcast subscription system, using Apple’s payment infrastructure and its podcast-playing apps, and open it to all podcast publishers. Listeners would still need to download a specific app—Apple’s app—but they could mix the free podcasts in Apple’s catalog with the ones they’re paying for.
This one feels a whole lot more likely to me. Yes, it means that Apple’s podcast directory would shift from its current emphasis on the open standards of RSS to a hybrid model that also features limited-access content. But if Apple wanted to encourage the commercial growth of the podcast world, it would be entirely within its powers to make it happen.
Of course, for an approach like this to work, Apple might need to expand its podcast-playing empire a little bit, which takes us to…
What Apple might do: Expand across platforms
For Apple to get podcast publishers on board with paid podcast subscriptions, it’s going to need to answer the questions about users on non-Apple platforms, most notably Android. The answer here is for Apple to create a version of the Podcasts app for Android, and it wouldn’t surprise me if it did. There’s already Apple Music on Android—why not Podcasts, too? Google Play Music’s support for podcasts is weird, and while there are a few good Android-based podcast players out there—Pocket Casts comes to mind—combining standard podcast functionality with the ability to get access to new, subscriber-only podcasts could be a winner.
iTunes for Windows already exists, but it would be great if Apple created a standalone Podcast app for Mac and Windows alike. Failing that, how about a Podcasts web app that syncs subscription status with mobile players? It works pretty well in Overcast today, and Apple’s upped its iCloud game lately. It could happen.
What Apple will probably do: Keep iterating on app and curation
Beyond offering subscriptions, the Podcast app could get better, with better speed adjustment settings and automatic silence removal. It’s a pretty solid basic player today, but there’s always room for improvement. Maybe it’s time to add chapter support?
There’s never going to be an ultimate solution to the problem of giving people good suggestions about what kinds of podcasts they might like, but I expect that Apple will always keep pushing in this direction, both with curated features like the ones currently in the Podcasts section in iTunes, and algorithmic lists tailored to individual listeners. Maybe there’s some intelligence to be gleaned from Apple Music’s equivalents to these features.
If I had to place a bet on a major change in Apple’s approach to podcasting, I’d place it on adding money to the equation. It’s an area Apple knows well, and it’s already got many of the pieces in place to quickly bring on publishers and create its own library of premium, subscription-only audio programs. All while taking its traditional 30 percent cut, of course, at least for the first year. And if it does that, I’d be surprised it it didn’t offer a version of its Podcasts app on Android, too, just to make publishers confident that they’ve got all their bases covered.
It would be the first major change in how Apple has approached podcasting in the 12 years of the iTunes podcast directory. But after 12 years of inaction, maybe Apple finally feels it’s time for podcasting to become more than just a hobby.
By Jason Snell
February 27, 2017 8:50 AM PT
Audio Hijack 3 has become my go-to tool for recording audio for podcasts and pretty much everything else on my Mac. But even if you’re already using Audio Hijack, you may not realize just how flexible its modular, block-building approach allows it to be.
Let me give you two examples. The first one comes from the Session I use for recording and live streaming podcasts on The Incomparable or Relay FM.
This session is doing an immense number of things at once. It’s recording my microphone as a full-quality mono WAV, deposited to my Desktop, named something like jason-20170227-0834, indicating the date and time the recording was started. It’s recording the audio from Skype and saving that to the Desktop, so I can use it as a reference (or backup if one of my panelists fails to record their own microphone). It’s routing the Skype audio into my USB audio interface so I can hear people’s voices in my headphones.
It’s also routing both sets of audio through Rogue Amoeba’s Loopback, a virtual output device that serves as the audio source for Nicecast, another Rogue Amoeba app that connects to our live-stream servers and lets me stream that mix of my voice and my panelists’ voices to live listeners. (There’s a Volume block on the Skype side, so I can reduce the volume of the skype audio a tad so that it’s the same volume as my own voice.)
Finally, that last mixdown of my voice and the Skype audio is also saved to the Desktop, with some very particular settings. Audio Hijack gives you remarkable control over the audio format your recordings can be saved as. In the case of this mixed-down file, I’m saving it as a 64kbps mono MP3, complete with tags and even custom album art.
Members of The Incomparable get access to a special podcast feed containing an archive of all of our live-streamed sessions. Audio Hijack makes the process dead simple—I upload that MP3 file, unchanged, to my server, because it’s in exactly the proper format, right down to the show art.
Here’s another example that’s one I use less often, but still goes a long way to showing just how powerful Audio Hijack can be. For The Incomparable’s beer episode, I had to record four people around a table in my house, as well as a bunch of people who were connected via Skype.
To do this, I connected my Zoom H6 portable recorder to my Mac in USB interface mode—one of the handy features of this device is that it can transform itself into a six-track USB audio interface on demand—and attached four table microphones for my in-person participants. I connected a multi-way headphone splitter to the output from my Mac, and each of us brought our own set of headphones.
Everything got routed by Audio Hijack: each individual track from the H6 was saved to its own file on my Desktop, and then routed to Skype via Loopback to everyone else on the call could hear us mixed together. I recorded the Skype audio to my Desktop and routed that out to the headphone splitter. Shockingly, the entire thing worked flawlessly, despite it being operated by increasingly tipsy people.
Back in the day at Macworld I was frustrated by how hard it was to set up a multi-microphone recording session in our podcast studio. Getting a civilian to understand how to properly configure GarageBand or Logic for a foolproof multi-microphone recording session? Forget it. But with Audio Hijack, I was able to make it simple, by creating a Session that recorded the output of all four microphones in the studio to individual files, a format I replicated for the beer episode.
Combined with tools like Loopback and Nicecast, there is not a single audio problem on my Mac I have not been able to solve with Audio Hijack. Its flexibility and clever interface continue to amaze me.
By Jason Snell
December 28, 2016 3:49 PM PT
I produce podcasts featuring different people using different microphones in all sorts of different homes, which is to say that the nature of the sound files I receive from my panelists can vary widely.
My goal is to make everyone sound as good as possible for the benefit of the listener—and eliminate telltale background noises that would come and go as different people speak. As a result, I spend a lot of time (and have spent more money than I’d expected) trying to remove noise from people’s audio files.
This sort of stuff isn’t for everybody—you don’t need to buy expensive software and spend a half an hour or longer processing all of your audio files in order to make a good podcast. (Also, in most cases the best long-term solution is to get your panelists to improve their equipment or technique, not to fix it in post.) In fact, there are times when I wonder if all the work I put into the removal of noise from audio files is something listeners even notice. But I notice. And I do think getting the noises out improves my podcasts.
Anyway, there’s a lot of software out there that will let you remove noise from your podcasts. Most of them work the same way: you “train” the software on a portion of the audio that contains only the noise you want to remove, which is generally a moment when your subject isn’t talking. In that moment of personal silence, the recording is pure noise: the whirr of a laptop fan, the buzz of a heater, and the hiss of a microphone that does a very good job of picking up room noise.
If you’d like to try this out, consider Audacity, which is free and offers a de-noising plug-in. Another option is the $149 SoundSoap. Adobe includes a de-noising effect with its audio-editing app Audition. As for me, for the last year or so I’ve been using the $249 iZotope RX 5, which is a combination of audio utilities that let you de-noise, de-hum, and de-reverb audio.
Here are some before-and-after samples. We’ll start with a particularly noisy track from my pal David J. Loehr, which may have actually been recorded in a hotel room, not his usual location. From the waveform, you can already tell this is a noisy track: The big spikes are when David is talking, but when he’s not talking there’s still a pretty thick line. That’s the sign of background noise. (There’s also a big empty gap in the middle; that’s when David muted his microphone entirely.)
iZotope RX 5 also provides a second way of visualizing audio, which is via an orange-tinted interface that indicates noises at specific frequencies. That’s most visible across the bottom of the screen. Those solid bars are background hums—they sit at specific frequencies and just keep on making noise.
Most de-noising plug-ins will take care of background hums, but iZotope RX 5 offers a separate de-hum plug-in that is especially effective at destroying those hums. To remove the hum, I select a portion of the audio that contains the hum and click the Learn button in the De-hum window. Then I select the entire track (or at least the portion of the track that contains the hum) and click Process to remove the hum from the selected area. As you can see in the image below, after I click Process the two orange bars at the bottom of the waveform have vanished from the selected portion of the audio file. That hum has vanished entirely.
While removing the background hum is a major part of the noise-removal puzzle, there’s still other background noise. That’s why I’ll now select a portion of audio and click Learn on the De-noise window. Then I select the entire track (or the portion of it containing the noise I want to remove) and click Process to remove the noise.
As you can see from the image below, the area I processed shows up with the thinnest of waveform lines and appears largely black, with no overlaid orange speckles indicating noise. This “silent” part of the track is now truly silent.
In truth, most of the “silent” portions of my guest’s audio tracks aren’t ever heard by podcast listeners. Whether you use a noise gate or a Strip Silence feature like Logic Pro or Ferrite (that’s my approach), quiet portions of someone’s audio tracks are automatically squelched.
The value in removing noise isn’t making the quiet parts quiet—it’s making it so that the parts in which your panelists are talking don’t also contain hums and other background noise. Even when someone’s talking, there are natural pauses through which the hums and noise can bleed through. If I can remove them from everybody’s audio track, you won’t get distracted when the character of the audio changes dramatically every time someone else starts talking.
The screen shots from iZotope RX 5 are fun, but hearing is believing: Here’s a section of that track from David Loehr, before and after I removed the hum and noise.
I mentioned above that iZotope RX 5 also includes a de-reverb effect. That’s actually the primary reason I upgraded to iZotope from SoundSoap—some of my panelists have very echoey recording spaces. In time, perhaps they’ll change their recording set-up and it won’t be a problem, but I’d like to be able to suppress as much room echo as I can in the meantime.
Musicians add reverb to tracks all the time, but the idea of removing reverb seems kind of crazy. In fact, it requires a whole lot of wacky mathematical modeling of sound decay at various frequencies. But you know what? When it works, it’s magical.
This Christmas, my friend James Thomson joined me (and David Loehr!) for a podcast about the “Doctor Who” Christmas Special. James couldn’t use his usual recording location, however, because his mother-in-law was in town and was sleeping in that room. So he recorded from his kitchen, which was not the ideal recording location. It was a bit echoey.
If you’d like to hear how James’s original audio sounded like, what it sounded like after de-reverbing, and then what it sounded like with de-noising added, here’s a sample file.
Should aspiring podcasters run out and spend several hundred dollars for professional audio software? No. Start with Audacity or, if you’re using Audition, the built-in de-noising features. But if you’re interested in taking the next step—or you’ve got some brutal audio that you need to improve—you’d be surprised at the quality of the results you can get with a little bit of time and some clever software.
By Jason Snell
December 16, 2016 3:17 PM PT
I’ve written about using Ferrite to edit podcasts on iOS, but sometimes a video does a better job of demonstrating how software works. So with that in mind, I edited (or to be more accurate, re-edited) this week’s episode of Clockwise in Ferrite on my iPad Pro and captured the audio and video while I was doing it. The full edit took about 25 minutes, but I’ve compressed it substantially in this annotated video of the process.
(You can see a time-lapse of me editing on the Mac in Logic Pro X if you’d like to compare.)
By Jason Snell
December 2, 2016 10:05 AM PT
There were some podcasters at the Úll conference in Ireland last month, and at one point when we were talking shop I complained again about how iOS doesn’t support files on external storage devices that aren’t photos or videos.
This means that if I travel to record a live podcast using a multi-track recorder like the Zoom H6, I have to bring a Mac with me to offload the files. Oh, sure, I can edit a podcast on iOS with ease, but how to get the files over there?
One of the people at Úll—I believe it was Elias—suggested I try the Toshiba FlashAir Wi-Fi SD card. There have been many Wi-Fi-enabled SD cards—I used an Eye-Fi for years—but this one has an iOS app that actually lets you select any file on the card and open it in any app.
There are a bunch of caveats, as you might expect. The FlashAir app isn’t particularly elegant, but it’s functional. The functionality to open a file in another app via the share sheet is off by default, so you have to turn it on. Wi-Fi cards can suck battery, though the FlashAir turns off its Wi-Fi functions after a few minutes if they’re not being used.
But the upside is tremendous! With this approach I can travel somewhere with only an iOS device and my portable recording set-up, record a live audio session, import those files to my iOS device, and then edit and post that audio session, all from iOS.
Now, this doesn’t get Apple off the hook—its card-reader accessory should really be able to read other file types, and more generally iOS should be able to connect to storage devices and let you see the files, whether they’re photos or Word documents. But it closes another gap for my own iOS-based podcast workflow, and so I’m excited about that.
By Jason Snell
December 2, 2016 9:32 AM PT
One of my recent tech quests has been to find a way to record and edit podcasts when traveling with an iOS device and no Mac. The best approach I’ve found so far—and I’ve used it a few times—is to talk on Skype on an iPhone with a pair of earbuds while simultaneously recording myself on a good microphone on an iPad.
Look, I didn’t say it was a good approach. Just that it was the best one I’d found so far. Though I never travel without my iPhone and iPad, the two-device approach to recording is inelegant to say the least. In addition, the person I’m talking to on Skype hears me through a lousy microphone, and I can’t hear my own voice being returned to my ears. (That’s important, because if you can hear your own voice you can tell when you’re not talking into the microphone, and it makes your own impression of your voice sound less like you’re talking with your ears full of water.)
In testing the Audio-Technica ATR2100-USB for my story about the sub-$100 podcast studio, I realized that I had a better option for iOS-only recording. It’s still clunky, but the person on the other end of the Skype call can hear me clearly, and I can hear my own voice in my ears.
Here’s the trick: The ATR2100-USB is a rarity, a microphone that offers both a USB port, for direct connection to a digital device, and an XLR port, for an analog connection to a mixing board or other audio interface. And you can use both connections simultaneously.
So I attach the ATR2100-USB to my iPad or iPhone with Apple’s Lighting-USB Adapter — the old model will work, my iPhone 7 was able to power the microphone itself, though it’s possible that some models might require a power assist from the newer Lightning-USB Camera Adapter. Once the microphone is attached to the iOS device, it becomes the audio input and output for all apps, including Skype.
I plug my headphones into the headphone jack on the microphone, so I’m getting zero-latency feedback from my own voice as well as hearing the audio from Skype, channeled back from my iOS device.
Once that’s hooked up, all I need to do is record my microphone audio on the recorder while conducting my podcast via Skype. In the end, I’ve had a clear conversation and been able to hear my own voice, and my recorder has a pristine copy of my microphone audio.
There’s one final step—transferring the audio file from my recorder back to the iOS device—which requires more hardware. And this setup still doesn’t let me walk away with a recording of the other side of the Skype conversation, which is useful as an insurance policy in case someone else’s recording fails.
If you don’t already have an ATR2100-USB and a portable recorder with XLR plugs, I don’t think I can recommend that you spend money on this option. But if you happen to have the component parts, like I do, you have a single-iOS-device podcast studio ready to go.
By Jason Snell
November 18, 2016 10:36 AM PT
Podcasting is rapidly becoming an industry, with big money and big companies rushing in. But it is also still what it always was: A place where anyone’s voice can be heard. Anyone can make a podcast and post it to iTunes and, with luck and perseverance, find an audience.
One of the biggest hurdles in making a good podcast has always been the expense of equipment. Audio equipment can be expensive, especially the stuff that’s made (and priced) for professionals. One of the good things about this latest podcast renaissance is that the price of pretty good recording equipment has come down a whole lot lately.
Since I write about podcasting a lot, I get asked a lot about what the right starter set-up should be for a podcaster. To be clear—you could use your iPhone’s microphone or a set of EarPods and record a podcast with no extra investment, and you don’t need to spend a dime to get started. GarageBand is free with every Mac, Audacity is free for everyone, and Ferrite is free on iOS with a couple cheap in-app purchases for extra features.
But if you do want to invest a little bit in a better microphone, where should you put that cash? Here’s my recommendation for how you can get a great set-up for under $100.
At this point my recommendation for a podcast starter microphone is the $79 Audio-Technica ATR2100-USB.
The ATR2100-USB excels at keeping out room echo and other background noise (the stuff that can make podcasts hard to listen to), though that means you’ll need to work on your microphone technique and never stray too far away from the mic, or your voice will fade out rapidly. The good news is, it’s also got a headphone jack in its base, so you can hear your own voice as you speak and get immediate feedback if you stray too far away from dead center.
It’s a really amazing value at $79, and it’s often discounted on Amazon to between $35 and $50. The ATR2100-USB even has an XLR port on the bottom, so if you do end up wanting to plug it into a mixer or portable recorder, you can.
The problem with audio hardware is that you need to buy a bunch of accessories. The good news about the ATR2100-USB is that it already comes with XLR and USB cables, a microphone clip, and a desk stand. You don’t need to buy those.
What you do need to buy is a $3 foam windscreen. The ATR2100-USB requires you to get up close to it (because it’s blocking out room noise and echo!), but getting up close to a microphone can lead to lots of ugly popping sounds from your mouth. The windscreen will help filter those out.
You should also probably buy a $11 shock mount to replace the basic microphone clip that comes with the ATR2100-USB. If your microphone is sitting on a desk or table, you will probably be doing things like typing on your keyboard and bumping the work surface with your elbows. These are noises that you won’t notice, but they’ll reverberate right up through the mic stand and sound like explosions on your recording. A shock mount isolates the microphone so that it floats on a springy set of elastic bands.
Getting it off the table
The prices of all of these products can fluctuate quite a lot, but as I write this, those three purchases meet our goal of staying under $100! If you want to spend a little bit more money, well, there’s always a way to spend more money with audio equipment.
The next purchase I’d suggest is a boom arm or mic stand, to elevate your microphone off of your desk or table entirely. If you’ve got a desk you’re willing and able to semi-permanently mount an arm, buy a boom arm like this one (I haven’t tested that one, fair warning). These arms clamp to your desk (so make sure you’ve got a place you can clamp one!) and generally you can screw on the shock mount you bought above rather than use the microphone clip that comes with the arm.
If you don’t have a permanent podcasting location—I didn’t for years after I began podcasting—consider a stand like this $20 model. I used this stand for quite a while when I was podcasting while sitting on my bed. When I was done, I could just fold the stand up and stash it under the bed.
No matter what your budget, podcasting can allow you to have your voice be heard. And if you do want to spend $100, you can have your voice sound that much better. The choice is up to you—but you don’t need to lay out a whole lot of money regardless.
[Thanks to Antony Johnston, author of the Podcast Guest Guide, for the topic suggestion.]
By Jason Snell
August 19, 2016 2:30 PM PT
Since then I’ve discovered a few new facts worth mentioning:
Auphonic’s got an iOS app, Auphonic Recorder. It’s iPhone only and designed mostly for audio recording, but it contains a share extension that allows me to export from Ferrite and immediately upload to Auphonic, without using something like Dropbox as an intermediary. If I’m using an Auphonic preset I’ve previously configured, it will even automatically begin processing my project using those settings once the upload is complete.
The $10 app TwistedWave Audio Editor will export in MP3 format, upload to Dropbox or a server via SFTP, and supports detailed MP3 tagging.
Depending on my needs, I could see myself using either of these tools. If I want to to audio post-processing and have a bit more fiddly control over every aspect of my tags, Auphonic will do the job. But TwistedWave seems to do the job when it comes to encoding and tagging.
By Jason Snell
August 10, 2016 3:15 PM PT
Last weekend my wife and I took a quick car trip to Ashland, Oregon to catch some plays at the Oregon Shakespeare Festival. The night before we left, I realized I hadn’t edited that weekend’s episode of The Incomparable yet, and I didn’t want to bring a laptop with me.
No problem—as I’ve written about before, I have used Ferrite Recording Studio to edit numerous podcasts over the past eight months or so. I can’t recommend it highly enough if you want to edit podcasts on iOS.
There’s just one thing: Ferrite won’t export projects in MP3 format 1. Neither do many other iOS apps, and the reason is that MP3 encoding is still encumbered by patents. Any app that builds in MP3 encoding is risking a bill of thousands of dollars from some of the patent holders—and so most of them just don’t do it. I’ve searched for an iOS app that would encode my audio into properly tagged MP3s, ready for uploading to my server, but have come up empty 2.
Instead, I turn to the web service Auphonic. Auphonic is free for two hours per month of processed audio, and charges for additional hours of encoding—I bought 10 hours of credits for $22, for example.
Getting my file from Ferrite to Auphonic is a little bit tricky. I export a file from Ferrite and instruct the app to save it to Dropbox. My iPad then uploads the file to Dropbox via the Dropbox app. Once that’s done, I use Dropbox’s Sharing feature to generate a link to the file, and tell Auphonic to use the contents of that URL as my audio source.
Within Auphonic, I can set show art (which I can upload directly from my Dropbox via Safari using iOS’s document-picker interface), tags, and even chapter markers with time codes, as well as the bit rate and file format of the final file. Auphonic also offers optional audio processing, creating a more level volume and reducing noise across the final track. Finally, you can add your own servers—SoundCloud, Libsyn, and any old server via SFTP—to your Auphonic account, and set Auphonic to automatically upload the result once it’s done processing the file.
I was able to export and upload The Incomparable while sitting at a comfortable table in an Ashland pub, drinking their beer and using their free Wi-Fi. Auphonic did the rest, re-encoding the file as an MP3, tagging it properly, and uploading the result to both my Libsyn account and to The Incomparable’s FTP server. When it was all done, I received an email alerting me that the entire process was completed. (It took a couple of minutes, start to finish.)
I wish there were a tool on my iPad that would do everything that the Auphonic web app does, but that may be impossible as long as the MP3 patent remains intact. Fortunately, as far as I can tell the final patents covering MP3 encoding will be expiring in 2017, at which point I’m sure Ferrite (and other tools) will add that feature. In the meantime, Auphonic is a solid and affordable alternative.
Yes, I could just post certain episodes of my podcasts in AAC format (and Ferrite will tag them), but I’d rather stay consistent, and it’s possible there are still some podcast clients out there that don’t like the AAC format. ↩
By Chip Sudderth
May 23, 2016 10:41 AM PT
Hobbyist and professional podcasters alike depend on Microsoft’s Skype for mustering panels and interviewing guests, even as they curse it under their breath for its occasional lack of stability and call quality. Skype is ubiquitous because it’s widely cross-platform, relatively easy to install and use, and free—but it may be time for Mac podcasters in particular to pursue more options.
Skype’s Mac user support forum has been abuzz since December with complaints that the ability to adjust conversation volume had been removed since version 7.25. A Skype community manager acknowledged that client and server changes were responsible, and that restoring the functionality would not be easy: “For the speaker volume controls we are still working out how to address this for the scenarios where OSX global speaker volume controls are not the answer.”
This did not amuse podcasters on the forum, because Skype for Mac now consistently outputs “hot” and distorted audio to both headphones and capturing software. “Double-ending,” or recording both sides of a Skype conversation at the source for the producer to sync, is a podcasting best practice. But if a guest is unable to independently record their side of the conversation or has a technical failure, the producer depends on the Skype track for backup. Since Skype for Mac 7.35, that track is likely to sound jarringly worse than the host’s.
The changes in Skype may relate to a new problem I have in putting together my panel podcast, The Audio Guide to Babylon 5, using Skype and one of Rogue Amoeba’s indispensable tools for podcasters, Audio Hijack. Audio Hijack typically and cleverly captures audio from any Mac application. Using the Skype preset, however, as soon as I press the “record” button Skype audio becomes even hotter and largely unusable if my co-hosts have a recording failure.
Audio Hijack’s technical support team researched the issue and responded to me by email (emphasis added):
We’ve been digging further, and it seems that there’s a bug or major change in Skype that’s affecting Audio Hijack’s ability to capture and split up the input and output audio, and we’re looking into ways of improving that behavior. We might suggest using an alternative method of capturing your audio, by disabling the setting to include audio inputs with Skype, and capturing your microphone separately.
That’s what I did. My new Audio Hijack session (pictured) includes two separate audio inputs: a direct link to my USB microphone interface on the left channel and Skype audio output minus my input on the right channel. (The two inputs don’t even have to be combined into the same file; Jason’s preferred Audio Hijack layout sends each audio source into a separate mono file.) The result is that my Skype recordings are still hot but no longer too hot to use in an emergency. 1
The short-term lesson here is that podcasting tools that directly integrate with Skype may be somewhat risky, as Microsoft changes its clients and underlying technology without considering edge cases. On the Mac side, guests can simply record their side of the conversation using QuickTime Player. Producers can record the Skype track and their own microphones separately.
In the long term, however, this serves as a warning to podcasters. Is podcasting support on the Mac so much of an edge case that we need to more thoroughly explore alternatives to Skype? FaceTime is Mac-only. Google Hangouts, which runs as an extension to Chrome, can integrate with Hangouts on Air and YouTube for live video, but it can be a strain on both bandwidth and resources.
Cast seems to be the most promising alternative for traditional podcasting. Even without using its online editing and hosting services, it seamlessly records and syncs native audio from guests. It’s perfectly designed for novice users: just open an emailed link in Chrome, choose your microphone, and go. The host can directly retrieve the individual MP3 files for editing. In my experiments with Cast, however, it seemed unforgiving to guests with spotty internet service or overburdened computer hardware, and Cast doesn’t support more than four participants at one time.
More challenging to many podcasters is the cost: Cast charges a minimum of $10 per month for 10 hours of recording time. For all its headaches—and if you’re confident you’re not going to need to use its audio output—Skype is free. However, as we’ve seen repeatedly in the social media sphere, if you’re not a service’s paying customer your needs are more likely to be less of a priority when technological underpinnings or business models change.
My podcasting community tends to grumble a lot about Skype. Maybe we should take our attention, and even our money, elsewhere.
Plenty of podcasters use raw Skype audio to begin with. While the resulting audio quality isn’t ideal, a guest with a fast, reliable internet connection and a high-quality microphone should sound all right. ↩
By Jason Snell
May 5, 2016 8:14 AM PT
The bulk of the podcasting I do involves me sitting alone in a room talking into a microphone to other people who are somewhere else, doing the same thing. There are lots of advantages to this approach: It lets me host podcasts with people who live all over the world, for one thing, but it also isolates everyone’s sound. We’re all recording in our own little isolation booths, and that can make editing a whole lot easier, since I can clip out the coughing fit or barking dog from your recording and it won’t bleed through from anyone else’s microphone.
Unfortunately, when you’re recording live and in person, the isolation booth is gone, and things get much more complicated. The environment itself can be noisy and challenging, and using more than one microphone at one time can make things complicated. But on the bright side, you won’t need to spend much time editing, because there’s not much point, since even if you clip the sound out of one microphone, it’ll still be audible on the others.
Here’s the set-up I use for remote recording:
The recorder. I recently upgraded to the $400 Zoom H6, which allows me to record up to six XLR microphones at one time (with an additional adapter for the extra two microphones). My previous recorder, the $160 Zoom H4N, is only capable of recording two XLR microphones alongside its own built-in mic, which wasn’t enough for the larger groups I find myself recording live, so I sold it and upgraded. It’s a great value as a starter recorder, and can double as a USB microphone interface when you attach it to a computer. (And yes, if your subjects are willing to snuggle up a little bit, you can record many people with just two or three microphones.)
I choose to use a portable recorder rather than a computer and a USB interface mostly because it’s a much simpler set-up. With a laptop (or iOS device), you need you make sure it’s got power, you need to tote along a second box for the XLR-to-USB interface (and it may need its own power source), and you have to count on your recording software not to let you down. Small portable recorders are self contained, writing their output to a SD card for later import to a computer for editing. They can be powered by AC power or AA batteries that you can find in any store, in a pinch. It’s better this way.
(You may be asking yourself, can I attach two or more USB microphones to a Mac and record that way? I don’t recommend it. I’ve tried it in the past and the microphones generally seem to get out of sync, so when it comes time to put the tracks together later, it gets all echoey and weird.)
The microphones. I have a small collection of XLR microphones. Look at Marco Arment’s review of XLR microphones for details, but if you’re recording live you’re going to need to buy more than one, so price will be a factor.
The best trait of a microphone for live recordings is that it rejects sound that isn’t directly in front of the microphone. If you record with microphones that tend to pick up a lot of room noise, that noise will be magnified and you’ll get a noisy, echoey recording. I have two $150 Shure Beta 58As, but I also have two $20 Pyle PDMIC58s. If you buy the excellent value $60 ATR-2100-USB, you can take advantage of the fact that this USB microphone can also work as an XLR microphone and let it pull double duty.
The accessories.. All the handheld microphones get covered with a $3 windscreen, and I screw their microphone clips onto a cheap fold-up mic stand. You’ll also need to buy XLR cables, and if your microphones are going to be spaced far away from each other and the recorder, you’ll need to make sure that they’re long enough to manage that.
If you need to use microphones in a space where there’s no table or desk, you could have everybody stand and hold the microphones as if they were ready to belt out some classic rock at the top of their lungs. Or you could invest in a few $25 boom stands. I bought one of these and it’s incredibly flexible—I’ve used it to record in all sorts of environments, and because it’s not attached to a table, it isn’t affecting by people doing something noisy like pounding on that table.
The environment. This is a tough one. Record where you can record; if you can avoid super echoey spaces (empty walls, high ceilings, huge glass windows or doors), do so. Recording outside can be surprisingly quiet, unless you’re standing on a crowded sidewalk next to a major road. If you can find a quiet, non-echoey space, you’ve hit the jackpot. But I’ve done some good-sounding outdoor podcasts and some lousy-sounding indoor ones.
By Jason Snell
April 26, 2016 9:24 AM PT
If you’re podcasting or recording voiceovers for video, you need a good microphone. Fortunately, there are good options to be found even if you’re on a tight budget. Unfortunately, there are so many options that it can be dizzying. I reviewed five low-cost USB audio interfaces in a search to find the best of the many options.
The USB/XLR choice
For most podcasters on a budget, the right microphone is almost certainly a USB microphone. They’re easy to use and convenient—just plug it in to your computer and start recording.
I’ve recommended the Blue Microphones Yeti for years after using one myself for several years, and it’s still a great balance of quality and price.
But as Marco Arment points out in his microphone mega-review, there are a lot of other good options. Right now the Audio-Technica ATR-2100-USB (sold in Europe as the Samson Q2U) seems to be the best buy; for a lot less money than the Yeti, you can get a USB microphone that doubles as an XLR microphone for more complex set-ups, with a built-in headphone jack. If you’re usually recording in an echoey room, this noise-killing dynamic microphone is a great choice.
However, there are reasons to choose XLR microphones over USB models. XLR microphones, differentiated by the large three-pinned XLR connector that’s been in use for ages and has plugged into many an analog sound board, come in many shapes and sizes, including some remarkably good-sounding microphones that are available for astonishingly low prices.
Unfortunately, XLR microphones won’t work with a computer or other audio recorder unless you can connect them to an interface that, in turn, connects to your computer via USB. If you’re planning on recording more than one microphone at a time, XLR interfaces are also handy, because you can connect many microphones to an interface box and then record it all on your computer.
They’re also flexible; I can connect my XLR microphones to anyone’s interface box or mixer, and on more than one occasion I’ve been a microphone short and been able to borrow one from a friend. I also own a Zoom H6 recorder that allows me to connect up to six microphones via XLR cables in a portable setting.
There are a lot of uses, but also a lot of parts—but if you take the XLR plunge, you’ll need not only the microphone, but the interface and (of course) XLR cables to connect them all.
By Jason Snell
March 25, 2016 10:00 AM PT
On Monday afternoon I recorded this week’s episode of Upgrade live from Interstate 280, driving home from the Apple event in Cupertino. It was an experiment—I thought it might be fun to do something different for our post-event podcast, and on a day absolutely packed with work, it also allowed me to do something productive with the long drive between Apple and my home north of the Golden Gate Bridge.
I’m pretty happy with the final result, though I wouldn’t recommend recording every episode of your podcast in a moving car. I’m impressed that we only seem to have received one complaint about the danger of podcasting while driving—if you’re opposed to all in-car phone calls, then we’ll just have to disagree—and happy to have heard from numerous people who were entertained by the sound of my turn signals, the beep of the Automatic connected to my car, and the sound of the sudden downpour that happened in the vicinity of San Francisco International Airport.
A few people were wondering what equipment we used to make the podcast, so here’s the scoop:
My microphone was a Sony ECM-77B, which is a small clip-on design that I usually use for recording videos. With it clipped to my shirt, I was able to record without taking my hands off the wheel of my 2005 Honda Civic Hybrid. I attached it to my Zoom H6 portable recorder, which I bought last year. It’s capable of recording six microphones at once, but in this case I was only recording the one.
Myke Hurley and I tried to chat via Skype, but that connection wasn’t stable, so we switched to the telephone. Myke loaded some credits into his Skype account and called my iPhone from Skype, and I kept one earbud in my ear (you can’t cover both ears while driving in California) and talked to Myke during my drive. Listeners to the live stream heard me sound like I was on the telephone, because I was.
Once the drive was over, I ran the file through the automatic dialogue denoiser plug-in in iZotope RX 5, and then sent it off to Myke so he could use it to replace the audio he had recorded of me talking via the telephone. He imported the file into Logic and manually ducked the audio when I wasn’t talking, so the sound would seem consistent—if we cut it off entirely when I wasn’t talking, the change in sound was really distracting. This was a lot of extra work on Myke’s part, but I think it made the end product sound that much better.
I left Apple in the early afternoon, and there was almost no traffic on my return home, so the podcast literally covers every moment I was driving from Cupertino to my house. We wrapped up the podcast with me sitting in my chair at home as I usually do! I leave the calculation of my average driving speed across the trip as an exercise to the listeners.
By Jason Snell
March 24, 2016 2:00 PM PT
Some additional items that came up after I posted my story about Apple’s new Lightning adapter:
As is detailed on the product’s spec sheet, it works with lots and lots of iPads. It also worked fine with my iPhone 6S, even though it’s not listed on Apple’s chart. The product’s marketing seems focused on the iPad Pro—and USB 3 transfer speeds can only be achieved on the 12.9-inch iPad Pro model—but it’s got more broad off-label utility.
Though my story was focused on the iPad as a podcasting platform, after I recorded with my iPhone 6S I realized how amazing it would be to record a full-quality podcast with nothing but a small, high-quality microphone and an iPhone. Talk about portability! Unfortunately, as I mentioned in the story, Apple needs to open up microphone access to multiple apps on iOS before this can work. (My preferred iOS audio editing app, Ferrite, works just fine on iPhone. It’s just cramped.)
Fraser Speirs asked on Twitter if the Camera app on the iPhone or iPad would automatically pick up the audio from an attached USB microphone. This morning, I attached the adapter and the Yeti to my iPhone 6S and took some video in the Camera app, and the sound that was captured came from the Yeti itself. So the answer is yes!
Phil Schiller said you can connect an iPad to Ethernet via a USB-to-Ethernet adapter, so I tried it. Even in Airplane Mode or with Wi-Fi turned off, I was able to connect to the Internet via my Ethernet adapter, so it works! However, I can’t figure out where you can see any evidence that you’re on Ethernet, or any way to adjust networking settings. But it does seem to work.
By Jason Snell
March 23, 2016 4:09 PM PT
While introducing the new 9.7-inch iPad Pro at Monday’s press event, Apple marketing chief Phil Schiller made an aside about a new accessory, the $39 Lightning to USB 3 Camera Adapter:
This is a really powerful accessory, a USB [adapter]. Sure, it lets you plug in your camera, which many of us do, but because it’s powered, you can use a lot of powered USB devices. For example, you can plug in an Ethernet adapter to get on your corporate network. And for those of you who are podcasters, you can plug in a microphone and do your podcast right from an iPad Pro.
I’m a podcaster and an iPad Pro user, so I considered letting out a cheer in the small Town Hall theater, but didn’t want to be the only one. I looked down at iMore’s Rene Ritchie, two rows in front of me, just as he started to clap, and then I joined in. (We did it, everyone, we got podcasting on an iPad to elicit a cheer at an Apple event!)
It’s two days later and I’ve taken delivery of one of these adapters, and have given it a try. The short version is, yes indeed, it works as Apple indicated. But there are also a few quirks to be aware of—and this doesn’t remove all the roadblocks to using an iPad Pro as a dedicated podcasting machine.
Powering microphones and mixers
Though there’s been a USB-Lightning adapter for some time now, the issue with using a USB microphone for podcasting has been all about power. As Schiller indicated, most common USB microphones require more power than the iPad can deliver—and so they just won’t work if you plug them in to the adapter. One workaround people discovered for this was to attach a powered USB hub to the adapter, and then plug a microphone into the hub… but it was a messy solution.
The new adapter solves this problem by getting wider, adding a Lightning port right next to the existing USB port. This means that you can use a USB device while powering your iPad, which wasn’t possible with the old model. (I sometimes stream live podcast audio via an external USB device, but had to be sure that my battery was fully charged before I did that. Similarly, if you want to hook your iPad to your corporate Ethernet network, as Schiller suggests, you’d probably also want to keep your battery topped up while you worked.)
The power that comes to the adapter via Lightning doesn’t just power the iPad—it’s also feeding the USB device you attach to the adapter. When I first tried to attach audio devices to my iPad Pro, I learned an important lesson: If you want to get power out of the adapter, you’ve got to put power into it. When I attached my USB-to-Lightning cable to Apple’s 5 watt USB power adapter—the tiny cube Apple includes with iPhones—I had no success. When I switched to the larger 12-watt brick, though, everything started to work.
I was able to attach both my Blue Yeti microphone and an XLR-based microphone via the Sound Devices USBPre 2 USB mixer to my iPad Pro with no problem. Both showed up as inputs in Ferrite Recording Studio immediately. This all worked on my iPhone 6S, too—same adapter, same microphones, same result.
One funny thing I noticed accidentally is that when I removed the USB end of my Lightning-USB cable from the power adapter and plugged it into my iMac, it didn’t register the iPad as being present—the adapter seems to only use its lightning port as a source of power.
But we’re not there yet
So once the applause from Phil Schiller mentioning iPads and podcasting on stage dies down, where does this leave us? If you’re someone who wants to record a podcast in person using an iOS device and a USB mixer or microphone, you’re set. But most of the podcasts I do are conversations that are conducted over the Internet, usually using Skype. And for the iPad to be a viable device for those kinds of podcasts, Apple needs to update its software.
In short, the audio inputs on iOS need to be accessible by more than one app at a time. Right now I can make a Skype call on my iPad, or I can record my voice to a file on my iPad, but I can’t do both at once—the moment a second app wants access to the microphone, the first one has to give it up. Changing that one behavior in iOS 10 would be enough to allow me to travel and record podcasts without bringing my MacBook Air with me. (I can already edit podcasts on iOS quite well—I edited this week’s Incomparable on my iPad Pro, in fact.)
There’s more Apple could do here, like offer apps access to system audio or the audio output of individual apps, so I could record the sound coming out of Skype, as I do with Call Recorder or Audio Hijack on my Mac today. This seems less likely to happen to me, but I can still dream. (Skype could also adopt Apple’s existing Inter-App audio, allowing other apps to record its output, but this seems even less likely to me.)
(An aside: Yes, you can record remote podcasts entirely on iOS today if you use two devices, such as an iPhone and an iPad. One of them serves as your Skype device while the other one acts as a recorder. It’s really not an ideal situation, especially if you want to hear both your own microphone input and the voices of the people you’re podcasting with.)
It would also be helpful if Apple improved importing files from USB devices and SD cards. Right now iOS is a whiz at importing photo and video files from attached USB devices and cards, but it fails at other file types. I travel with an audio recorder that saves files to an SD card (and also can attach via USB)—but once I record audio there, there’s no way to transfer it to my iPad. It would be great if external media was accessible via standard iOS open and import sheets. Right now, if I want to travel and record something on my fancy six-track USB recorder, I am unable to work with those files on my iPad without the intervention of a Mac.
So there’s more work to do on this front, but this new adapter removes another barrier. Podcasters like me are now one step closer to the dream of doing it all on iOS. I hope Apple eliminates the final roadblock with iOS 10 this fall. Until then, my MacBook Air will be mandatory equipment whenever I’m traveling and podcasting simultaneously.
By Jason Snell
January 11, 2016 5:03 PM PT
On Monday, Rogue Amoeba released Loopback, a $99 (currently on sale for $75) audio utility that dramatically enhances the flexibility of Mac audio. If you’re a podcaster, DJ, or other person who spends time trying to route audio between different Mac apps, you may find Loopback to be an essential tool.
OS X frustratingly doesn’t let you route audio directly from specific apps and input devices to other apps. With Loopback, you can create virtual audio inputs and outputs that appear in the Sound preference pane and in just about any app that works with audio. (It’s a trick that I previously used Ambrosia Software’s WireTap Anywhere tool for, but that app broke in Lion and is no longer being developed. The open-source tool Soundflower does the same thing, although I find its interface confusing and its compatibility and reliability wanting.)
Loopback uses the audio smarts of the makers of Audio Hijack to create an audio utility that’s reliable and offers an interface that’s much more easily understandable. I’ve been using Loopback during its lengthy beta period, and have found it to be an invaluable tool for some very specific audio needs.
Here’s a simple example of how Loopback can be helpful: Even if you’ve got a multi-channel input device attached to your Mac, Skype will only ever use the first channel. With Loopback, you can create a virtual input device that mixes all the channels of your mixer into a single channel. (When I was at Macworld, we had a ridiculous setup where Skype used an iMac’s audio-input jack as its microphone, fed by an output from our mixing device, so that the people on the other end of Skype could hear all four microphones in our studio at one time. Ridiculous.)
Alternately, if you’ve got two USB microphones, you can plug them both in, and create a virtual input that combines them both. Switch to Skype, choose the new virtual interface as your “microphone”, and the app will be none the wiser. (Rogue Amoeba also suggests that screencasters will like Loopback because you can combine apps and input devices exactly as you want them when you’re recording.)
You define what goes where via Loopback’s simple, drag-and-drop interface. You can also create a “pass-thru device,” which serves as both an input and an output, so that you route sound directly from one app to another—for example, from GarageBand to Skype.
One of the frustrations I’ve had for a while is an inability to play audio clips into a Skype conversation. I actually figured out a way to work around this using Audio Hijack 3, but the approach is only functional when Audio Hijack 3 is actively recording my session. With Loopback, I can create a virtual device that combines my microphone and either iTunes or a soundboard app, and use that device as Skype’s “microphone.”
Is this an esoteric audio tool that will only be of value to people who do weird things with Mac audio? Yep. But if you’re one of those people, Loopback is potentially a workflow-shattering experience—and in the best way.
By Jason Snell
December 4, 2015 11:01 AM PT
Recording a podcast with other people over the Internet can be complicated. Everyone needs microphones, sure, but they also need to connect to you so you can hear one another, and for the best audio quality, they need to record their end of the conversation and then send that file to you.
The new web service Cast makes the recording process easy by not requiring that panelists install any special software (beyond Google Chrome—it doesn’t work with Safari yet) or sign up for anything in order to be a part of the conversation. You just send them a link, they open it in Chrome, and they’re up and running. (The service also provides basic in-browser audio editing and podcast hosting, all in the aim of making it easier than ever to get your podcast heard.)
I tried Cast a few times this summer as a part of the service’s beta test, and wasn’t thrilled with the results, but now that the service is officially ready for the world, I gave it a spin this week. Dan Moren and I recorded a short podcast available to Six Colors subscribers using Cast.
I was pretty happy with the sound quality of the conversation, both as we were talking and when we played it back. There weren’t any noticeable artifacts, and the final version on the server sounded good. Cast works by streaming live audio while simultaneously recording your microphone locally and uploading a higher-quality version in the background.
Cast is limited to three guests (plus the host), but large panels are unruly and difficult to edit (take it from me), so I’m not sure it’s a major limitation.
Cast’s recording interface also takes care to add some features that will be quite useful to hosts and panelists alike. A Show Notes button lets hosts write down information about the recording, including when there were issues that will require attention when it’s time to edit the podcast. And the Raise Your Hand button allows a panelist to indicate that they’ve got something to say, which can help smooth out the conversation—I know a lot of podcasters who type the word “hand” into their Skype windows to get the same effect.
Once the recording is done, you can jump into Cast’s editing interface, or—and I like this feature a lot—just walk away with everyone’s files, recorded locally and uploaded invisibly behind the scenes, and pop them into your audio editor of choice. Since the host controls the start and stop of the recording session, the files all start at the same point, which saves you from having to manually synchronize them. Files come down as 128kbps MP3s, which is absolutely acceptable quality for a spoken-audio podcast. (The first time I tried this with the files from my session with Dan, the download failed. I went back later and tried again, and there was no problem.) The show notes are also downloadable as a text file, tagged to the time code of your recording.
Editing in Cast is pretty basic, as you might expect from a browser-based editor. You can edit out chunks of the entire recording, which is useful to make the beginning and end of the show line up perfectly, as well as remove any digressions or mistakes in the middle. You can also adjust the volumes of various tracks, so you can balance out the relative volumes of all your guests. Unfortunately, you can’t trim out noise from a single track, so if someone has a coughing fit while someone else is talking, Cast can’t help you.
You can add new audio layers to the Cast editor, letting you overlay audio (say, sound effects or music) on your session. There’s also a clever “Wedges” feature, which lets you insert audio that pauses your session, plays the audio file, and then continues your session—useful for introductions, ads, and that sort of thing.
Once you’re done, click Mix and Cast with collapse all your audio files into a single mixed-together file. You can choose Standard mix, which leaves your audio alone, or a dynamic-compression mix, which is supposed to smooth out your audio levels. Unfortunately, I found the dynamic-compression mix to be too aggressive—the whole thing sounded overmodulated.
Cast is $10/month (for up to 10 hours of recording time) or $30/month (for 100 hours of recording). I didn’t test Cast’s podcast-hosting feature, but offering unlimited hosting certainly sweetens the deal if you’re currently playing for hosting with a service like Libsyn or Podbean. I published an excerpt of my podcast with Dan to Cast if you’d like to give it a listen and, in the process, test out Cast’s hosting infrastructure.
If you’re a podcast host who has a lot of different guests, non-technical panelists, or panelists who don’t remember to press the recording button or send you their file in time, Cast offers an appealing and simple way to get good quality audio out of guests without asking them to install Skype. If you’re a podcaster or potential podcaster who is frustrated or confused by the Skype-and-local-recording rigamarole, Cast also seems like a service worth trying. And if you don’t want to do more than basic editing, Cast can potentially be a one-stop shop for all your recording, editing, and hosting, which is quite compelling.
Check Cast out for yourself at tryca.st.
By Jason Snell
November 24, 2015 4:46 PM PT
When I wrote about editing a podcast on iOS using the Ferrite Recording Studio app, and then discussed it on The Talk Show, I heard from a bunch of people who wanted to know what I used to record audio on the iPad.
That’s an easy answer—I didn’t—with a more complex issue wrapped inside it. This is a tough one. Even Federico Viticci of MacStories, who uses iOS to do his entire job, still uses a Mac for recording podcasts.
Audio on iOS is primitive when compared to OS X. Only one app can play audio at a time—if you’re playing music and you open YouTube and start playing a video, your music doesn’t keep playing (as would happen on the Mac)—the music is stopped and then YouTube begins to play. And while the Mac’s innate audio-input abilities are not great (thank goodness for utilities like Audio Hijack and Sound Siphon and Call Recorder for Skype), they’re a darn sight better than what’s available on iOS.
As with playing audio, only one app can record audio on iOS at one time. And yet most of the podcasts I create on iOS require that I use a communications app—usually Skype—to talk to the other people on the podcast. The moment Skype begins a call on iOS, it grabs control of the microphone and any other recording app is stopped in its tracks.
There may be some workarounds possible—GarageBand and other apps have been written to use an app called Audiobus to send audio back and forth across apps. It’s a clever hack, but I’m unclear if it could work with Skype (given that it’s sending and receiving call audio all the time, which is more complex than either playing or recording alone), and even so, it would require Skype to be updated to support the feature. (Skype could, of course, offer a feature that let you record your own microphone locally, or offer a recording of your call in the cloud, but Microsoft seems uninterested in pursuing such features.)
So the best hope here is that iOS gets an update at some point that allows multiple apps to have access to audio input. Every year I hope it’s one of those little features that Apple displays on a slide at WWDC that says, “100+ other great features!” or somesuch. It’s never been there.
In the meantime, there is a way to make a Skype call and also record on a high-quality microphone using only iOS. It’s just kind of ridiculous: You make the Skype call on your iPhone, presumably with iPhone earbuds or other compatible headphones with a microphone, while sitting in front of an iPad that’s attached to a microphone and recording locally. The people on Skype hear your bad microphone, but your good microphone is what gets used on the actual podcast. Serenity Caldwell used this method for both this week’s Incomparable Radio Theater and Upgrade episodes. The risk is that if your recording fails, all that remains is a lousy recording of your voice on a set of earbuds via Skype—not a great backup.
I’ve got a Zoom H6 recorder, so if I wanted to travel with just iOS devices, I think I would just record my microphone locally using that, then transfer the file for editing. That also allows me to bypass another problem with recording on an iPad or iPhone: support for external microphones.
There are a few microphones and mixers out there with a native Lightning connector, but most USB devices that rely on Apple’s Lightning to USB Camera Adapter. Unfortunately, the Lightning connector is limited in the amount of power that it can supply; most USB devices won’t work with it unless you connect the microphone via a powered USB hub. Things get messy quickly. It’s workable—I discovered that even my Sound Devices USBPre2 audio interface can work with the iPad if you bring a powered USB hub and put it in a special compatibility mode—but it’s not ideal.
That’s the longer answer. The short answer is, recording podcasts on iOS today is not as easy as editing them. It can be done, but only with a number of workarounds that aren’t necessary on the Mac, which has a more mature sound system that can handle playing and recording multiple audio streams in multiple apps simultaneously.
Ah, well. Maybe in iOS 10.
By Jason Snell
November 13, 2015 5:04 PM PT
Like a lot of iPad users, I dream of traveling with just the iPad, and no laptop. I’m not sure what it saves me, really—my 11-inch MacBook Air is about as small as they come. But still, it’s a dream.
What gets in the way of it, for me: podcasting. iOS has come a long way in terms of power and functionality, but when it comes to audio there have always been lots of issues. iOS basically doesn’t allow two apps to use the microphone simultaneously, and Skype for iOS doesn’t support built-in recording or a pass-through technology like Audiobus, so if you want to talk on Skype while also recording your microphone’s input, you either need to use two devices or a Mac. 1
Using an iPad to do the kind of multi-track podcasting editing I do in Logic on my Mac has been possible for quite a while. Auria is the app I’ve liked the most for this sort of thing, but its interface always struck me as ungainly. I could edit a podcast in that app, but it was slow, and not very much fun.
This week writer/podcaster Fraser Speirs mentioned a new podcast editor he liked, Wooji Juice’s Ferrite Recording Studio. I had been looking for a project to take on in order to test out the iPad Pro, so I took Ferrite for a spin.
In a word, wow: This is the iOS multitrack editor that I’ve been waiting for. Ferrite has all the features that have made my podcast editing workflow so efficient: Strip Silence, compression, noise gate, ripple delete, quick selection of all following clips. It’s all there. And it’s all built inside an attractive interface that’s a pleasure to use. It’s like Ferrite read my mind.
Only later did I realize that Ferrite did, in a way, read my mind. Canis, the lead developer of Ferrite, has listened to my podcasts and read my articles about podcast editing, and apparently some of that rubbed off on the product? During development, he asked me to send him some of my sample podcast files so that he could test using real-world examples, and I sent him a zipped folder full of the raw files that I use to edit The Incomparable. I just hadn’t connected the dots.
Like Logic, Ferrite will break long podcast tracks into short blocks by removing the silence between noisy passages; just select a track and choose the Strip Silence command from a pop-over menu, then specify a couple of settings. It’s got a built-in compressor and noise gate (able to be turned on via an in-app purchase), to level out volume. Trimming individual blocks of sound is as easy as tapping and sliding a finger left or right. And when I want to pull everything in the project forward or backward in time, I just tap on a clip, then triple-tap to select all of the following clips.
Ferrite works much better for me with a keyboard than without, mostly because I spend an awful lot of time pressing the space bar to toggle playback on and off. There’s a play/pause icon on the interface, of course, but it’s way down in the bottom right corner, which is not a convenient location, especially on the enormous iPad Pro screen. I also needed to use the keyboard to rapidly delete clips that were full of stray noise, because Ferrite’s touch-based multiple-clip selection feature is a little bit finicky.
Still, the fact is that my temporary can-I-do-this experiment with Ferrite iPad Pro never reached the stage where I bailed out and decided that I couldn’t do it. A couple of hours later (par for the course for these things), I had an entire finished episode of The Incomparable ready to go 2. (I did have to export the final file back to my Mac to re-encode it as an MP3; Ferrite currently only lets you output projects as AAC files.)
Will I edit next week’s episode on an iPad? Probably not, but that’s more a function of the tools that surround my editing experience (MP3 taggers and encoders, track-sync utilities, and the like) than the core editing experience itself. But for the first time I can see myself traveling with just an iPad and using it to edit podcasts wherever I go. (But if I need to record a podcast on the road, I’ll need to record on my iPad while I’m talking on Skype using my iPhone…)
One final note: I did this all on an iPad Pro, but Ferrite works on other iPad models, and even iPhones. So even if I don’t end up sticking with the iPad Pro, I suspect that I’d have no problem editing a podcast on my iPad Air 2.
Ferrite is free to download from the App Store, with its more advanced features accessible via two $10 in-app purchases. If you’re a podcast editor who dreams of using an iPad to do the job, I highly recommend you give it a try.
By Jason Snell
October 9, 2015 1:51 PM PT
Today Marco Arment released Overcast 2, a free update to his iOS podcast app. There are a lot of great iOS podcast apps out there, but Overcast remains my favorite, thanks to its excellent Smart Speed and Voice Boost features, as well as its flawless speed-boosting features.
Speaking of those features, in previous versions of Overcast they were unlocked when you made an in-app purchase. Beginning with Overcast 2, they’re free. The entire app is free, in fact, with Marco going to a patronage model—he requests donations if you use and like Overcast, to help support its continued development.
It’s an interesting move, but Marco was right to be concerned that the 80 percent of his users who didn’t pay weren’t seeing his app’s most notable features. Now everyone can use those features—and if a small percentage of Overcast users figure that it’s worth paying to thank Marco for his work, it should all work out.
That’s the End of That Chapter
An inside joke in the tech podcasting community has been that, for quite some time now, there have been some vocal podcast listeners who will strongly and repeatedly suggest that real podcasts embed chapter marks. It’s not fair to say that people are almost always German—sometimes they’re Austrian or Swiss.
For a long time I made AAC versions of my podcasts specifically to create chapter marks using GarageBand. But years ago, I gave up and went to MP3 versions only. However, it turns out that the MP3 format does support chapter marks too—it’s just never been supported in most podcast-creation tools or podcast-playing clients 1.
Today, with the release of Overcast 2, the number of people who can take advantage of podcast chapter marks has skyrocketed. If you’re a podcaster wondering how you can add chapter marks to your podcast, your options are limited right now.
In fact, right now I know of only one, and it’s what I’ve been using for Clockwise for the last couple of years: the web app Auphonic. Auphonic is an audio processing tool—you upload your file and then set it to encode it, add chapter marks, provide leveling and filtering, and even automatically upload it to your host. You can process two hours of content per month for free, and there’s a sliding scale of what you need to pay for more processing time.
Auphonic also sells a Mac app called Auphonic Leveler Batch Processor, which does all the leveling and filtering, but unfortunately doesn’t (yet?) support adding MP3 chapter marks.
So for now, if you’re a podcaster and you want to experiment with chapter marks, I’d recommend that you check out Auphonic. But it’s hard to believe that someone won’t build a tool—even a quick and dirty one—to make this something you can do right on your Mac 2.
By Jason Snell
August 12, 2015 2:54 PM PT
I used to edit podcasts in GarageBand, but switched a few years ago to Apple’s $200 Logic Pro. I don’t use most of Logic’s high-end audio production features, but it’s got a few features that make it much better than GarageBand for my purposes.
However, GarageBand is perfectly suitable for podcast editing, and don’t let anyone tell you different. Every Mac comes with GarageBand, meaning every Mac user has access to a free multitrack audio editor capable of generating high-quality podcasts. And while it’s true that the latest version of GarageBand (version 10) lacks some of the podcast-specific features of GarageBand 6.0.5 and earlier, it’s not true that you can’t edit a podcast in the current version of GarageBand. You can! (Earlier on Six Colors I wrote about editing podcasts in more depth.)
GarageBand 10, in fact, based on the same core set of features as Logic, which means you can take advantage of some plug-ins to make your podcasts sound much better—if you can figure out how to use those features. GarageBand doesn’t make it easy. Let me give you a tour of where these features are and make you some suggestions about how you can use them to make a better podcast in GarageBand 10.
By Jason Snell
July 17, 2015 2:18 PM PT
This month Apple’s celebrating “10 Years of Podcasts”, meaning that it’s been a decade since Apple introduced podcasting features into GarageBand and iTunes and added a podcast directory to the iTunes store.
Of course, podcasts have been around for more than 10 years. I remember Shawn King broadcasting radio on the Internet in 1994, and several other Apple-themed podcasts date from the early 2000s. 1 Leo Laporte founded TWiT in 2005, though in a fit of pique about Apple making noises about owning the word podcast, he re-dubbed them netcasts and you still hear that word on TWiT’s promos today.
Prompted by Rene Ritchie, I looked up the first podcast I actually hosted. It was probably Macworld Podcast 27, February 8, 2006, live from a cruise ship in the Pacific Ocean—though I more vividly remember the very next episode, which featured Leo Laporte and was largely conducted in a shipboard bar. As Leo and I talked, more geek cruisers stopped to watch. By the end of the chat, Leo and I had gathered a studio audience, which applauded when we concluded. It was awesome.
The first podcast of my own was the original TeeVee podcast, in July of 2006. It was sporadic and didn’t last very long. I didn’t resume podcasting independently from my job until August of 2010, when The Incomparable debuted. Hard to believe it’s been nearly five years, until I look at the calendar and see that I’ve got to prep episode 256 for posting tomorrow.
These days I host or co-host four weekly podcasts and produce several more. 2 Thanks to the rise of podcast sponsorships (and my departure from my old job), I can say that I’m not just a writer and editor who podcasts on the side—I’m also a professional podcaster.
That’s weird, but it’s good. I love to listen to podcasts and I love to make them. It’s good to be doing something you love. If podcasting couldn’t help me make a living, well, I’d still be doing it. (Just probably not quite as much of it!)
Upgrade (Mondays), Clockwise (Wednesdays), TV Talk Machine (Fridays), and The Incomparable (Saturdays) are my four weekly podcasts. I also do Total Party Kill fortnightly, TeeVee weekly during “Doctor Who” and “Game of Thrones” seasons, Robot or Not irregularly, and parts of the Incomparable Game Show. ↩
By Jason Snell
June 24, 2015 12:21 PM PT
One of the reasons I promote Call Recorder as a tool for Mac podcasters is that it records what you hear on Skype. Whatever microphone is selected as an input in Skype, that’s the one Call Recorder records. So if I can hear you, and you sound good, and you’re using Call Recorder, you’re going to give me a recording of your microphone that sounds good.
When people don’t use Call Recorder, I often discover that while they sounded great on Skype—their fancy high-quality external microphone was selected as the input in Skype’s Audio/Video settings—they were accidentally recording their conversation using their computer’s built-in microphone.
It’s very sad. It means I have to choose with a local recording of a bad microphone or a Skype recording of a good microphone. The Skype recording is generally of pretty good quality, though I prefer a local recording because it doesn’t ever get weird Skype sound artifacts (common when someone has a dodgy Internet connection) and it’s an isolated version of the one person’s voice. A recording of a Skype conversation contains everyone in the conversation, and when they all talk at once there’s nothing you can do to pick them apart.
Anyway, this scenario happened this week. One of my guests accidentally recorded using their computer microphone rather than the good microphone we heard on Skype. So I was going to have to use the Skype recording, but I had local recordings of the other guests.
This is doable, and in fact what I have to do when someone’s local recording utterly fails. (The most recent episodes of Total Party Kill feature a recording failure, so when one person talks I have to delete everyone else’s voices and use the everyone-on-Skype track instead.)
But in this case, I did have a track from the person. It did record a voice, just not one at a quality I could use. To save the day (and my time), I cheated. Here’s what I did.
First, I had to trim the local recording so that it synced perfectly with my Skype reference track. Then I dropped both tracks into Logic and synced all the other local audio files with them, using the Skype track as a reference.
I use Logic’s Strip Silence feature to make noisy areas in a track visible, and remove all areas of a track that contain silence. Once I run the Strip Silence command, only areas containing noise remain on any given track.
In this case, I could use Strip Silence to my advantage. I ran Strip Silence on the local recording of the computer microphone, meaning that Logic was only using that track at times when that panelists was speaking. It was, essentially, a map of when that person talked and when they were silent.
If only I could use that set of Strip Silence-created audio blocks as a sort of audio mask (forgive me, that’s my Photoshop creeping in)? After all, when the panelist is taking, it’s going to be (mostly) just them talking in the Skype track, too.
So that’s what I did. I quit Logic, opened both the local recording and the Skype reference track in Sound Studio, copied the Skype reference track, and pasted it right over the local computer-microphone recording, replacing it entirely. Then I saved the file and quit Sound Studio.
When I opened Logic back up, it did yell at me—it looks like this file has changed!—but then continued on its way. In the place of the old local audio was now the audio from the Skype reference track, but only the moments when my panelist was talking.
At that point, I still had some work to do—stripping out coughs and microphone clicks that weren’t actual talking, removing other audio tracks when there truly was cross-talk, and the like—but it was clean-up work. And much less work than having to manually cut in the Skype track (and cut out all the other tracks) every time the panelist with the bad recording spoke.
By Jason Snell
June 19, 2015 10:53 AM PT
Here’s some podcast/audio nerdery that won’t be of interest to most people, but it’s saved my bacon more than once and just this morning it appears to have saved the bacon of a fellow podcaster, so here goes.
I broadcast my podcasts live using Nicecast, a $59 utility from Rogue Amoeba. One of Nicecast’s, er, nice features is that it’ll also optionally save an archive of your broadcast locally. I’ve enabled this feature, mostly just in case the recording software I usually use—Call Recorder—fails.
There are a lot of reasons I use Call Recorder, most specifically that it records whatever microphone is selected as an input in Skype, so if you sound okay to your fellow podcast participants, your recording will sound okay too. You won’t believe how many times I’ve had it happen that someone has sounded great on Skype, only to send me a local recording of themselves that was made not with their fancy USB microphone, but with the lousy microphone embedded in their laptop or with the (somewhat less lousy) microphone on their earbuds.
But a failure in Call Recorder can be catastrophic. Call Recorder saves its files as QuickTime movies, and if the program doesn’t finish saving that file—say, there’s a crash or a power failure—the entire thing is unsalvageable. So it’s good to have a backup, if not more than one.
Anyway, I had a recording failure a few weeks ago and turned to my Nicecast backup. When I opened the file, I discovered something curious: The file was a stereo recording with both my voice and voices on Skype on the left side, but only the voices of my panel on the right side. This probably happened because I’m using a stereo USB audio interface but only a single microphone.
So I had a thought. Having an isolated audio track of my own voice would improve the quality of the recording and reduce the amount of time I’d spend editing the podcast. Could I somehow subtract the content of the right side from the left, leaving me with a recording of just my own voice?
The answer turned out to be yes. I used my basic audio touch-up tool of choice, Sound Studio, to recover my microphone audio and save the day. First, I copied out each side of the stereo track into their own individual mono files. Then I selected the entire contents of the right track (the one containing just my panelists’ voices) and chose Audio: Invert Signal Polarity.
Go back to high school physics for a second. A wave can be cancelled out by an identical, but inverted wave. This works in the ocean (where two waves can interact and end up cancelling each other out) and it works in sound, too. It’s also a principle used in noise-cancelling headphones.
Anyway, once I inverted the polarity of the panelist-only signal, I copied the result and switched to the window containing the audio of my voice and the panelists together. Using Sound Studio’s Mix Paste command, I pasted the inverted sound over top of the original. And, much to my surprise, it actually worked! The mix paste had subtracted the other voices from the file, resulting in a track that contained only my voice.
Though the quality of the Nicecast archive wasn’t as high as my Call Recorder file (because I was using lower quality settings for the backup), it was still pretty good. I used the track in an episode of The Incomparable and I’m pretty sure nobody noticed a thing.
Like I said, I’m not sure how often this sort of thing comes up in the real world. But if you ever run into this problem, I hope you’ll remember this story and try this approach. It might save your bacon like it saved mine.
By Glenn Fleishman
March 16, 2015 7:00 AM PT
Back in the depths of time, newsletters were a big business. Printed inexpensively in small quantities for investors or people in specialized industries, like lumber or printing, subscriptions could run hundreds to thousands of dollars a year. A few hundred subscribers made them profitable; a few thousand, lucrative, allowing for staff writers and researchers.
These newsletters had timely information that wasn’t found in daily newspapers or weekly magazines. There was no cable news network, and radio was local with generally directed nationally syndicated programs.
Even when cable TV started to add channels, news radio proliferated, and early dial-up services added news features, focused information important for someone’s profession was difficult to find. Newsletters were exceedingly lucrative and appreciated. Some newsletters included or offered cassette tapes — proto-podcasts! — or omitted the paper part entirely and were just audio tapes. I knew of one on desktop publishing that was mass-faxed, and if I recall right, cost $495 per year for weekly dispatches. (MacPrePress, produced by the late Kathleen Tinkel and Steve Hannaford, for those with long memories.)
The Internet’s emergence derailed a lot of these newsletters, because scarce data became easily available, and specialized information sites rose quickly. Often some data previously acquired with scarcity justifying its high expense was suddenly or within a few years available freely, ubiquitously, and instantaneously. In some cases, the dollars shifted: the money paid on postal-dispatched paper shifted to online subscriptions to Web sites, email newsletters, and access to databases. The excellent credit-card industry newsletter, the Nilson Report, now delivers 23 issues in PDF form and snail mail each year for $1,495, for instance, plus the past fives years as an electronic archive. In others, they evaporated entirely.
Blogs were certainly part of the reason. Once easy-to-use blogging software appeared in the early 2000s, a hundred million blogs bloomed, and a tiny portion were dedicated to reporting and analysis, including my own Wi-Fi Networking News (WNN) blog. In a previous era, WNN would have been a newsletter that, based on the interest I saw on the site, would have grossed hundreds of thousands of dollars a year. (The blog brought in $30,000 to $40,000 a year in the mid 2000s.)
Many of the most prolific and focused bloggers who covered tech and finance turned those blogs into businesses, were acquired by larger media companies, or were hired by publications to write for them and often start in-house blogs.
By Jason Snell
February 2, 2015 7:18 AM PT
Once you’ve recorded your podcast, it’s time to edit. Editing can be incredibly simple—trim the beginning and end point and be done with it—or as complicated as you want to make it. I use a few different editing approaches based on my tools and the needs of the particular shows I do. Let me describe them to you now…
By Jason Snell
January 14, 2015 1:16 PM PT
I do a lot of podcasting. And I am often asked about what tools I use and how I produce my podcasts. So in a series of articles on this site, I hope to detail my approach to making podcasts. What I don’t intend is to suggest that this is the only way to make podcasts—it’s just the way that I make them. If I can provide some sort of inspiration—or even a cautionary example of what not to do—I’m glad to do so.
While I think it’s true that many people underestimate how much work goes into making a podcast, I also get the sense that other people overestimate the time I spend. And depending on what kind of a podcast you’re creating, the amount of time required to put it together can vary widely. The average episode of The Incomparable probably takes three or four hours to edit; the average TV Talk Machine I can turn around in 10 minutes.
By Jason Snell
November 28, 2014 11:37 AM PT
I make podcasts as part of my job now, but despite my year spinning records at my high-school radio station, I don’t have much of a background in audio. Like many podcasters, I’ve learned as I’ve gone along, and I’ve upgraded my hardware and software along the way.
I’m frequently asked for product recommendations for podcasting, and while I can’t claim to have tried every USB microphone out there, I have tried many of them and heard the recording results of many more. I’ve also talked to audio experts, sometimes even voluntarily.
Last night I had a couple of exchanges on Twitter that really irked me. I mentioned that the Blue Yeti, the microphone that I use, was on sale at Amazon. (That sale has since ended.) It seems like every time I mention the Yeti on Twitter, I’m immediately sea-lioned by an audio expert who wants to point out that the Yeti is not suitable for professional use.
Point one: I wasn’t recommending it to professionals, I was recommending it to podcasters who are not pros, the ones using headsets and Blue Snowballs and Apple EarPods. Point two: It’s the microphone I’ve used for the last two years, so I think maybe calling it unfit for professional use is not only insulting to me, but wrong on its face.
Anyway, the great thing about podcasting is that anyone can do it. You don’t need to have access to a broadcasting company’s radio transmitter and studios packed with equipment. You can reach people with your voice right now. Yes, these days there are a lot of big names (often from those big broadcasting companies) doing podcasts, but there’s also an incredible diversity of voices and subjects.
If you’re just starting out, don’t allow yourself to be intimidated by all this audio talk. If you have something to say, say it.
I don’t deny that I’ve heard some pretty awful sounding podcasts in my day. Audio quality does matter. I’d just argue that beyond a certain point, it only matters to audio snobs. My favorite podcast, The Flop House, often has some severe audio problems—but it doesn’t matter, because the content is great.
So start with the equipment you’ve got. You could literally do a podcast by talking into your iPhone and posting it. (I don’t recommend it, but you could do it.) Every Apple laptop comes with a built-in microphone. Again, I don’t recommend you use that microphone, but you could. You could use the EarPods that come with your iPhone—and I’d recommend them over that laptop microphone any day. Add an external microphone when you get the chance. Learn how to use GarageBand or Audacity to edit your podcast—both of them are free.
Beyond that, here’s a tiny bit about hardware.
By Jason Snell
November 24, 2014 9:20 AM PT
Today Moisés Chiullan announced that Brett Terpstra’s Systematic podcast and Christina Warren and Brett Terpstra’s podcast Overtired are moving from 5by5 to ESN.fm.
There’s been a lot of podcast movement lately, which isn’t really surprising given how young this medium (or whatever) is. Not everyone finds podcast networks valuable, but they can helpfully group shows of similar sensibilities together, provide exposure for new shows that might otherwise be missed, and offer a technical or financial infrastructure that can be convenient for people who have something to say but don’t want to build a podcasting business1.
And sometimes after a while, those hosts or shows are ready to spread their wings, creatively or technically. Plenty of talented hosts have left 5by5, but you know what? My pals Merlin Mann and Andy Ihnatko are still there, and the indefatigable Dan Benjamin’s producing new audio and video shows all the time.
Since we moved Clockwise from IDG (with the blessing of some nice folks in IDG management) to Stephen Hackett and Myke Hurley’s new Relay FM network, the audience of that show has more than doubled. Being on Relay helped expose the show to a great audience of tech-podcast listeners, and has also helped us grow Upgrade rapidly.
I should mention that as of the most recent episode of The Incomparable, I’m no longer posting episodes to the 5by5 network. We started the show in 2010 and quickly Dan started recruiting me. A little more than a year later, we joined 5by5, and it helped expose my odd little pop-culture show to a much wider audience2.
As time wore on, I decided I wanted to build something on my own, and launched spin-off shows on The Incomparable Network. That project also allowed me to add show metadata that 5by5 simply couldn’t or wouldn’t offer, like a page of all our Star Wars episodes or an index of show topics.
At that point the clock was ticking. I began posting the show to both networks. After a communication failure at 5by5 forced me to abandon a live episode just as it was starting, we set up our own live-stream system that we could control. And most recently, I gave Dan notice that we were changing ad-sales teams. The relationship was at an end. It was time to make it official.
I’m a believer in the medium—it’s one of the ways I expect to support myself and my family now that I’m on my own. But these are the early days. Things are changing rapidly. There are always new podcasts and new networks. (And yes, it’s worth reminding ourselves that this is not the only new-media opportunity out there.)
This reminds me of nothing more than the early days of the web. The younger people out there might not remember, but that period was like the wild west. Things changed every day. Podcasting’s going through something similar.
Anyway, thanks to everyone out there who has listened to some of my podcasts. And best of luck to Brett and Christina on their new adventures with Moisés at ESN.
John Gruber, Marco Arment, John Siracusa, and Merlin Mann were unlikely to have devoted the time to podcasting when they started—but Dan Benjamin offered technical expertise and an ad-sales infrastructure, as well as being an excellent conversational foil.↩
Nothing really changed with the production of the show when we moved—I’ve produced and edited almost every episode, and Dan never had any input into the content.↩
By Jason Snell
November 17, 2014 10:11 AM PT
There’s a lot of talk about podcasting these days, mostly because big names from public radio are doing interesting new things with the medium, and people who write for major media outlets tend to listen to public radio. All of a sudden, thanks to the imprimatur of big media, podcasting is apparently back. Even though all the tech geeks have been listening to podcasts for years now, and it’s been growing as a medium all this time.
Still, as a huge fan of the medium (you may have noticed), I’m happy that more attention is being paid to it. A rising tide lifts all boats—and this stamp of approval from mainstream media will reach future podcast listeners and future podcast advertisers alike. It’s a good thing.
Media outlets aren’t the only ones suddenly paying attention to podcasting. Today Ingrid Lunden at TechCrunch reports that Spotify’s app includes hidden references to podcasting features. This follows the purchase of podcast service Stitcher by Spotify competitor Deezer last month.
More importantly for Spotify, Deezer gave me smart explanation of why podcasting was interesting: Deezer is making a big move to do more with in-car services, and podcasts and talk radio are especially popular in that setting. It could be that Spotify, which also has a number of connected car integrations in place, is thinking along the same lines.
Podcasts are replacing the radio for tech savvy car commuters, and once less savvy commuters are exposed to podcasting I suspect they’ll do the same. I’m not entirely convinced that Spotify is the best vehicle for this, but someone’s going to crack it. As Marco Arment wrote yesterday, it may take some time:
Smartphone podcast apps and Bluetooth audio in cars have both helped substantially, but both have also been slow, steady progressions that are nowhere near complete. No smartphone app has caused a massive number of new listeners to suddenly flood to podcasts, and people don’t upgrade their cars frequently enough for any automotive media features to cause market booms. A lot of people still listen to podcasts in iTunes, and a lot of cars still don’t have Bluetooth audio. We’ll get there, but it takes a while.
If one of the biggest concentrations of podcast listenership is in the car, then the difficulty of connecting podcasts to cars becomes the biggest barrier to the success of the medium. Car tech has traditionally been terrible, thanks to the weird dance between automakers and their equipment suppliers—but that’s starting to change, mostly thanks to Google and Apple. The new Android Auto and CarPlay features allow most new smartphones to project a simplified version of their interfaces onto the screens of compatible car-entertainment devices.
Yes, as Marco points out, this will take years to trickle down to most cars, but it will. It makes too much sense to let the likes of Google and Apple drive these entertainment systems with the much better hardware and software that’s in the pocket of almost every driver.
While I think there’s a huge opportunity to bring the podcast medium to a broader collection of listeners—if I were to do a tech startup, it would probably be something related to this—I’m not convinced that the Spotifys of the world are the right companies to do it. Spotify’s brand is about music, not talk. It’s also unclear what Spotify’s terms would be, and as someone who thinks Stitcher’s terms are really crappy, that’s a serious concern.
No, the company that could do the most to make podcasting a success is Apple. Apple’s got the biggest directory of podcasts on the planet at iTunes and the two most popular podcast-listening apps (Podcasts and iTunes). In the mid-2000s, Apple tried to make podcasting the next big thing, and the world wasn’t ready. Apple’s commitment to podcasting dramatically receded after that—remember when GarageBand was for podcasting?—but with iOS 8 it added Podcasts as a default app, so maybe the tide is turning.
It’s great that podcasting is having a moment in the spotlight. Maybe this is the right time for Apple and other tech companies to forget about the false-start of 2005 and bring this amazing medium to the masses. I’m pretty sure they’re going to love it.
[Hat tip to Federico, Stephen, and Casey.]
By Jason Snell
September 29, 2014 8:08 AM PT
[This is probably the first in a series of posts about nerdy podcast things. Apologies to everyone who’s not a podcaster. Are there people left who aren’t yet hosting their own podcasts? Your time will come…]
At WWDC this year, I hosted a bunch of podcasters in IDG’s podcast studio. (You can drive up to Mill Valley and use my garage next year, folks.) During the recording of Accidental Tech Podcast, I noticed something interesting: Marco Arment was streaming his show live from his iPad.
As someone who streams his own podcasts live, I was intrigued by Marco’s setup. And while Marco uses this particular setup when he’s on the road (he has a mixing board when he’s at home), for the past few months I’ve been using the same setup to stream The Incomparable. From an iPad mini. (I usually use Nicecast from Rogue Amoeba, but various aspects of my Mac’s audio system began behaving strangely when I started using the Yosemite betas.)
In fact, one of the great advantages to this approach is that you don’t have to deal with the Mac’s finicky sound system, which should be much better than it is. (I’d like to be able to, for example, route a couple of USB microphones and the audio from a couple of Mac apps into a virtual input that gets sent out over Skype. There was some great software that used to do this, but most of it died when Lion was released, believe it or not.) Some new software is slowly starting to appear that fills in the gaps, but the beauty of using an iOS device to stream audio is that your Mac doesn’t have to worry about any of that—all it has to do is play sound, which it’s doing already.
The centerpiece of what I’ve taken to calling the Marco Method is the Behringer UCA202, a $30 USB audio interface. Combine that with Apple’s Lightning to USB Camera Adapter, plug into your iOS device, and you’ve got the start of something. (Yes, iOS devices supply enough power to the UCA202 to keep it running, which is not the case with many USB-based audio interfaces.)
Next up is an RCA-to-minijack cable. The RCA inputs plug into the UCA202, and the minjack goes where I would normally plug my headphones—when I’m podcasting, that’s my Blue Yeti USB microphone. The UCA202 has its own headphone jack and volume plug, so I plug my headphones in there and can ride the volume wheel to get the right volume for my ears, separate from the right volume for the live stream.
That’s the hardware side. On the software side, Marco discovered a $5 app by Anthony Myatt called iCast Pro. It’s not much to look at, and it’s an iPhone app so it runs in blown-up mode on an iPad, but it connects directly to an Icecast server, which is what both of us use to stream live. The Icecast server then relays the audio stream to anyone who wants to tune in.
This approach doesn’t provide any way to charge the battery of the iOS device you’re using to stream, but my fully charged iPad mini could probably stream for five hours before running out of juice. I haven’t yet had the chance to test out this setup in the field, but it really allows you to stream live from just about anywhere. Thanks for the tip, Marco.