It’s an interesting subject, and I wasn’t aware of some details but there’s a few discrepancies.
I get the whole normalization happening with streaming services (pandora, spotify, etc.) but doesn’t the fact that 90% of all the music they are streaming still adhere to current volume trends sort of ignore the effort (if there is one)?
>>> At the moment, sure. But that’s going to change quickly once people realize that mastering to the current volume trend ends up sounding worse than leaving a bit more dynamics in. <<<
You’re a mastering engineer, and I’m not going to act like I know more than I do; but don’t you think this is a bit of a sidestep? I mean, our work as engineers has always been to make lasting work that sounds intentional at whatever volume, on whatever listening system. To then go: “No wait, this is going to be streamed, it needs a special version for that.”, doesn’t that seem counteractive?
>>> It’s one more step in the process FOR SOME PEOPLE, but I don’t think it’s a sidestep. It’s more like an intermediary step for those people who are concerned with how their music will sound in both systems. Most artists will probably just ignore it until they can’t any longer, but some of my clients clearly hear the difference and want a version that sounds the best for both delivery methods. <<<
“The new 1770 standard only allows for raising and lowering of the overall song volume, the audio content itself is not being altered at all.”
So basically, peak normalization. No dynamic changes, just volume.
(And let’s be real, everything these days peaks at 0dB, or close, so we’re turning down volume most likely)
>>> Correct, it’s peak normalization. <<<
“In short, more and more ways we consume audio are being set up to automatically adjust the playback volume in order to achieve consistent loudness.”
This sounds like compression, or RMS normalization. Which begs the question, what are we talking about here?
If the 1770 standard does peak normalization, as you first stated, then I don’t believe there’s a need for concern. If 1770 is RMS, then OK, time to panic. If the RMS normalization you’re speaking of is a separate entity like the kind that we already have a la: iTunes, Youtube, Streaming, etc. then how is this news? Any audiophile worth their salt knows to turn normalization off on their playback devices, so where’s the concern?
>>> The measuring of the loudness of the files is the only part of this that involves any sort of RMS aspect. The track is analyzed using the guidelines in the spec to determine it’s loudness based on methods that approximate the way the human ear works. It’s assigned an RMS value if you will, and then all tracks are scaled with peak normalization so that they all playback at the same perceived loudness. <<<
I personally find it quite scary that your clients have requested 2 separate masters because of this (to me still not altogether clear) concern. Again, in my opinion, a master is a master. Just like how I don’t expect a master to compress the hell out of dynamics because it needs to sound loud, I don’t expect it to sound thin because it needs more dynamics at low volume. It just doesn’t make much sense to me; not to mention, they’re complete opposites!
>>> I have some clients now who ask for specific Soundcloud versions too, knowing their steaming encoders do weird things to bass heavy tracks sometimes. Or people who want alternated Mastered For iTunes versions. At the moment I don’t see wanting a seperate 1770 version as any different. Some people will care enough to bother with the versions, some won’t until they have no choice.
The concern here is that a track mastered to current loudness levels likely has a LOT of limiting and compression done to it, and often not for artistic reasons, but in order to “compete” with other tracks while browsing (say Beatport). When you play back these tracks in a peak normalized system, instead of being louder than other tracks, they end up quieter and sounding flat by comparison. Because they are flat in comparison, they’ve had a lot more dynamic range removed already. <<<
Now for the whole “loudness wars”, I don’t want to write a whole thesis, so I’ll just ask, how is this related? It sounds like this 1770 standard, is just that, a volume standard. How will this derail us from current dynamic/volume practices? It’s already factual based on psycho-acoustic analysis and the fletcher munson curve that we as human beings perceive sound non-linearly; which as an effect makes us believe that louder sounds better.
Whether we take this practice too far by over compression, is debatable, but you can’t argue against peak volume consistency, with the ceiling being 0dB.
>>> I think I answered that above. <<<
To wrap this up, though I realize I probably started a few discussions, I’m confused by the concern. I’d still like to believe that the 1 final master coming out of the studio should be the reference point, not to over complicate things.
>>> In an ideal world, that’s definitely the case. And I think the shift to this new way of working will lead us to that faster than anything else too. Right now there is no consistancy in how albums and songs from different artists are played among various media outlets. This will fix that issue if enough get people onboard, one master should translate the same in terms of loudness compared to other songs by other artists. It’s an attempt to solve an issue now, not create a new one in the future.
In the short term, yes it makes a little more work with alternate versions (for those that actually care), but in the long term this should give the artist a LOT more flexibility in how they mix and master their tracks. I think it’s important to realize there’s very little differences in the versions too. It mainly eliminates the need to over limit for volume’s sake alone, allowing the artist to focus on more creative uses for buss processing. These aren’t drastically different versions, or a requires new way of writing a song. Most people will just limit their mixdowns less and everything else will be the same. <<<