1 February 2024 to 27 April 2018 · tagged listen/tech
¶ We Will Know Ourselves by Our Love; We Will Know You By What You Let Go · 1 February 2024 listen/tech
Collective listening is a cultural investment. Collected listening data can be valuable for music streaming services' selfish business purposes, of course, but it's generated by music and listeners, and should be valuable to the world and to music first.
It was my job, for a while, to try to turn music-listening data into cultural knowledge. My opinion, from doing that, is that there are four fundamental kinds of socially valuable music-cultural knowledge that can be learned, with a little attentive work but no need for inscrutable magic, from listening.
The first is popularity. The most fundamental change in our knowledge about music and love, from the physical era to the streaming era, is that we now know what every listener plays, instead of only what they buy. In its simplest form this produces playcounts, and thus the most basic form of streaming transparency and accountability is showing those playcounts. Streaming services have to track plays for royalty purposes, obviously, but music accounting is done by track, and cultural accounting is done by recording and song. At a minimum, we consider the single and the reappearance of that same exact audio on the subsequent album to be one cultural unit, not two, and thus want to see the total plays for both tracks combined in both places. Most major current services do this adequately, albeit at different levels of precision (and one major service glaringly does not display playcounts at all). But really, as people we know that the live version of a song is the same song as the studio version, and if we ask each other what the most popular song on a live album is, we do not mean which of those literal live recordings has been played the most, we mean which of those compositions has been conjured into the air the most across all its minor variations. So far no service has attempted to show this human version of popularity in public, although probably all of them have some internal representation of the idea for their own purposes. (I have worked on various logistical and cultural issues around song identity and disambiguation over the course of my time in music data, but never on the actual mechanics of music recognition, ala Shazam.)
The second kind of knowledge, derived from the first, is currency. We would like to know, I think, what music people are playing "now". Ariana Grande's new song is currently hotter than her old ones, even though it is nowhere near the total playcount of the old ones yet. This can be calculated with windows of data-eligibility, or by prorating plays by age, and most major services do some version of this, but only share it selectively. Spotify, for example, uses an internal version of currency to select and rank an artist's 10 most "Popular" tracks, but only those 10, and the only numbers you actually see there and elsewhere in the app are the all-time playcounts. I worked on a currency algorithm at the Echo Nest, before we were acquired by Spotify, but it's hard to do this very well without actual listening data, and the one Spotify had already devised from better data, without us, produced better results without being any more complicated.
The third kind of knowledge, moving a big step beyond basic transparency, is similarity. Humans listen to music non-randomly, and thus the patterns of our listening encode relationships between songs and between artists. Most current services have some notion of song similarity for use in song radio and other song-level recommendations, and also some notion of artist similarity for behind-the-scenes use in artist radio and more explicit use as some kind of exploratory artist-level navigation ("Related Artists", "Similar Artists", "Fans Also Like", etc.).
I worked on multiple generations of these algorithms in my 12 years at the Echo Nest and then Spotify, and as of my departure in December 2023 the dataset for the "Fans Also Like" lists you see on artists pages in the Spotify app was my personal work. In my time there I had many occasions to compare competing similarity algorithms, both in and out of music, and in a better world less encumbered by petty confidentiality clauses, I would cheerfully bore you with the tradeoffs between them at great length. In my experience simple methods can always beat complicated methods because they're so much easier to evaluate and improve, and time spent refining the inputs is usually at least as productive as tweaking the algorithms themselves, but much less appealing in engineering terms. I consider the calculated similarity network of ~3 million Spotify artists, as I left it, to be a historically monumental achievement of collective listening made mostly possible by streaming itself, but having had to do a lot of internal lobbying on behalf of the musical cogency of similarity results over the years, I am forced to concede that my personal stubbornness is more relevant than any one individual ought to be in this process. Spotify still has my code, but stripped of my will and belief I'm not sure it will thrive or even survive. My individual layoff doesn't necessarily express a Spotify corporate opinion on any larger subject, but it's hard to deny that if Spotify cared, organizationally, about giving the assisted self-organization of the world's listening back to the world, my individual production role in this specific form of it would have been a trivial and uncontroversial excuse for not letting me go. If they give up on this whole feature as a result of one person's absence, it will be a tragic and unforced loss for everybody.
The fourth key form of music knowledge, moving up one more level of abstraction from pairwise similarity, is genre. Genres are the vocabulary by which we understand and discuss music, and genres as communities are the way in which music clusters together in the world. Genres are communities of artists and/or listeners and/or practice, and usually some combination of all three. AI music will be meaningless and inherently point-missing if it attempts to apply sonic criteria without any references to communities of creation or reception, and it will turn out be just one more non-scary new tool in the long history of creative tools if it ends up rooted in how communities sing to themselves about their love. There is no "post-genre" music future, or at least no non-nihilistic one, because music creates genres as it goes.
There are three ecosystemic ways to approach the data-modeling of musical genres: you can let artists self-identify, you can crowd-source categorization from listeners, or you can moderate some combination of those inputs with human expertise.
Two of those ways don't work. Artists self-identify aspirationally, not categorically. If you try to make a radio station of all the rappers who describe themselves as simply "hip hop", you will get a useless pool of 75,000 artists from which most will never be selected. Listeners, conversely, describe music contextually, so two different listeners' "indie pop" playlists may be using the phrase "indie pop" in totally unrelated ways, and thus may have no cultural connection at all. But motivated humans, especially if they know some things about music and are willing to learn more, can mediate these difficulties and channel noisy signals into guided and supervised extrapolations.
You might expect that a global music-streaming service, in recognition of its dependence on music and thus its responsibility to steward music culture, would have a large dedicated team working constantly on systematic, culturally-attuned genre-modeling. Spotify did not. It had editors making playlists, which is sometimes a form of genre curation and sometimes is not. It had ML engineers trying to find correlations between words in playlist titles and tracks, despite playlist titles very much not being a track-tagging interface at all, never mind a genre-categorization tool. It had a handful of people doing specific genre-curation work, mostly on our own initiative because we knew it was worthwhile. And it had me maintaining the genre system, with all its algorithms and all its curation tools. I invented the system (at the Echo Nest, before we were even acquired), I ran it, I supervised it, I tweaked it, I defended it, I believed in it, I helped people apply it to other music and business problems. I had a Slack trigger on the word "genre", so you could summon me from anywhere in Spotify by just typing it. The system grew from hundreds of genres to thousands. My own personal site, everynoise.com (which also predated the Spotify acquisition), was a way to share a sprawling holistic view of it that would never have made sense inside a black-and-green Spotify window or even a white Rdio window before that. I never managed, in ten years of trying, to get genres integrated into the actual daily Spotify music experience (I wanted there to be a list of Fans Also Like genres on artist pages right under the list of Fans Also Like artists; both of these are forms of cultural context and collective knowledge), but I know, from years of emails and stories and other people's independent enthusiasm (including, only shortly before the layoffs, this one in The Pudding, which said "an always-updating catalog of 6,000 genre is groundbreaking" with unfortunate foreshadowing) that I wasn't the only person who understood the value of this whole earnest and unruly and seemingly-endless project.
Will I be proven wrong about the "endless" part? Here, again, we cannot simply conclude that Spotify does not care about genres and music culture because I got laid off. The code remains. Some of the other people who did genre-curation work are still there. Spotify could just keep the internal system running, even if nobody but me would have the inclination or expertise to improve it any further. And maybe they will. I hope they will. It doesn't cost much in computing terms. Spotify is the world's most dominant music-streaming service and genres are how music evolves and exists. Surely one cares about the other.
But if they cared, and one person in a still-8000-person company is basically the smallest practical unit of care, keeping me around would have been self-evidently worthwhile. The genre system wasn't even the only thing I did. The genre system and Fans Also Like weren't even the only things I did. The genre system and Fans Also Like and Wrapped weren't even the only things I did. The public toys I made were the tiniest fraction of my work. If everything I did do wasn't enough, maybe they don't care, and maybe all these things will be unceremoniously abandoned.
But what comes from us, and is made out of our love, of course we can and will rebuild over and over. Spotify is not the only collector of collective listening. These were not the first attempts to connect artists through their shared fans, or to model the genres into which we assemble, and they were never going to be the last. Maybe we will look back on these meager, patchwork networks of only 3 million artists, and only 6000 genres, like we keep the absurdly self-important book reports our kids wrote when they were 9. We are proud of their care and their ambition, not their page-counts. We remember what they dreamed of becoming, and then we hug the people they are in the midst of becoming, and then we think about what we are going to do and become tomorrow.
It was my job, for a while, to try to turn music-listening data into cultural knowledge. My opinion, from doing that, is that there are four fundamental kinds of socially valuable music-cultural knowledge that can be learned, with a little attentive work but no need for inscrutable magic, from listening.
The first is popularity. The most fundamental change in our knowledge about music and love, from the physical era to the streaming era, is that we now know what every listener plays, instead of only what they buy. In its simplest form this produces playcounts, and thus the most basic form of streaming transparency and accountability is showing those playcounts. Streaming services have to track plays for royalty purposes, obviously, but music accounting is done by track, and cultural accounting is done by recording and song. At a minimum, we consider the single and the reappearance of that same exact audio on the subsequent album to be one cultural unit, not two, and thus want to see the total plays for both tracks combined in both places. Most major current services do this adequately, albeit at different levels of precision (and one major service glaringly does not display playcounts at all). But really, as people we know that the live version of a song is the same song as the studio version, and if we ask each other what the most popular song on a live album is, we do not mean which of those literal live recordings has been played the most, we mean which of those compositions has been conjured into the air the most across all its minor variations. So far no service has attempted to show this human version of popularity in public, although probably all of them have some internal representation of the idea for their own purposes. (I have worked on various logistical and cultural issues around song identity and disambiguation over the course of my time in music data, but never on the actual mechanics of music recognition, ala Shazam.)
The second kind of knowledge, derived from the first, is currency. We would like to know, I think, what music people are playing "now". Ariana Grande's new song is currently hotter than her old ones, even though it is nowhere near the total playcount of the old ones yet. This can be calculated with windows of data-eligibility, or by prorating plays by age, and most major services do some version of this, but only share it selectively. Spotify, for example, uses an internal version of currency to select and rank an artist's 10 most "Popular" tracks, but only those 10, and the only numbers you actually see there and elsewhere in the app are the all-time playcounts. I worked on a currency algorithm at the Echo Nest, before we were acquired by Spotify, but it's hard to do this very well without actual listening data, and the one Spotify had already devised from better data, without us, produced better results without being any more complicated.
The third kind of knowledge, moving a big step beyond basic transparency, is similarity. Humans listen to music non-randomly, and thus the patterns of our listening encode relationships between songs and between artists. Most current services have some notion of song similarity for use in song radio and other song-level recommendations, and also some notion of artist similarity for behind-the-scenes use in artist radio and more explicit use as some kind of exploratory artist-level navigation ("Related Artists", "Similar Artists", "Fans Also Like", etc.).
I worked on multiple generations of these algorithms in my 12 years at the Echo Nest and then Spotify, and as of my departure in December 2023 the dataset for the "Fans Also Like" lists you see on artists pages in the Spotify app was my personal work. In my time there I had many occasions to compare competing similarity algorithms, both in and out of music, and in a better world less encumbered by petty confidentiality clauses, I would cheerfully bore you with the tradeoffs between them at great length. In my experience simple methods can always beat complicated methods because they're so much easier to evaluate and improve, and time spent refining the inputs is usually at least as productive as tweaking the algorithms themselves, but much less appealing in engineering terms. I consider the calculated similarity network of ~3 million Spotify artists, as I left it, to be a historically monumental achievement of collective listening made mostly possible by streaming itself, but having had to do a lot of internal lobbying on behalf of the musical cogency of similarity results over the years, I am forced to concede that my personal stubbornness is more relevant than any one individual ought to be in this process. Spotify still has my code, but stripped of my will and belief I'm not sure it will thrive or even survive. My individual layoff doesn't necessarily express a Spotify corporate opinion on any larger subject, but it's hard to deny that if Spotify cared, organizationally, about giving the assisted self-organization of the world's listening back to the world, my individual production role in this specific form of it would have been a trivial and uncontroversial excuse for not letting me go. If they give up on this whole feature as a result of one person's absence, it will be a tragic and unforced loss for everybody.
The fourth key form of music knowledge, moving up one more level of abstraction from pairwise similarity, is genre. Genres are the vocabulary by which we understand and discuss music, and genres as communities are the way in which music clusters together in the world. Genres are communities of artists and/or listeners and/or practice, and usually some combination of all three. AI music will be meaningless and inherently point-missing if it attempts to apply sonic criteria without any references to communities of creation or reception, and it will turn out be just one more non-scary new tool in the long history of creative tools if it ends up rooted in how communities sing to themselves about their love. There is no "post-genre" music future, or at least no non-nihilistic one, because music creates genres as it goes.
There are three ecosystemic ways to approach the data-modeling of musical genres: you can let artists self-identify, you can crowd-source categorization from listeners, or you can moderate some combination of those inputs with human expertise.
Two of those ways don't work. Artists self-identify aspirationally, not categorically. If you try to make a radio station of all the rappers who describe themselves as simply "hip hop", you will get a useless pool of 75,000 artists from which most will never be selected. Listeners, conversely, describe music contextually, so two different listeners' "indie pop" playlists may be using the phrase "indie pop" in totally unrelated ways, and thus may have no cultural connection at all. But motivated humans, especially if they know some things about music and are willing to learn more, can mediate these difficulties and channel noisy signals into guided and supervised extrapolations.
You might expect that a global music-streaming service, in recognition of its dependence on music and thus its responsibility to steward music culture, would have a large dedicated team working constantly on systematic, culturally-attuned genre-modeling. Spotify did not. It had editors making playlists, which is sometimes a form of genre curation and sometimes is not. It had ML engineers trying to find correlations between words in playlist titles and tracks, despite playlist titles very much not being a track-tagging interface at all, never mind a genre-categorization tool. It had a handful of people doing specific genre-curation work, mostly on our own initiative because we knew it was worthwhile. And it had me maintaining the genre system, with all its algorithms and all its curation tools. I invented the system (at the Echo Nest, before we were even acquired), I ran it, I supervised it, I tweaked it, I defended it, I believed in it, I helped people apply it to other music and business problems. I had a Slack trigger on the word "genre", so you could summon me from anywhere in Spotify by just typing it. The system grew from hundreds of genres to thousands. My own personal site, everynoise.com (which also predated the Spotify acquisition), was a way to share a sprawling holistic view of it that would never have made sense inside a black-and-green Spotify window or even a white Rdio window before that. I never managed, in ten years of trying, to get genres integrated into the actual daily Spotify music experience (I wanted there to be a list of Fans Also Like genres on artist pages right under the list of Fans Also Like artists; both of these are forms of cultural context and collective knowledge), but I know, from years of emails and stories and other people's independent enthusiasm (including, only shortly before the layoffs, this one in The Pudding, which said "an always-updating catalog of 6,000 genre is groundbreaking" with unfortunate foreshadowing) that I wasn't the only person who understood the value of this whole earnest and unruly and seemingly-endless project.
Will I be proven wrong about the "endless" part? Here, again, we cannot simply conclude that Spotify does not care about genres and music culture because I got laid off. The code remains. Some of the other people who did genre-curation work are still there. Spotify could just keep the internal system running, even if nobody but me would have the inclination or expertise to improve it any further. And maybe they will. I hope they will. It doesn't cost much in computing terms. Spotify is the world's most dominant music-streaming service and genres are how music evolves and exists. Surely one cares about the other.
But if they cared, and one person in a still-8000-person company is basically the smallest practical unit of care, keeping me around would have been self-evidently worthwhile. The genre system wasn't even the only thing I did. The genre system and Fans Also Like weren't even the only things I did. The genre system and Fans Also Like and Wrapped weren't even the only things I did. The public toys I made were the tiniest fraction of my work. If everything I did do wasn't enough, maybe they don't care, and maybe all these things will be unceremoniously abandoned.
But what comes from us, and is made out of our love, of course we can and will rebuild over and over. Spotify is not the only collector of collective listening. These were not the first attempts to connect artists through their shared fans, or to model the genres into which we assemble, and they were never going to be the last. Maybe we will look back on these meager, patchwork networks of only 3 million artists, and only 6000 genres, like we keep the absurdly self-important book reports our kids wrote when they were 9. We are proud of their care and their ambition, not their page-counts. We remember what they dreamed of becoming, and then we hug the people they are in the midst of becoming, and then we think about what we are going to do and become tomorrow.
¶ bemused complicated partial credit layoff friday morning · 26 January 2024 listen/tech
The glum Digital Music News headline reads "Spotify Daylist is Blowing UpToo Bad the Creator Was Laid Off", and although I haven't specifically talked to the person who came up with Daylist since the layoffs, I don't think they were affected in this round. The explanation in the body of the story is a little more specific:
This is mostly true in what it actually says. I wasn't the only person working on the Spotify genre-categorization project, but I started it, I ran it, I wrote all of its tools and algorithms, and I worked on many applications of it to internal problems and app features. Without me it probably will not survive. And that genre system is one of the ingredients that feeds into Daylist.
The DMN piece is derived from an earlier article at TechCrunch, where the assertion is more carefully phrased: "Spotifys astrology-like Daylists go viral, but the companys micro-genre mastermind was let go last month". And more carefully reported:
...
The "look no further" flourish is misguided, since I didn't curate every individual genre myself, and maybe didn't personally configure any of the ones they cite. We did not make up the name "egg punk", either.
USA Today, drawing from both of these stories, kept the plot twist out of the headline ("How to find your Spotify Daylist: Changing playlists that capture 'every version of you'") and saved it for a rueful final paragraph:
The judicious "help" there is fair enough. And as none of these say, in addition to working on genres I was also a prolific source of this kind of internal personalization experiment, and thus part of an environment that encouraged it.
Daylist itself was absolutely not my doing, though. You'd have to ask its creator about their influences, but so far I haven't seen Spotify give public named credit for the feature, and in a period of sweeping layoffs, in particular, I encourage you to take note of the general corporate reluctance to acknowledge individual work. But while we're at it, I did not have anything to do with Discover Weekly, nor did anybody from the Echo Nest, which was the startup whose acquisition brought me to Spotify and which I did not found. These are not secret details, and a reporter could easily discover them by asking questions. None of three people who wrote those three articles about Daylist talked to me before publishing them.
And although the Daylist feature itself is charming and viral, and I support its existence, it also demonstrates three recurring biases in music personalization that are worth noting for their wider implications.
The most obvious one is that Daylist is based explicitly on the premise that listening is organized by, or at least varies according to, weekdays and dayparts. It is not the first Spotify feature to stipulate this idea, and clearly there are listeners for which it is relevant. But I think both schedule-driven and the similar activity-driven models of listening (workout music, study music, dinner music..) tend to encourage a functional disengagement from music itself. Daylist mitigates this by describing its daypart modes in mostly non-functional terms, including sometimes genres and other musical terminology, and of course you aren't required to listen to nothing but Daylist and thus it isn't obliged to provide all important cultural nutrients. But the eager every-few-hours updating does make a more active bid for constant attention than most other personalization features. Discover Weekly and Release Radar are only weekly, and short. Daily Mix is only (roughly) daily, although it's both endless and multiple. I don't think the cultural potential of having all the world's music online is exactly maximized by encouraging you to spend every Tuesday afternoon the same way you supposedly always have.
The second common personalization bias in Daylist is that it manifestly draws from a large internal catalog of ideas, but you have no control over which subset you are allowed to see, and there is no way to explore the whole idea-space yourself. This parsimonious control-model is not at all unique to Spotify, but it's certainly pervasive in Spotify personalization features, from the type and details of recommendations you see on the Home page to the Mixes you get to the genre and mood filters in your Library. Daylist's decisions about your identity are friendly but unilateral. It's not a conversation. To its credit, Daylist is the first of these features that explains its judgments in interactive form, so you can tap a genre or adjective and see what that individual idea attempts to represent. But this enables only shallow exploration of the local neighborhood of the space. There's still no way to see a complete list of available terms or jump to a particular one even if you somehow know it exists. Obviously everynoise.com demonstrates my strong counterbias towards expansive openness and unrestricted exploration, but one might note that even after 10 years of me working on this genre project at Spotify, there's no place other than my own personal site to see the whole list of genres.
And the third common personalization bias demonstrated unapologetically by Daylist is the endemic tech-company fondness for unsupervised machine learning over explicit human curation. As you can see for yourself by comparing the "genre" mixes you find through Daylists with the corresponding genre pages on everynoise, the genre system is only one of Daylist's inputs. All the non-genre moods and vibes in Daylist obviously come from a different system, but even the genre terms are also filtered through other influences. I did help with those other systems, too, creditwise, but I didn't invent and wasn't running them.
Nor, honestly, do I trust them. You will learn to trust or distrust your own Daylists, if you spend time listening to them or even just inspecting them, but if you follow conversations about them online to get a wider sample than just your own, you will quickly find that they do not always make sense. Mine, right now, claims to be giving me japanese metal and visual kei, but much of it is actually idol rock and a mysterious number of <100-listener Russian metalcore bands that I have never played and which have no evident connection to bands I have. The "Japanese Metal" mix is mostly Japanese, but only sporadically metal. The "Visual Kei" mix is mostly Japanese, and does contain some visual kei, but you'd have to already know what visual kei is to pick those songs out. The "Laptop" mix opens with Morbid Angel's "Visions from the Dark Side", a song that not only was not made on a laptop (to put it mildly), but which narrowly predates the commercial availability of laptops entirely.
The genre system was not error-proof, either. But it was built on intelligible math, it was overseen by humans, and those humans had both the technical tools and moral motivation to fix errors. We did not have a "laptop" genre because "laptop" is not a community of artists or listeners or practice, but if we had, and the system had put Morbid Angel on it, I would have stopped all other work until I was 100% confident I understood why such an egregious error had happened and had taken actions to both prevent that error from recurring and improve the monitoring processes to instill programmatic vigilance against that kind of error.
But once you commit to machine learning, instead of explicit math, you mostly give up on predictability. This doesn't prevent you from detecting errors, but it means you will generally find it hard to correct errors when you detect them, and even harder to prevent new ones from happening. The more complicated your systems, the weirder their failure modes, and the weirder the failures get, the harder it is to anticipate them or their consequences. If you delegate "learning" to machines, what you really mean is that you have given up on the humans learning. The real peril of LLM AI is not that ChatGPT hallucinates, it's that ChatGPT appears to be generating new ideas in such a way that it's tempting to think you don't need to pay people to do that any more. But people having written is why ChatGPT works at all. If generative AI arrives at human truths, sometimes or ever, it's because humans discovered those truths first, and wrote them down. Every problem you turn over to interpolative machines is a problem that will never thereafter be solved in a new way, that will never produce any new truths.
The problem with a music service laying off its genre curator is not the pettiness of firing the person responsible for a shiny new brand-moment. I was responsible for some previous shiny new brand-moments, too, as recently as less than a week before the layoff, but not this one and mere ungratefulness is sad but not systemically destabilizing. Daylist was made by other people, and will be maintained by other people. The problem is that I insisted on putting human judgment and obstinate stewardship in the path of demand-generation, and if that isn't enough to keep you from getting laid off from a music-streaming company, it's hard to imagine anybody else having the idiotic courage to keep trying it.
This is mostly true in what it actually says. I wasn't the only person working on the Spotify genre-categorization project, but I started it, I ran it, I wrote all of its tools and algorithms, and I worked on many applications of it to internal problems and app features. Without me it probably will not survive. And that genre system is one of the ingredients that feeds into Daylist.
The DMN piece is derived from an earlier article at TechCrunch, where the assertion is more carefully phrased: "Spotifys astrology-like Daylists go viral, but the companys micro-genre mastermind was let go last month". And more carefully reported:
...
The "look no further" flourish is misguided, since I didn't curate every individual genre myself, and maybe didn't personally configure any of the ones they cite. We did not make up the name "egg punk", either.
USA Today, drawing from both of these stories, kept the plot twist out of the headline ("How to find your Spotify Daylist: Changing playlists that capture 'every version of you'") and saved it for a rueful final paragraph:
The judicious "help" there is fair enough. And as none of these say, in addition to working on genres I was also a prolific source of this kind of internal personalization experiment, and thus part of an environment that encouraged it.
Daylist itself was absolutely not my doing, though. You'd have to ask its creator about their influences, but so far I haven't seen Spotify give public named credit for the feature, and in a period of sweeping layoffs, in particular, I encourage you to take note of the general corporate reluctance to acknowledge individual work. But while we're at it, I did not have anything to do with Discover Weekly, nor did anybody from the Echo Nest, which was the startup whose acquisition brought me to Spotify and which I did not found. These are not secret details, and a reporter could easily discover them by asking questions. None of three people who wrote those three articles about Daylist talked to me before publishing them.
And although the Daylist feature itself is charming and viral, and I support its existence, it also demonstrates three recurring biases in music personalization that are worth noting for their wider implications.
The most obvious one is that Daylist is based explicitly on the premise that listening is organized by, or at least varies according to, weekdays and dayparts. It is not the first Spotify feature to stipulate this idea, and clearly there are listeners for which it is relevant. But I think both schedule-driven and the similar activity-driven models of listening (workout music, study music, dinner music..) tend to encourage a functional disengagement from music itself. Daylist mitigates this by describing its daypart modes in mostly non-functional terms, including sometimes genres and other musical terminology, and of course you aren't required to listen to nothing but Daylist and thus it isn't obliged to provide all important cultural nutrients. But the eager every-few-hours updating does make a more active bid for constant attention than most other personalization features. Discover Weekly and Release Radar are only weekly, and short. Daily Mix is only (roughly) daily, although it's both endless and multiple. I don't think the cultural potential of having all the world's music online is exactly maximized by encouraging you to spend every Tuesday afternoon the same way you supposedly always have.
The second common personalization bias in Daylist is that it manifestly draws from a large internal catalog of ideas, but you have no control over which subset you are allowed to see, and there is no way to explore the whole idea-space yourself. This parsimonious control-model is not at all unique to Spotify, but it's certainly pervasive in Spotify personalization features, from the type and details of recommendations you see on the Home page to the Mixes you get to the genre and mood filters in your Library. Daylist's decisions about your identity are friendly but unilateral. It's not a conversation. To its credit, Daylist is the first of these features that explains its judgments in interactive form, so you can tap a genre or adjective and see what that individual idea attempts to represent. But this enables only shallow exploration of the local neighborhood of the space. There's still no way to see a complete list of available terms or jump to a particular one even if you somehow know it exists. Obviously everynoise.com demonstrates my strong counterbias towards expansive openness and unrestricted exploration, but one might note that even after 10 years of me working on this genre project at Spotify, there's no place other than my own personal site to see the whole list of genres.
And the third common personalization bias demonstrated unapologetically by Daylist is the endemic tech-company fondness for unsupervised machine learning over explicit human curation. As you can see for yourself by comparing the "genre" mixes you find through Daylists with the corresponding genre pages on everynoise, the genre system is only one of Daylist's inputs. All the non-genre moods and vibes in Daylist obviously come from a different system, but even the genre terms are also filtered through other influences. I did help with those other systems, too, creditwise, but I didn't invent and wasn't running them.
Nor, honestly, do I trust them. You will learn to trust or distrust your own Daylists, if you spend time listening to them or even just inspecting them, but if you follow conversations about them online to get a wider sample than just your own, you will quickly find that they do not always make sense. Mine, right now, claims to be giving me japanese metal and visual kei, but much of it is actually idol rock and a mysterious number of <100-listener Russian metalcore bands that I have never played and which have no evident connection to bands I have. The "Japanese Metal" mix is mostly Japanese, but only sporadically metal. The "Visual Kei" mix is mostly Japanese, and does contain some visual kei, but you'd have to already know what visual kei is to pick those songs out. The "Laptop" mix opens with Morbid Angel's "Visions from the Dark Side", a song that not only was not made on a laptop (to put it mildly), but which narrowly predates the commercial availability of laptops entirely.
The genre system was not error-proof, either. But it was built on intelligible math, it was overseen by humans, and those humans had both the technical tools and moral motivation to fix errors. We did not have a "laptop" genre because "laptop" is not a community of artists or listeners or practice, but if we had, and the system had put Morbid Angel on it, I would have stopped all other work until I was 100% confident I understood why such an egregious error had happened and had taken actions to both prevent that error from recurring and improve the monitoring processes to instill programmatic vigilance against that kind of error.
But once you commit to machine learning, instead of explicit math, you mostly give up on predictability. This doesn't prevent you from detecting errors, but it means you will generally find it hard to correct errors when you detect them, and even harder to prevent new ones from happening. The more complicated your systems, the weirder their failure modes, and the weirder the failures get, the harder it is to anticipate them or their consequences. If you delegate "learning" to machines, what you really mean is that you have given up on the humans learning. The real peril of LLM AI is not that ChatGPT hallucinates, it's that ChatGPT appears to be generating new ideas in such a way that it's tempting to think you don't need to pay people to do that any more. But people having written is why ChatGPT works at all. If generative AI arrives at human truths, sometimes or ever, it's because humans discovered those truths first, and wrote them down. Every problem you turn over to interpolative machines is a problem that will never thereafter be solved in a new way, that will never produce any new truths.
The problem with a music service laying off its genre curator is not the pettiness of firing the person responsible for a shiny new brand-moment. I was responsible for some previous shiny new brand-moments, too, as recently as less than a week before the layoff, but not this one and mere ungratefulness is sad but not systemically destabilizing. Daylist was made by other people, and will be maintained by other people. The problem is that I insisted on putting human judgment and obstinate stewardship in the path of demand-generation, and if that isn't enough to keep you from getting laid off from a music-streaming company, it's hard to imagine anybody else having the idiotic courage to keep trying it.
¶ New Releases by...Something · 25 January 2024 listen/tech
Whatever my job is or, currently, isn't, I still really want to know about new music, and there were exactly two excellent methods for this and I made them both and I'm not allowed to use either one any more. So I keep poking at approaches to this problem using public tools, and the Spotify API is still the best of these. I've gone so far as to implement a version of the obviously-intractable brute-force approach of getting a genre's artists, expanding to all those artists' Fans Also Like artists (refcounting along the way), and then getting the catalog for each artist in the resulting longer list in order to filter their releases by release-date.
This does sort of work, eventually. It's not especially convincing for coverage, because the public API only exposes 20 Fans Also Like artists, where the internal Spotify datasource behind that (at least as long as they keep maintaining the similar-artist system I wrote) had up to 250 for each seed artist. And it takes just about forever, because it requires thousands of individual queries, and even with only one instance of it running my API key quickly hits its rate-limit and gets throttled to wait in between calls.
As I have noted, Spotify could mostly fix this problem by enabling genre: filtering in the /search API when searching for "albums" (which actually means releases despite my 10 years of trying to persuade a nominal music service to take the difference between singles and albums seriously), since this API already has a tag:new filter for getting new releases (from the last 2 weeks, which is also kind of arbitrary but at least means the last full release-week is always completely covered). There's already internal data for artists' "extended" genres, which is the extrapolated version using collective artist similarity. Or at least there is if they keep maintaining the genre system I wrote.
You can see exactly how viable this is, if you're curious and not unmanageably triggered by a thing that takes the shape of our loss without salving it, because the search API does already allow release searches to be filtered by label: the same way it could allow searching by genre. Any API app could do this, there's nothing special about my access or techniques here. But I looked and didn't find one, so I made it. This is what my job was like, too, and apparently I was literally correct when I used to say that I'd be doing it even if they weren't paying me.
Thus: New Releases by Label for, e.g., a list of 58 metal-related labels.
The chances are decent that if too many people try this at once it will slow down or die, too, but for each label it requires as few as two queries: one to get that label's new releases (in pages of 50 if a single label has more than 50 new releases from the last two weeks), and then at least one follow-up query (in pages of 20) to get those albums' tracks. This is reasonable overhead.
Labels are no direct substitute for genres, obviously, not least because if you care about music you need not also care about labels or whether artists are even on one.
And even if you do care about labels, label data is messy. It's something of a stretch to call it "label data", in the current music-distribution ecosystem. There's a text field for "label", and humans type stuff into it. If the humans doing a given bit of typing are diligent, and none of the other undiligent humans stray into the diligent namespaces by accident or nefarity, then you can kind of pretend it's data. I spent a while in my former job trying to do slightly better than this by aggressively normalizing name-variations and algorithmically distinguishing between actual labels and whatever words people who aren't on labels would type into that box, with some success:
I see there that my past-job self could have combined "Hell's Headbangers" and "Hells Headbangers Records" by removing apostrophes, which either didn't occur to me or caused more problems than it solved, and I no longer remember which and can't check.
There are, though, many labels that exist to release a certain kind of music according to some kind of unifying principle, and those principles tend to align with genres, or more accurately tend to be part of the social structure that builds music-based communities, which are what I usually mean when I talk about genres. So this approach is wildly incomplete, but seems at least potentially helpful to me. You can try it with some labels you like, and see if it helps you, too.
The one small catch with this is that the API label filter is very literal. You have to know the exact way the label you're looking for is typed in the "label" fields. And, inconveniently, that label field does not actually appear in the Spotify app.** What you see instead are the copyright and publishing credits, at the bottom of an album's tracklist like this:
You might guess that the "label" here could be either "AFM Records" or "Soulfood Music Distribution GmbH", and as it turns out "AFM Records" is right, but it didn't have to be either of those and guessing is tedious anyway.
But I have a thing for exercising my petty annoyances about how to display albums, already, so if you look up an artist in the everynoise research tools, you can now see each release's label next to its release date.
But looking labels up album by album is tedious, too. The one automated tool left in my new-release workflow is Release Radar, which provides a subsistence level of new-release awareness if you take the time to follow all the artists you know you care about. And I have a thing for exercising my petty annoyances about how to display playlists, too, so I added a label column to it, which you can even click on to sort or group a whole playlist by label:
And if your playlist happens to represent a subset of new releases you know you care about, look at the bottom of that page for a little helpful link to feed all the labels from a given playlist back into New Releases by Label.
Obviously it would be better if there was also a link to find new releases from this playlist's genres, instead of just its labels, and of course that's what this link used to do. And could do, again, in a better future.
We will get better futures. That is, we'll get them if we build them, and we will build them, one way or another, because it's too annoying not having them.
** 1/30: An alert reader notes that the label actually is available in the Spotify app, not on the release itself but in the Song Credits dialog for any of its songs, at the bottom labeled "Source".
This does sort of work, eventually. It's not especially convincing for coverage, because the public API only exposes 20 Fans Also Like artists, where the internal Spotify datasource behind that (at least as long as they keep maintaining the similar-artist system I wrote) had up to 250 for each seed artist. And it takes just about forever, because it requires thousands of individual queries, and even with only one instance of it running my API key quickly hits its rate-limit and gets throttled to wait in between calls.
As I have noted, Spotify could mostly fix this problem by enabling genre: filtering in the /search API when searching for "albums" (which actually means releases despite my 10 years of trying to persuade a nominal music service to take the difference between singles and albums seriously), since this API already has a tag:new filter for getting new releases (from the last 2 weeks, which is also kind of arbitrary but at least means the last full release-week is always completely covered). There's already internal data for artists' "extended" genres, which is the extrapolated version using collective artist similarity. Or at least there is if they keep maintaining the genre system I wrote.
You can see exactly how viable this is, if you're curious and not unmanageably triggered by a thing that takes the shape of our loss without salving it, because the search API does already allow release searches to be filtered by label: the same way it could allow searching by genre. Any API app could do this, there's nothing special about my access or techniques here. But I looked and didn't find one, so I made it. This is what my job was like, too, and apparently I was literally correct when I used to say that I'd be doing it even if they weren't paying me.
Thus: New Releases by Label for, e.g., a list of 58 metal-related labels.
The chances are decent that if too many people try this at once it will slow down or die, too, but for each label it requires as few as two queries: one to get that label's new releases (in pages of 50 if a single label has more than 50 new releases from the last two weeks), and then at least one follow-up query (in pages of 20) to get those albums' tracks. This is reasonable overhead.
Labels are no direct substitute for genres, obviously, not least because if you care about music you need not also care about labels or whether artists are even on one.
And even if you do care about labels, label data is messy. It's something of a stretch to call it "label data", in the current music-distribution ecosystem. There's a text field for "label", and humans type stuff into it. If the humans doing a given bit of typing are diligent, and none of the other undiligent humans stray into the diligent namespaces by accident or nefarity, then you can kind of pretend it's data. I spent a while in my former job trying to do slightly better than this by aggressively normalizing name-variations and algorithmically distinguishing between actual labels and whatever words people who aren't on labels would type into that box, with some success:
I see there that my past-job self could have combined "Hell's Headbangers" and "Hells Headbangers Records" by removing apostrophes, which either didn't occur to me or caused more problems than it solved, and I no longer remember which and can't check.
There are, though, many labels that exist to release a certain kind of music according to some kind of unifying principle, and those principles tend to align with genres, or more accurately tend to be part of the social structure that builds music-based communities, which are what I usually mean when I talk about genres. So this approach is wildly incomplete, but seems at least potentially helpful to me. You can try it with some labels you like, and see if it helps you, too.
The one small catch with this is that the API label filter is very literal. You have to know the exact way the label you're looking for is typed in the "label" fields. And, inconveniently, that label field does not actually appear in the Spotify app.** What you see instead are the copyright and publishing credits, at the bottom of an album's tracklist like this:
You might guess that the "label" here could be either "AFM Records" or "Soulfood Music Distribution GmbH", and as it turns out "AFM Records" is right, but it didn't have to be either of those and guessing is tedious anyway.
But I have a thing for exercising my petty annoyances about how to display albums, already, so if you look up an artist in the everynoise research tools, you can now see each release's label next to its release date.
But looking labels up album by album is tedious, too. The one automated tool left in my new-release workflow is Release Radar, which provides a subsistence level of new-release awareness if you take the time to follow all the artists you know you care about. And I have a thing for exercising my petty annoyances about how to display playlists, too, so I added a label column to it, which you can even click on to sort or group a whole playlist by label:
And if your playlist happens to represent a subset of new releases you know you care about, look at the bottom of that page for a little helpful link to feed all the labels from a given playlist back into New Releases by Label.
Obviously it would be better if there was also a link to find new releases from this playlist's genres, instead of just its labels, and of course that's what this link used to do. And could do, again, in a better future.
We will get better futures. That is, we'll get them if we build them, and we will build them, one way or another, because it's too annoying not having them.
** 1/30: An alert reader notes that the label actually is available in the Spotify app, not on the release itself but in the Song Credits dialog for any of its songs, at the bottom labeled "Source".
¶ The problem with new releases · 14 December 2023 listen/tech
There were a lot of things on Every Noise at Once that updated daily or weekly, and some of them will lose value only slowly now that they can't be updated. The one that loses almost all of its value at once is the weekly New Releases by Genre.
There were no new-release features in Every Noise at Once at its literal outset, pre-Spotify when it was powered by second-hand Echo Nest data and the Rdio API. I found a 2014-07-30 message from me on I Love Music sharing the URL of the first single-page version of a Spotify new-release list, but the earliest capture in the Internet Archive is from 2014-09-24, when Maroon 5's "Maps" was the top single of the week, and the righthand column listed "all 3304 releases this week". I called this version the Spotify Sorting Hat, because certain things hadn't happened yet.
By 2019 that version became untenable, and I rewrote it from scratch to separate the genre collation from the raw list, and introduce more control so you had a prayer of finding the subset of music you actually cared about. That version is dead now without the internal Spotify feeds that provided its data.
You might imagine that there would be alternatives by now, but there aren't any on the same scale. Spotify has by far the best API for this kind of idea, but it isn't quite set up to provide any of the three things that a serious new-release tool should offer:
- "all" new releases: in truth I started imposing thresholds on what the Sorting Hat would include before switching to NRbG, and NRbG never showed literally everything, either. But it showed a lot. The Spotify API for searching allows you to specify "tag:new" and get only things released in the last two weeks, but you can only get 50 of them at a time, and only 1000 total, sorted by popularity. Most weeks there are more than 100,000.
- new releases by genre: you can filter by genre in the Spotify API, but only in artist searches. And you can only use "tag:new" in album searches. So currently they can't be combined. Unlike the all-release issue, this one would be a reasonable feature request for the API, as it fits the existing usage-models and wouldn't be particularly onerous to support. Assigning artist-level genres to albums can get existential if you think too hard about it, but if you stick to the idea that genres are communities, then calling an album "atmospheric black metal" is shorthand for saying that it's an album made by a band that is part of the atmospheric black metal community, which makes sense even if the particular album is acoustic folk pastiches, and in the new-release case gets it to the audience that wants to know about that album, so it's fine.
- discovery: the more complicated thing NRbG did was to try to distribute new releases by bands who aren't really the canonical representatives of any genre to the genres whose fans would be their most likely audiences. This absolutely can't be done using the existing API, for the same reasons that you can't extract a "full" list. You can get the 20 most similar artists for any artist, but for matching unknown artists to genres you need to go in the other direction, finding the 100s of artists whose Fans Also Like lists includes the known artists in a genre. But adding this feature to the API wouldn't be much harder than adding the genre filter itself.
Absent those features, the Spotify API can't be used to build this, and so far all the other services are even farther from being able to provide the tools for it. NRbG worked because the end results weren't confidential, just inaccessible, and I could solve the inaccessibility by running a set of carefully interlocking internal queries. Now I can't.
I don't really know how we can do new-release discovery now, without this. We can go back to the human, community-based modes of knowledge we used to use, of course: mailing lists, discussion forums, blogs, playlists maintained by individual experts. One genre at a time, these ways are usually better than queries, detailed and contextual and exultory. But they can't be aggregated the way data can. You can keep track of a genre or two or five this way, but not 20. Not 100. I've been monitoring hundreds of genres every week, for years. Now I am as lost as you.
There were no new-release features in Every Noise at Once at its literal outset, pre-Spotify when it was powered by second-hand Echo Nest data and the Rdio API. I found a 2014-07-30 message from me on I Love Music sharing the URL of the first single-page version of a Spotify new-release list, but the earliest capture in the Internet Archive is from 2014-09-24, when Maroon 5's "Maps" was the top single of the week, and the righthand column listed "all 3304 releases this week". I called this version the Spotify Sorting Hat, because certain things hadn't happened yet.
By 2019 that version became untenable, and I rewrote it from scratch to separate the genre collation from the raw list, and introduce more control so you had a prayer of finding the subset of music you actually cared about. That version is dead now without the internal Spotify feeds that provided its data.
You might imagine that there would be alternatives by now, but there aren't any on the same scale. Spotify has by far the best API for this kind of idea, but it isn't quite set up to provide any of the three things that a serious new-release tool should offer:
- "all" new releases: in truth I started imposing thresholds on what the Sorting Hat would include before switching to NRbG, and NRbG never showed literally everything, either. But it showed a lot. The Spotify API for searching allows you to specify "tag:new" and get only things released in the last two weeks, but you can only get 50 of them at a time, and only 1000 total, sorted by popularity. Most weeks there are more than 100,000.
- new releases by genre: you can filter by genre in the Spotify API, but only in artist searches. And you can only use "tag:new" in album searches. So currently they can't be combined. Unlike the all-release issue, this one would be a reasonable feature request for the API, as it fits the existing usage-models and wouldn't be particularly onerous to support. Assigning artist-level genres to albums can get existential if you think too hard about it, but if you stick to the idea that genres are communities, then calling an album "atmospheric black metal" is shorthand for saying that it's an album made by a band that is part of the atmospheric black metal community, which makes sense even if the particular album is acoustic folk pastiches, and in the new-release case gets it to the audience that wants to know about that album, so it's fine.
- discovery: the more complicated thing NRbG did was to try to distribute new releases by bands who aren't really the canonical representatives of any genre to the genres whose fans would be their most likely audiences. This absolutely can't be done using the existing API, for the same reasons that you can't extract a "full" list. You can get the 20 most similar artists for any artist, but for matching unknown artists to genres you need to go in the other direction, finding the 100s of artists whose Fans Also Like lists includes the known artists in a genre. But adding this feature to the API wouldn't be much harder than adding the genre filter itself.
Absent those features, the Spotify API can't be used to build this, and so far all the other services are even farther from being able to provide the tools for it. NRbG worked because the end results weren't confidential, just inaccessible, and I could solve the inaccessibility by running a set of carefully interlocking internal queries. Now I can't.
I don't really know how we can do new-release discovery now, without this. We can go back to the human, community-based modes of knowledge we used to use, of course: mailing lists, discussion forums, blogs, playlists maintained by individual experts. One genre at a time, these ways are usually better than queries, detailed and contextual and exultory. But they can't be aggregated the way data can. You can keep track of a genre or two or five this way, but not 20. Not 100. I've been monitoring hundreds of genres every week, for years. Now I am as lost as you.
¶ Our ghosts as echoes · 7 December 2023 listen/tech
The job I used to do at Spotify was to experiment with what could be made out of listening data, but the experiments that worked were then automated to run on appropriate schedules. These stopped working temporarily when my employee account was deactivated, but were revived, and some data-driven daily playlists are updating again. Currently no data updates are making it to everynoise.com, though, and all the big weekly jobs run on Fridays, so it remains to be seen what my ghosts there are still allowed to sing.
¶ everynoise.com · 4 December 2023 listen/tech
everynoise.com is and was a projection of things I worked on at the Echo Nest, and then Spotify. As of this morning, I'm not working there, and I expect most of the things I did will no longer be done. The domain is mine, but the processes that updated the site have already stopped running, and the current server may stop serving at any moment. Once that happens, I will try to bring back at least a wistful memorial to what it was.
But even failing that, its 10-year history lives on in the Wayback Machine. And music, of course, survives our fragile and temporary machines for organizing it.
More information when I have it.
But even failing that, its 10-year history lives on in the Wayback Machine. And music, of course, survives our fragile and temporary machines for organizing it.
More information when I have it.
¶ PopCon 2020 · 5 September 2020 listen/tech
The 2020 Pop Conference starts this week! It's online and free, so you can come!
But you do need to register!
My PopCon contribution this year is not a talk, it's a web explorer for the music that defines, unites and distinguishes kids around the world. It's called The Aqueduct of Youth.
It's part of a panel called The Platforms of Youth: Meme-ing, Marketing & Streaming, with five other explorations of "Old Town Road", Christian musicals, aging, Tiktok and VSCO Girls:
You can watch videos from the other participants, drink from the Aqueduct, and then come and ask us questions in our live Zoom discussion session on PopCon's opening day, Wednesday 9/9, from 10:30-11 Pacific / 1:30-2pm Eastern.
But you do need to register!
My PopCon contribution this year is not a talk, it's a web explorer for the music that defines, unites and distinguishes kids around the world. It's called The Aqueduct of Youth.
It's part of a panel called The Platforms of Youth: Meme-ing, Marketing & Streaming, with five other explorations of "Old Town Road", Christian musicals, aging, Tiktok and VSCO Girls:
You can watch videos from the other participants, drink from the Aqueduct, and then come and ask us questions in our live Zoom discussion session on PopCon's opening day, Wednesday 9/9, from 10:30-11 Pacific / 1:30-2pm Eastern.
¶ 2019 in Music · 6 January 2020 essay/listen/tech
I starting making one music-list a year some time in the 80s, before I really knew enough for there to be any sense to this activity. For a while in the 90s and 00s I got more serious about it, and statistically way better-informed, but there's actually no amount of informedness that transforms a single person's opinions about music into anything that inherently matters to anybody other than people (if any) who happen to share their specific tastes, or extraordinarily patient and maybe slightly creepy friends.
Collect people together, though, and the patterns of their love are sometimes very interesting. For several years I presided computationally over an assembly of nominal expertise, trying to find ways to turn hundreds of opinions into at least plural insights. Hundreds of people is not a lot, though, and asking people to pretend their opinions matter is a dubious way to find out what they really love. I'm not really sad we stopped doing that.
Hundreds of millions of people isn't that much, yet, but it's getting there, and asking people to spend their lives loving all the innumerable things they love is a more realistic proposition than getting them to make short numbered lists on annual deadlines. Finding an individual person who shares your exact taste, in the real world, is not only laborious to the point of preventative difficulty, but maybe not even reliably possible in theory. Finding groups of people in the virtual world who collectively approximate aspects of your taste is, due to the primitive current state of data-transparency, no easier for you.
But it has been my job, for the last few years, to try to figure out algorithmic ways to turn collective love and listening patterns into music insights and experiences. I work at Spotify, so I have extremely good information about what people like in Sweden and Norway, fairly decent information about most of the rest of Europe, the Americas and parts of Asia, and at least glimmers of insight about literally almost everywhere else on Earth. I don't know that much about you, but I know a little bit about a lot of people.
So now I make a lot of lists. Here, in fact, are algorithmically-generated playlists of the songs that defined, united and distinguished the fans and love and new music in 2000+ genres and countries around the world in 2019:
2019 Around the World
You probably don't share my tastes, and this is a pretty weak unifying force for everybody who isn't me, but there are so many stronger ones. Maybe some of the ones that pull on you are represented here. Maybe some of the communities implied and channeled here have been unknowingly incomplete without you. Maybe you have not yet discovered half of the things you will eventually adore. Maybe this is how you find them.
I found a lot of things this year, myself, some of them in this process of trying to find music for other people, and some of them just by listening. You needn't care about what I like. And if for some reason you do, you can already find out what it is in unmanageable weekly detail. But I like to look back at my own years. Spotify's official forms of nostalgia so far define years purely by listening dates, but as a music geek of a particular sort, what I mean by a year is music that was both made and heard then. New music.
I no longer want to make this list by applying manual reductive retroactive impressions to what I remember of the year, but I also don't have to. Adapting my collective engines to the individual, then, here is the purely data-generated playlist of the new music to which I demonstrated the most actual listening attachment in 2019:
2019 Greatest Hits (for glenn mcdonald)
And for segmented nostalgia, because that's what kind of nostalgist I am, I also used genre metadata and a very small amount of manual tweaking to almost automatically produce three more specialized lists:
Bright Swords in the Void (Metal and metal-adjacent noises, from the floridly melodic to the stochastically apocalyptic.)
Gradient Dissent (Ambient, noise, epicore and other abstract geometries.)
Dancing With Tears (Pop, rock, hip hop and other sentimental forms.)
And finally, although surely this, if anything, will be of interest to absolutely nobody but me, I also used a combination of my own listening, broken down by genre, and the global 2019 genre lists, to produce a list of the songs I missed or intentionally avoided despite their being popular with the fans of my favorite genres.
2019 Greatest Misses (for glenn mcdonald)
I made versions of this Misses list in November and December, to see what I was in danger of missing before the year actually ended, so these songs are the reverse-evolutionary survivors of two generations of augmented remedial listening. But I played it again just now, and it still sounds basically great to me. I'm pretty sure I could spend the next year listening to nothing but songs I missed in 2019 despite trying to hear them all, and it would be just as great in sonic terms. There's something hypothetically comforting in that, at least until I starting trying to figure out what kind of global catastrophe I'm effectively imagining here. I'm alive, but all the musicians in the world are dead? Or there's no surviving technology for recording music, but somehow Spotify servers and the worldwide cell and wifi networks still work?
Easier to live. I now declare 2019 complete and archived. Onwards.
Collect people together, though, and the patterns of their love are sometimes very interesting. For several years I presided computationally over an assembly of nominal expertise, trying to find ways to turn hundreds of opinions into at least plural insights. Hundreds of people is not a lot, though, and asking people to pretend their opinions matter is a dubious way to find out what they really love. I'm not really sad we stopped doing that.
Hundreds of millions of people isn't that much, yet, but it's getting there, and asking people to spend their lives loving all the innumerable things they love is a more realistic proposition than getting them to make short numbered lists on annual deadlines. Finding an individual person who shares your exact taste, in the real world, is not only laborious to the point of preventative difficulty, but maybe not even reliably possible in theory. Finding groups of people in the virtual world who collectively approximate aspects of your taste is, due to the primitive current state of data-transparency, no easier for you.
But it has been my job, for the last few years, to try to figure out algorithmic ways to turn collective love and listening patterns into music insights and experiences. I work at Spotify, so I have extremely good information about what people like in Sweden and Norway, fairly decent information about most of the rest of Europe, the Americas and parts of Asia, and at least glimmers of insight about literally almost everywhere else on Earth. I don't know that much about you, but I know a little bit about a lot of people.
So now I make a lot of lists. Here, in fact, are algorithmically-generated playlists of the songs that defined, united and distinguished the fans and love and new music in 2000+ genres and countries around the world in 2019:
2019 Around the World
You probably don't share my tastes, and this is a pretty weak unifying force for everybody who isn't me, but there are so many stronger ones. Maybe some of the ones that pull on you are represented here. Maybe some of the communities implied and channeled here have been unknowingly incomplete without you. Maybe you have not yet discovered half of the things you will eventually adore. Maybe this is how you find them.
I found a lot of things this year, myself, some of them in this process of trying to find music for other people, and some of them just by listening. You needn't care about what I like. And if for some reason you do, you can already find out what it is in unmanageable weekly detail. But I like to look back at my own years. Spotify's official forms of nostalgia so far define years purely by listening dates, but as a music geek of a particular sort, what I mean by a year is music that was both made and heard then. New music.
I no longer want to make this list by applying manual reductive retroactive impressions to what I remember of the year, but I also don't have to. Adapting my collective engines to the individual, then, here is the purely data-generated playlist of the new music to which I demonstrated the most actual listening attachment in 2019:
2019 Greatest Hits (for glenn mcdonald)
And for segmented nostalgia, because that's what kind of nostalgist I am, I also used genre metadata and a very small amount of manual tweaking to almost automatically produce three more specialized lists:
Bright Swords in the Void (Metal and metal-adjacent noises, from the floridly melodic to the stochastically apocalyptic.)
Gradient Dissent (Ambient, noise, epicore and other abstract geometries.)
Dancing With Tears (Pop, rock, hip hop and other sentimental forms.)
And finally, although surely this, if anything, will be of interest to absolutely nobody but me, I also used a combination of my own listening, broken down by genre, and the global 2019 genre lists, to produce a list of the songs I missed or intentionally avoided despite their being popular with the fans of my favorite genres.
2019 Greatest Misses (for glenn mcdonald)
I made versions of this Misses list in November and December, to see what I was in danger of missing before the year actually ended, so these songs are the reverse-evolutionary survivors of two generations of augmented remedial listening. But I played it again just now, and it still sounds basically great to me. I'm pretty sure I could spend the next year listening to nothing but songs I missed in 2019 despite trying to hear them all, and it would be just as great in sonic terms. There's something hypothetically comforting in that, at least until I starting trying to figure out what kind of global catastrophe I'm effectively imagining here. I'm alive, but all the musicians in the world are dead? Or there's no surviving technology for recording music, but somehow Spotify servers and the worldwide cell and wifi networks still work?
Easier to live. I now declare 2019 complete and archived. Onwards.
¶ Hyperspace House Concerts · 26 October 2018 listen/tech
One of the questions we've yet to exactly answer, about the new streaming-based music business, is how you get started in it. In the old business, you mostly got started by playing your music for people near you. The new one has the potential to be strictly better than this, it seems to me, both by giving you more power to reach the people around you even when you aren't on a stage in one of their bars, and by giving listeners the ability to effectively warp to your town to hear you without leaving theirs.
For locally-popular artists seeking even-wider audiences, at least, Every Place at Once is an experimental partial answer: an algorithmic semi-global explorer of the music distinctively popular in individual cities. That relies on a fair amount of listening signal to operate, though, and thus doesn't really answer the question about getting started. How do you get to be locally popular? How do you move from you playing your songs for friends to strangers listening to your songs of their own accord?
I had been mostly ignoring this problem, having tried pushing the thresholds of Every Place at Once lower with results that were more worse than better, but periodically some new potential computational approach occurs to me. And usually also doesn't work. But this week, actually, one of these failed to fail as conclusively. It turns out that even at very low listening levels (on the order of tens of listeners, not even hundreds), if most of a song's listeners are in a single place, there's a pretty good chance that there's a reason for that. And, usefully, "That's where the band is from" turns out to be the most common one.
So I made another thing. If artists with tens of fans are the scale where you might play house concerts, this thing is an attempt at algorithmic semi-global Hyperspace House Concerts.
You can see pretty quickly that it's at least sometimes working: if you're listening to Harvard or MIT a cappella groups, you're probably in Cambridge. Mjangles is a rapper born in Ghana and raised in the Bronx, but he's currently a sophomore at Harvard. And even the music that isn't from Cambridge sometimes turns out to have interesting local stories. Jocelyn Hagen is a choral composer from North Dakota, but the week the list above was generated, her piece "Moon Goddess" was being rehearsed by a Harvard choir for an imminent concert. I didn't find as clear an explanation for André Caplet's similarly-lovely choral piece, but it was once performed by the Radcliffe Choral Society, so perhaps it still lingers in the walls.
And if some of this signal turns out to be noise, maybe that's OK. Sometimes the music in a place is coming out of the open windows of a passing car. And if 10 cars pass you blasting the same song, now it's part of your city.
So poke around, listen, see what you can find. Start where you live, and then try some places where you don't. There are lists for 500+ cities around the world, automatically updated every week, and more will appear as listening allows. And your listening can be part of it, part of how music travels and how careers begin and how we all find out what we're like.
For locally-popular artists seeking even-wider audiences, at least, Every Place at Once is an experimental partial answer: an algorithmic semi-global explorer of the music distinctively popular in individual cities. That relies on a fair amount of listening signal to operate, though, and thus doesn't really answer the question about getting started. How do you get to be locally popular? How do you move from you playing your songs for friends to strangers listening to your songs of their own accord?
I had been mostly ignoring this problem, having tried pushing the thresholds of Every Place at Once lower with results that were more worse than better, but periodically some new potential computational approach occurs to me. And usually also doesn't work. But this week, actually, one of these failed to fail as conclusively. It turns out that even at very low listening levels (on the order of tens of listeners, not even hundreds), if most of a song's listeners are in a single place, there's a pretty good chance that there's a reason for that. And, usefully, "That's where the band is from" turns out to be the most common one.
So I made another thing. If artists with tens of fans are the scale where you might play house concerts, this thing is an attempt at algorithmic semi-global Hyperspace House Concerts.
You can see pretty quickly that it's at least sometimes working: if you're listening to Harvard or MIT a cappella groups, you're probably in Cambridge. Mjangles is a rapper born in Ghana and raised in the Bronx, but he's currently a sophomore at Harvard. And even the music that isn't from Cambridge sometimes turns out to have interesting local stories. Jocelyn Hagen is a choral composer from North Dakota, but the week the list above was generated, her piece "Moon Goddess" was being rehearsed by a Harvard choir for an imminent concert. I didn't find as clear an explanation for André Caplet's similarly-lovely choral piece, but it was once performed by the Radcliffe Choral Society, so perhaps it still lingers in the walls.
And if some of this signal turns out to be noise, maybe that's OK. Sometimes the music in a place is coming out of the open windows of a passing car. And if 10 cars pass you blasting the same song, now it's part of your city.
So poke around, listen, see what you can find. Start where you live, and then try some places where you don't. There are lists for 500+ cities around the world, automatically updated every week, and more will appear as listening allows. And your listening can be part of it, part of how music travels and how careers begin and how we all find out what we're like.
¶ Music Demographics, Accountability and Short Blasts of Finnish Metal · 27 April 2018 listen/tech
On 27 April 2018 I gave a talk at the 2018 Pop Conference called "Panic, Death and Other Gender Patterns in Spotify Listening".
The talk itself involved enough snapshots of changing data and short blasts of music that I am not going to attempt to transpose it directly to blog form, but the main point of it was to introduce some new things I have added to everynoise.com having to do with music and gender and numbers and genres and accountability and hope.
Demographics and Listening
The first of these things is an interactive index to the distinctive musical tastes of Spotify listeners divided by country/age/gender demographic groups. It is here:
Every Demographic at Once
This is similar in structure to Every Place at Once, which explores listening by city, and I retrofitted some new features from the demographic version to the geographic one, so both now try to describe each group's tastes in both genre and song terms.
The demographic groups include, where data allows, country/age groups who self-identify as nonbinary, which Spotify began offering as a gender choice in September 2016.
There's even a playlist of the music most distinctively popular among nonbinary listeners as a global group, which I played some songs from in the talk, and which you can hear here:
Streamshare
The second thing I discussed is an automatically-updating dataview breaking down the current share of Spotify streams in each country by artist-gender. That's here:
Female-Artist Streamshare by Country
It has data for each country overall, but also separately for "direct" listening (people going directly to album or artist pages and explicitly playing something), for Spotify's algorithmic personalized Discover Weekly playlists, and for Spotify's featured editorial playlists.
The initial state of this data was about like this:
- Per-country female-artist streamshare, by the most optimistic definition of "female artist" I could construct from my existing data, which is "music whose performers are not exclusively male", varies from about 16-33%.
- You might hope, or even expect, that listeners might seek out more female artists on their own than when they are just listening to other people's playlists or editorial programming, but in fact it's currently the reverse, and female streamshare of direct listening is consistently lower.
- Discover Weekly has notably lower female-artist streamshare than overall or direct listening. This is not very surprising, because DW is powered by co-reference in Spotify playlists, and globally male listeners make more and longer playlists than female listeners on average, so this is a pretty textbook example of algorithmic confirmation bias due to inherent asymmetries in the data inputs.
- Spotify editorial programming, on the other hand, actually has a significantly higher percentage of female-artist streams in general and in almost every country. (Which is an intentional state brought about and maintained by specific Spotify editorial effort.)
Genre Patterns
The third thing is a similar automatically-updating dataview of genres and their shares of both female listeners and streams from female or mixed-gender artists.
Gender Listening Patterns by Genre
There is a wide array of interesting individual differences across the genre spectrum, but the discouragingly unsurprising overall insight is that male listeners stream less from female artists than female listeners do, and in general even female listeners don't stream equally.
In an attempt to see what I could do to push against these inclinations, I took this genre and gender data and produced an experimental set of playlists that try to collect music by female or mixed-gender artists in every individual specific microgenre. The "genre" column in this table links to these, and initially about 1000 genres had enough artist-gender and genre-listening data to support playlists.
Obviously this is not the first time anybody has tried to make female-artist genre playlists, but to me these efforts too often use only a very reductive high-level notion of genre, like "Women in Pop" and "Women in Rock", and thus seem inherently tokenistic. This also tends to implicitly characterize the presence of women in music as a novelty and a separate subject from music itself. The converse ideal, I think, is for it to be true and well-understood that there are women making every kind of music, and for it to be possible to find and listen to filtered subsets of genre music by female or mixed-gender artists without this seeming token or novel. At best these lists would be amazing just because music made by humans is consistently and profoundly amazing.
In many genres, as you can trivially discover by exploring these lists, we are nowhere near this gender-balanced ideal. In some cases my robots couldn't find any plausible intersections between genre and gender at all, and in some of the playlists they did generate, the artists may be "mixed gender" on dubious technicalities, or the music may be more culturally adjacent to the genre than in it.
But the good news about this set of battles in the fight to undo the miserable legacies of chauvinism and patriarchy is that at worst you'll hear some additional music by men, which is sometime surprisingly decent, or you'll end up listening to some music that isn't exactly what you asked for, which is also not a terrible idea for life.
[The playlist I used for example snippets during the talk will make very little explanatory sense on its own, but it has music, and so I include it here for posterity, in case posterity runs out of music of its own.]
[5 June 2018: For anybody who wants a concise tl;dr instead of these detailed breakdowns, I added a summary.]
The talk itself involved enough snapshots of changing data and short blasts of music that I am not going to attempt to transpose it directly to blog form, but the main point of it was to introduce some new things I have added to everynoise.com having to do with music and gender and numbers and genres and accountability and hope.
Demographics and Listening
The first of these things is an interactive index to the distinctive musical tastes of Spotify listeners divided by country/age/gender demographic groups. It is here:
Every Demographic at Once
This is similar in structure to Every Place at Once, which explores listening by city, and I retrofitted some new features from the demographic version to the geographic one, so both now try to describe each group's tastes in both genre and song terms.
The demographic groups include, where data allows, country/age groups who self-identify as nonbinary, which Spotify began offering as a gender choice in September 2016.
There's even a playlist of the music most distinctively popular among nonbinary listeners as a global group, which I played some songs from in the talk, and which you can hear here:
Streamshare
The second thing I discussed is an automatically-updating dataview breaking down the current share of Spotify streams in each country by artist-gender. That's here:
Female-Artist Streamshare by Country
It has data for each country overall, but also separately for "direct" listening (people going directly to album or artist pages and explicitly playing something), for Spotify's algorithmic personalized Discover Weekly playlists, and for Spotify's featured editorial playlists.
The initial state of this data was about like this:
- Per-country female-artist streamshare, by the most optimistic definition of "female artist" I could construct from my existing data, which is "music whose performers are not exclusively male", varies from about 16-33%.
- You might hope, or even expect, that listeners might seek out more female artists on their own than when they are just listening to other people's playlists or editorial programming, but in fact it's currently the reverse, and female streamshare of direct listening is consistently lower.
- Discover Weekly has notably lower female-artist streamshare than overall or direct listening. This is not very surprising, because DW is powered by co-reference in Spotify playlists, and globally male listeners make more and longer playlists than female listeners on average, so this is a pretty textbook example of algorithmic confirmation bias due to inherent asymmetries in the data inputs.
- Spotify editorial programming, on the other hand, actually has a significantly higher percentage of female-artist streams in general and in almost every country. (Which is an intentional state brought about and maintained by specific Spotify editorial effort.)
Genre Patterns
The third thing is a similar automatically-updating dataview of genres and their shares of both female listeners and streams from female or mixed-gender artists.
Gender Listening Patterns by Genre
There is a wide array of interesting individual differences across the genre spectrum, but the discouragingly unsurprising overall insight is that male listeners stream less from female artists than female listeners do, and in general even female listeners don't stream equally.
In an attempt to see what I could do to push against these inclinations, I took this genre and gender data and produced an experimental set of playlists that try to collect music by female or mixed-gender artists in every individual specific microgenre. The "genre" column in this table links to these, and initially about 1000 genres had enough artist-gender and genre-listening data to support playlists.
Obviously this is not the first time anybody has tried to make female-artist genre playlists, but to me these efforts too often use only a very reductive high-level notion of genre, like "Women in Pop" and "Women in Rock", and thus seem inherently tokenistic. This also tends to implicitly characterize the presence of women in music as a novelty and a separate subject from music itself. The converse ideal, I think, is for it to be true and well-understood that there are women making every kind of music, and for it to be possible to find and listen to filtered subsets of genre music by female or mixed-gender artists without this seeming token or novel. At best these lists would be amazing just because music made by humans is consistently and profoundly amazing.
In many genres, as you can trivially discover by exploring these lists, we are nowhere near this gender-balanced ideal. In some cases my robots couldn't find any plausible intersections between genre and gender at all, and in some of the playlists they did generate, the artists may be "mixed gender" on dubious technicalities, or the music may be more culturally adjacent to the genre than in it.
But the good news about this set of battles in the fight to undo the miserable legacies of chauvinism and patriarchy is that at worst you'll hear some additional music by men, which is sometime surprisingly decent, or you'll end up listening to some music that isn't exactly what you asked for, which is also not a terrible idea for life.
[The playlist I used for example snippets during the talk will make very little explanatory sense on its own, but it has music, and so I include it here for posterity, in case posterity runs out of music of its own.]
[5 June 2018: For anybody who wants a concise tl;dr instead of these detailed breakdowns, I added a summary.]