Tag: open source

VP of Technology and Platforms Tyson Singer Shares How Developers Can Solve Complexity With Backstage

As Spotify’s VP of Technology and Platforms, Tyson Singer keeps a watchful and anticipatory eye on the company’s tech infrastructure. He focuses most on ensuring that our platform is always evolving behind the scenes—while still working for our users. He also makes sure that we’re at the forefront of tech innovation through effective long-term investments in areas like open source and sustainability. No small feats. 

His team’s most recent success is commercializing Backstage, a developer portal that Spotify created internally and then open-sourced in 2020. “Open-source software is code that is designed to be publicly accessible—anyone can see, modify, and distribute it as they see fit,” Tyson explained to For the Record. “Opening up Backstage to the open-source community enabled external contributions that kept improving the tool, and the wide array of viewpoints made it an even more diverse platform for us and everyone using it.”

What does Backstage actually do? What is the problem it solves for enterprises? 

Backstage solves complexity—the kind of everyday complexity that can really bog engineers and their teams down, which then slows your whole organization down. Developers have access to more technologies than ever before, which comes with more responsibilities than ever before. Whether working at small or large companies, engineers use countless systems that all come with their own interfaces and ways of doing things. 

A Backstage developer portal puts everything in one place and is customized to whatever tools a company and individual developer team are using. So instead of switching between all these different tools and dashboards and systems, there’s just one front end for all of it — a single pane of glass. This unlocks speed, improved collaboration, or even just a smoother day. 

How did a company like Spotify come to create a tool for developers, and how has it evolved over time?

Years ago when Spotify was just starting to grow into the platform it is today, the software on the back end was becoming increasingly fragmented and complex. We needed to find a way to simplify the messy ecosystem and make it easier for developers to focus on what they do best: creating. It is central to our philosophy that happy developers make happy code.  

As I mentioned, we went on to open source it because it is fundamental to our entire platform, so we were incentivized to make the best product possible and make it the industry standard. It took off as an open-source tool and currently has over 700 adopters from companies like Netflix, Peloton, American Airlines, and more. 

In December, we took our first step into commercial software by selling a bundle of plugins to enhance the open-source version of Backstage. If you think of Backstage like your phone, then the plugins are like the apps—they are what make your phone more valuable and useful. We’re really excited about generating revenue that allows us to continue to invest in and support our open-source work. 

What are some innovative strides Spotify is making in technology and how does that align with our overall mission?

Spotify is obviously best known as the world’s largest audio-streaming platform, but that last part is often overlooked: platform. Our tech platform powers over 500 million creators with cutting-edge technology at a scale that is constantly growing. That means our engineers are building more, faster, while also making sure our platform continues to run smoothly. A core part of our company mission is to unlock human potential and creativity. Developers are creators—some of the most prolific creators, given that every company is now a technology company, and we’re really excited about the technologies we are building that empower them to do their best work. We plan to share more of those technologies so developers outside of Spotify can achieve the same experience. 

And finally, what’s on your “recently played”?

Mostly I listen to podcasts. One of my long-time favorites is Invest Like the Best with Patrick O’Shaughnessy. I’m not a professional investor, but his guests often have very forward-looking and insightful views on technology and business. And I love his closing question, “What’s the kindest thing anyone has done for you?” as it reminds us and these successful folks that kindness and luck are a big part of success.  

Music-wise, my “recently played” is completely unpredictable, even to me. I thought when I checked I’d see Soundgarden, The Beatles, or Queen show up, but it was actually Norah Jones, Buena Vista Social Club, and Gipsy Kings. That’s the beauty of Spotify for me, there’s always great recommendations for every mood.  

Build the best developer portal possible with Backstage.

Rachel Bittner on Basic Pitch: An Open Source Tool for Musicians

orange open source and coding symbols on a blue, green, and white background

Music creation has never been as accessible as it is now. Gone are the days of classical composers, sheet music, and prohibitively expensive studio time when only trained, bankrolled musicians had the opportunity to transcribe notes onto a page. As technology has changed, so too has the art of music creation—and today it is easier than ever for experts and novices alike to compose, produce, and distribute music. 

Now, musicians use a computer-based digital standard called MIDI (pronounced “MID-ee”). MIDI acts like sheet music for computers, describing which notes are played and when—in a format that’s easy to edit. But creating music from scratch, even using MIDI, can still be very tedious. If you play piano and have a MIDI keyboard, you can create MIDI by playing. But if you don’t, you must create it manually: note by note, click by click. 

To help solve this problem, Spotify’s machine learning experts trained a neural network to predict MIDI note events when given audio input. The network is packaged in a tool called Basic Pitch, which we just released as an open source project

“Basic Pitch makes it easier for musicians to create MIDI from acoustic instruments—for example, by singing their ideas,” says Rachel Bittner, a research manager at Spotify who is focused on applied machine learning on audio. “It can also give musicians a quick ‘starting point’ transcription instead of having to write down everything manually, saving them time and resources. Basically, it allows musicians to compose on the instrument they want to compose on. They can jam on their ukulele, record it on their phone, then use Basic Pitch to turn that recording into MIDI. So we’ve made MIDI, this standard that’s been around for decades, more accessible to more creators. We hope this saves them time and effort while also allowing them to be more expressive and spontaneous.”

For the Record asked Rachel to tell us more about the thinking and development that go into Basic Pitch and other machine learning efforts, and how the team decided to open up the tool for anyone to access and to innovate on.

Help us understand the basics. How are machine learning models being applied to audio?

Rachel Bittner

On the audio ML (machine learning) teams at Spotify, we build neural networks—like the ones that are used to recognize images or understand language—but ours are designed specifically for audio. Similar to how you ask your voice assistant to identify the words you’re saying and also make sense of the meaning behind those words, we’re using neural networks to understand and process audio in music and podcasts. This work combines our ML research and practices with domain knowledge about audio—understanding the fundamentals of how music works, like pitch, tone, tempo, the frequencies of different instruments, and more.

What are some examples of machine learning projects you’re working on that align with our mission to give “a million creators the opportunity to live off their art”?

Spotify enables creators to reach listeners and listeners to discover new creators. A lot of our work helps with this in indirect ways—for example, identifying tracks that might go well together on a playlist because they share similar sonic qualities like instrumentation or recording style. Maybe one track is already a listener’s favorite and the other one is something new they might like.

We also build tools that help creative artists actually create. Some of our tech is in Soundtrap, Spotify’s digital audio workstation (DAW), which is used to produce music and podcasts. It’s like having a complete studio online. And then there’s Basic Pitch, which is a stand-alone tool for converting audio into MIDI that we just released as an open source project. We open sourced Basic Pitch and built an online demo, so anyone can use it to translate musical notes in a recording (including voice, guitar, or piano).

Unlike similar ML models, Basic Pitch is not only versatile and accurate at doing this, but it’s also fast and computationally lightweight. So the musician doesn’t have to sit around forever waiting for their recording to process. And on the technological and environmental side, it uses way less energy—we’re talking orders of magnitude less—compared to other ML models. We named the project Basic Pitch because it can also detect pitch bends in the notes, which is a particularly tricky problem for this kind of model. But also because the model itself is so lightweight and fast.

What else makes Basic Pitch a unique machine learning project for Spotify?

I mentioned before how computationally lightweight it is—that’s a good thing. In my opinion, the ML industry tends to overlook the environmental and energy impact of their models. Usually with ML models like this—whether it’s for processing images, audio, or text—you throw as much processing power as you can at the problem as the default method for reaching some level of accuracy. But from the beginning, we had a different approach in mind: We wanted to see if we could build a model that was both accurate and efficient, and if you have that mindset from the start, it changes the technical decisions you make in how you build the model. Not only is our model as accurate as (or even more accurate than) similar models, but since it’s lightweight, it’s also faster, which is better for the user, too. 

What’s the benefit of open sourcing this tool?

It gives more people access to it since anyone with a web browser can use the online demo. Plus, we believe the external contributions from the open source community help it evolve as software to create a better, more useful product for everyone. For example, while we believe Basic Pitch solves an important problem, the quality of the MIDI that our system (and others’) produces is still far from human-level accuracy. By making it available to creators and developers, we can use our individual knowledge and experience with the product to continue to improve that quality. 

What’s next for Basic Pitch in this area?

There’s so much potential for what we can do with this technology in the future. For example, Basic Pitch could eventually be integrated into a real-time system, allowing a live performance to be automatically accompanied by other MIDI instruments that “react” to what the performer is doing.

Additionally, we shared an early version of Basic Pitch with Bad Snacks, an artist-producer who has a YouTube channel where she shares production tips with other musicians. She’s been playing around with Basic Pitch, and we’ve already made improvements to it based on her feedback, fixing how the online demo handles MIDI tempo, and other things to make it work better for a musician’s workflow. We partnered with her to use Basic Pitch to create an original composition, which she released as a single on Spotify. She even posted a behind-the-scenes video on her channel showing how she used Basic Pitch to create the track. The violin solo section is particularly cool.

But it’s not just artists and creators that we’re excited about. We’re equally looking forward to seeing what everyone in the open-source developer community has been doing with it. We expect to discover many areas for improvement, along with new possibilities for how it could be used. We’re proud of the research that went into Basic Pitch and we’re happy to show it off. We’ll be even happier if musicians start using it as part of their creative workflows. Share your compositions with us!

Create a cool track using Basic Pitch? Share it on Twitter with the hashtag #basicpitch and tag the team @SpotifyEng.