General Knowledge (97)
What Netflix buying Warner Bros means for music and for sync licensing
Written by Sounds SpaceWhat Netflix buying Warner Bros means for music and for sync licensing
On December 5, 2025 Netflix announced it will acquire Warner Bros.’ studio and streaming assets in a blockbuster deal that media outlets put in the $72–83 billion range — a move that instantly reshapes the global entertainment map.
For anyone whose work touches music — artists, labels, publishers, composers, music supervisors, and sync houses — this is a huge moment. The new combined company will control not only enormous film and TV IP (Harry Potter, DC, Game of Thrones, HBO’s prestige catalogue) but also the output pipelines and distribution muscle of one of the world’s largest streamers. That combination changes bargaining power, catalogue strategy, and the architecture of sync licensing. Below, I break down the likely short-, medium- and long-term effects, plus practical moves music-rights holders should consider.
Immediate realities: what actually changed and what hasn’t
First: the transaction creates a vertically integrated content powerhouse. Netflix gains rights to decades of Warner Bros. library and HBO programming and will control how that content is distributed on a global streaming platform with massive data and user reach. The deal is subject to regulatory approval and is expected to take many months — but the intention and market signals are clear.
Second: ownership of film/TV IP does not automatically transfer music rights. Songs in older shows and films often have complex split ownership — record labels, music publishers, composers, and sometimes third parties own different pieces (master vs composition). Netflix owning a film studio makes it a much larger licensor of sync placements (it controls the media where music appears), but it doesn’t mean Netflix suddenly owns every song in every scene. Still, control of future scoring, soundtrack decisions, and new franchise exploitation becomes far easier for Netflix internally.
Why this matters for sync licensing and big-picture mechanics
Sync licensing sits at the intersection of content owners (studios, streamers) and music rights holders (publishers, labels, composers). Historically, studios license music from publishers and labels for films/TV; in turn, studios may bundle soundtrack exploitation, trailers, ads, games and theme-park uses into separate negotiations. When the same company controls both content production and the global streaming pipe, a few structural shifts follow:
-
Increased leverage over upstream terms. A dominant studio+streamer can internalize more of the value chain — meaning it can prefer in-house composers, commission bespoke tracks under work-for-hire, or negotiate catalogue licenses with broad, global scopes (longer terms, extended media, etc.). That bargaining position pressures publishers and labels to accept either larger one-off fees or buyouts, or to secure better deal protections.
-
Greater possibility of “buyout” models. Netflix already experimented with buying-out certain music rights for global use, and large-scale ownership of studio IP incentivizes wider use of buyouts for global sync clearance — particularly in TV series where Netflix’s economics favor owning perpetual streaming rights rather than repeated per-territory renewals. This can be a double-edged sword: predictable income for some creators but a reduction in recurring backend streams/licensing revenue for others.
-
Data-driven placement and catalogue recycling. Netflix’s user data and recommendation engine could create more targeted uses of songs — resurfacing older tracks into playlists, promos, trailers and algorithmic placements that drive streaming spikes. That’s huge for catalogue owners who can get renewed streaming revenue and downstream sync fees. Conversely, it means Netflix could prioritize cheaper internal options when data suggests a track’s audience lift would be negligible.
-
Bundling across franchises and formats. With control of theatrical, streaming, and merchandising pipelines, Netflix can package sync uses across movies, shows, trailers, games and theme-park experiences. That consolidation makes “one-stop” licensing attractive for Netflix and complicates negotiation tactics for rights holders who want to keep leverage across different media.
(Those are not hypothetical: industry analyses on how studios convert ownership into licensing leverage have been circulating since the acquisition talks began).
Concrete short-term effects (0–18 months)
-
Negotiations will harden. Publishers and labels will quickly test Netflix’s appetite for broad, long-term licenses vs narrower deals. Expect stiffer offers and more insistence on exclusivity or bundled rights for tentpole franchises.
-
Composers may see more staff/composer-in-house opportunities. Netflix already invests in original scoring; studio ownership boosts demand for franchise continuity and in-house scoring teams. That can be good for steady work but may pressure freelance composers to accept different terms (e.g., buyouts, non-recoupable fees).
-
Sync houses and music supervisors become more strategic partners. Supervisors who can provide tailored catalogue solutions or bespoke tracks will be in demand — but they’ll need to be nimble around Netflix’s preferred rights scopes and reporting/data formats.
Medium-term structural shifts (18 months–5 years)
-
Consolidation of licensing platforms and metadata standards. As Netflix scales its internal licensing and potentially licenses its own catalogue to third parties, there will be pressure to standardize metadata, split sheets, and payment reporting — a space already seeing startups and services modernising the sync market. That can reduce friction (good) but also enable faster, lower-cost internal clearances (which could reduce fees for some licensors).
-
More catalogue re-packaging and remastering. Old tracks tied to Warner films/TV can be repurposed into new formats and playlists, creating renewed streaming and sync value. Publishers that move fast to re-negotiate or clear stems and alternate masters will profit.
-
Regulatory and marketplace pushback. Antitrust scrutiny is real; regulators may impose remedies (divestitures, non-exclusive licensing mandates, behavioural remedies) that could blunt Netflix’s ability to monopolize certain licensing windows. This will affect how exclusive or non-exclusive deals get structured.
Risks — who loses and how
-
Smaller publishers and independent composers risk being squeezed by a giant licensor that can prioritize internal or cheaper catalogues. If Netflix standardizes buyouts for large shows, the long-tail income that small-rights owners count on could shrink.
-
Transparency & backend royalty issues. Large-scale internal use raises questions about reporting fidelity. Ensuring accurate use reporting, splits and divisor calculations is critical — missing or opaque reporting can cost creators dearly.
-
Market concentration undermining bargaining power. If other studios follow suit with vertical integrations, collective bargaining power for rights holders could be weakened, pushing rates down.
Opportunities — who can win and how
-
Catalogue owners who act fast. Publishers that proactively repitch their catalogues for franchises, create stems and alternative masters, and build sync-friendly metadata will catch the algorithmic and editorial attention of Netflix’s content teams.
-
Artists who own masters/compositions. Creator-owned masters and publishing provide the best negotiation leverage; artists with their rights intact can demand better terms or carve out higher-value sync deals.
-
Tech-enabled licensing platforms. Companies that can offer rapid, auditable, global licensing (with granular usage reporting) will be valuable partners — both to Netflix (which wants efficiency) and to rights holders (who want transparency). That market was already evolving pre-deal, and this acquisition accelerates its importance.
-
Music supervisors & bespoke composers. With more original series and films to score, premium supervision and tailored compositions will remain necessary — especially for high-profile franchises where bespoke music is a differentiator.
Practical playbook for rights holders (10 action points)
-
Audit rights now. Know exactly which compositions and masters you control, for what territories and media.
-
Clean your metadata. Improve ISRCs, splits, writer/publisher info — Netflix-scale buyers want neat data.
-
Create stems and alternate masters. These increase the chance a track gets reused (trailers, promos, games).
-
Consider selective exclusivity. For high-value placements, negotiate rolling exclusives or premium windows rather than blanket buyouts.
-
Build reporting safeguards into contracts. Define audit rights, payment cadence, and data formats.
-
Leverage boutique sync firms. They can package your catalogue for franchises and understand Netflix-style contracts.
-
Protect composer revenues. Avoid one-time buyouts when possible; insist on backend/royalty participation for major franchises.
-
Explore co-marketing deals. Tie-in playlisting, social activations, or soundtrack releases can amplify streaming income.
-
Watch regulatory updates. Any antitrust remedies could create windows of opportunity for third-party licenses.
-
Invest in IP ownership. If you’re an artist, control your masters and publishing — it’s the most direct hedge against market consolidation.
Final thoughts: an industry in flux, but not hopeless
This deal is systemic: it changes incentives, packing power into companies that can own IP, distribution and audience data at scale. For music rights holders that means both threat and chance. The big risk is commoditization — blanket buyouts, less recurring income, and harder negotiations for smaller players. The big opportunity is visibility and reuse: a single placement on a Netflix-distributed tentpole can still send an artist’s streams and sync demand skyrocketing.
Regulators will shape how far Netflix can push exclusivity and vertical control, so the landscape will keep shifting over the next 12–24 months. In the meantime, the music world’s best defence is straightforward: clean metadata, controlled rights, flexible licensing strategies, and partnerships with supervisors and platforms that understand the new rules.
Spotify’s Big Video Push: How the Streaming Giant Is Transforming Into a Hybrid Music-Video Platform in 2025
Written by Sounds SpaceSpotify’s Big Video Push: How the Streaming Giant Is Transforming Into a Hybrid Music-Video Platform in 2025
For over a decade, Spotify has defined itself as the audio-first platform — a place where music lived, playlists ruled, and podcasts became the new frontier. But in 2025, everything is changing. Spotify is no longer content being “just” a music app. It’s now pushing hard into video, integrating music videos, visual content, and even short-form features directly into its core experience.
This shift — from audio-only streaming to a hybrid audio-video ecosystem — is one of the biggest transformations in Spotify’s history. And it’s raising important questions:
What does the “music video” mean in 2025?
How will this move affect artists?
Will user behaviour change?
And what does this mean for the future monetisation of streaming?
Let’s break it all down.
Why Spotify Is Making a Video Push Now
For years, Spotify has been the king of audio streaming… but audio alone doesn’t dominate the cultural conversation anymore.
Music consumption today is shaped by TikTok, YouTube Shorts, Instagram Reels, and visual storytelling. Artists gain momentum and virality not from traditional music videos, but from high-impact visual clips, fan edits, behind-the-scenes snippets, and performance content.
In short:
✔ The world now consumes music visually
✔ Platforms built on video dominate attention
✔ Music discovery happens through images as much as sound
Spotify knows this — and knows that if it doesn’t evolve, it risks becoming a background app rather than the centre of culture.
So in 2025, Spotify is pivoting:
-
Adding full music videos
-
Integrating short-form vertical clips
-
Introducing video playlists
-
Expanding video podcasts
-
Offering artists visual content slots inside their track pages
-
Testing new video-based discovery surfaces
Spotify doesn’t want users to leave the app to watch visuals anymore — whether it’s on TikTok, YouTube, or anywhere else.
It wants to become a one-stop shop for music and visuals.
The New Meaning of “Music Video” in 2025
The traditional music video used to be a big-budget, cinematic centerpiece released on MTV or YouTube to promote a single. But in recent years, its cultural power has faded — replaced by quick, viral, attention-grabbing clips.
Spotify stepping into video changes the game again.
Music videos are no longer just promotional tools — they become part of the streaming experience itself.
This has major implications:
1. Music videos become more integrated and interactive
A music video on Spotify isn’t a separate destination like YouTube.
It becomes part of:
-
the song page
-
the playlist
-
the artist’s hub
-
the discovery feed
Fans can watch, like, save, share, comment, and even sync video clips with audio playlists.
This makes music videos a functional part of listening — not something external.
2. Artists can release multiple types of videos per song
Instead of one expensive video, a track might have:
-
the official music video
-
live session versions
-
vertical edits
-
fan-clip compilations
-
behind-the-scenes
-
motion graphics loops
-
animated versions
Spotify’s visual infrastructure makes all of these seamlessly accessible.
3. Short-form visuals become the new “album artwork”
Cover art used to be the identity of a song.
Now?
Video loops, canvas clips, and vertical snippets take the lead.
A song might be defined by a 5-second clip — the part users see repeatedly as they stream.
Spotify’s expansion gives these micro-visuals a home and elevates their creative importance.
4. Visuals become a new form of branding
Artists no longer just “release music.”
They release experiences — sound + visuals packaged together.
In 2025, the music video isn’t dead.
It’s evolving.
How This Shift Affects Artists: Opportunities & Challenges
Spotify’s video push brings both powerful benefits and new pressures for artists.
Opportunities for Artists
1. More ways to express creativity
Instead of choosing between a $20,000 music video and nothing, artists can now deliver:
-
affordable visualizers
-
animated loops
-
vertical edits
-
lyric videos
-
photo slideshows
-
fan-generated visuals
All inside Spotify.
This lowers the barrier to visual creativity.
2. Better discovery potential
A visually striking video or clip could now become a major discovery tool.
Imagine:
A user plays a playlist → a new song appears with gripping visuals → they’re instantly hooked.
Spotify hasn’t had this kind of “visual discovery” power before.
This is TikTok’s biggest advantage — and now Spotify is closing the gap.
3. Stronger monetization down the line
Spotify’s video rollout sets the stage for future revenue options:
-
Video ads
-
Sponsored visuals
-
Paid exclusive video content
-
Video-based fan subscriptions
-
Virtual merch or interactive video items
-
Premium artist video hubs
Artists will be able to earn more not just through streams, but through hybrid content releases.
4. Greater control over fan engagement
Artists can create:
-
episodic content
-
behind-the-scenes diaries
-
short film tie-ins
-
dance challenges
-
story-driven visual arcs
All without relying on YouTube or TikTok algorithms.
Spotify essentially becomes a platform where artists can build deeper, more controlled fan ecosystems.
Challenges for Artists
1. More pressure to create visual content
Not every musician is a filmmaker.
Not every band has the resources to pump out videos.
This shift may create a new kind of competition:
Who can produce the best visuals?
Who has the team to execute consistently?
Who can afford regular video content?
Artists might feel forced to invest in video even if they don’t want to.
2. Budget imbalance
Music videos, even short ones, cost time and money.
Independent artists might struggle while major labels flood the platform with polished visuals.
The gap could widen.
3. Creative burnout
Platforms often demand constant output.
Spotify’s new visual surfaces may increase the expectation that artists “feed” the platform regularly with:
-
new clips
-
new edits
-
new video versions
For creators already overwhelmed, this could be exhausting.
How User Behaviour Will Change
Spotify adding video isn’t just a technical upgrade — it will fundamentally change how people use the app.
1. Spotify becomes a “lean-in” platform
Audio is passive.
Video is active.
Users will now pick up their phones more often, scroll more, watch more, engage more.
Spotify becomes a place where people watch as much as they listen.
2. Playlists evolve into video playlists
A workout playlist might become:
-
20 songs
-
each with energetic visualizers
A chill playlist might include:
-
calm animations
-
nature clips
-
ambient visuals
This transforms playlists into immersive experiences, not just collections of songs.
3. Music discovery becomes visual
A catchy visual loop can hook a listener in seconds.
Spotify knows this.
That’s why it’s leaning into video for discovery.
Soon, “discovering music” on Spotify will feel closer to browsing Reels or TikTok — but focused entirely on songs.
4. The app keeps users for longer
Video dramatically increases retention.
The more surfaces Spotify adds:
-
video feeds
-
video-based recommendations
-
artist video stories
-
top video charts
…the more time people spend inside the app.
This reduces the need to jump to:
-
YouTube for music videos
-
TikTok for viral clips
-
Instagram for behind-the-scenes content
Spotify becomes a unified hub.
The Future of Monetization: How Spotify’s Video Push Changes the Business
Spotify’s video strategy isn’t just a creative decision — it’s a financial one.
Here’s how this evolution impacts monetization.
1. New video ad formats
Advertisers love video because it:
-
grabs attention
-
boosts engagement
-
increases recall
Spotify can now introduce:
-
pre-roll video ads
-
mid-roll video ads
-
sponsored visual playlists
-
artist video sponsorships
This opens the door to huge new revenue streams.
2. Premium video content tiers
Spotify might begin offering:
-
paid video episodes
-
exclusive artist videos
-
visual albums
-
behind-the-scenes documentaries
-
special sessions or live performances
Users could pay extra for enhanced visual content.
3. Video-based fan monetization
Artists may soon be able to offer:
-
paid video diaries
-
exclusive monthly content
-
locked premium videos
-
virtual meet-and-greets via video
This mirrors Patreon — but inside Spotify.
4. Brand partnerships with integrated video
Brands might sponsor:
-
video songs
-
playlist videos
-
video-based events
-
artist video series
Suddenly, sponsorship becomes more dynamic and profitable.
What This Means for the Future of Music Streaming
Spotify’s move signals a broader shift:
Streaming platforms are no longer competing over music. They’re competing over attention.
In 2025, audio isn’t enough.
-
TikTok dominates musical virality
-
YouTube dominates music video culture
-
Instagram dominates artist storytelling
-
Twitch dominates live performances
-
Spotify dominates audio
But Spotify wants more.
It wants to sit at the centre of all music-related content.
This means:
✔ Streaming platforms will become hybrid ecosystems
✔ Artists will release songs + videos simultaneously
✔ Visual storytelling will become standard
✔ Fan engagement will deepen
✔ Monetisation will expand dramatically
We’re entering a new era where music experiences are not defined by sound alone — but by sound, visual identity, and the emotional world an artist creates through both.
Conclusion: Spotify’s Video Push Is Reshaping Music Culture
Spotify’s expansion into video is more than a platform update — it’s a cultural shift that will redefine how music is consumed, discovered, and monetized in 2025 and beyond.
For artists, it means new creative opportunities — and new pressures.
For fans, it means richer, more engaging music experiences.
For the industry, it signals that pure audio streaming is no longer enough.
We’re witnessing the birth of a hybrid world where music and visuals merge into one unified experience — and Spotify is positioning itself at the centre of that evolution.
If this works out the way Spotify envisions, the future of music streaming won’t be audio-first.
It will be fully audiovisual — immersive, interactive, and integrated in ways we’re only beginning to understand.
How AlphaTheta’s CDJ-3000X is finding an unexpected new audience
Written by Sounds SpaceHow AlphaTheta’s CDJ-3000X is finding an unexpected new audience
AlphaTheta has refined, not revolutionised, its flagship player — which could become ubiquitous not just in premier nightclubs, but your average home DJ setup
If you’ve spent any time near a club booth in the last decade, you know the look: aluminium chassis, jog wheel gleaming under dim LEDs, and a screen full of waveform lines and hot cues. For years, Pioneer DJ (now under the AlphaTheta umbrella) owned that aesthetic and the market; the CDJ series was the shorthand for “professional DJ player.” The new CDJ-3000X doesn’t rip that script up. Instead, it takes the trusted formula and polishes each corner until the whole thing feels modern in a way that matters — and that refinement is what’s nudging the 3000X out of pro booths and into a growing number of home setups.
Not a revolution, an essential evolution
There’s a temptation to expect every new product to be a manifesto of change. The 3000X is wiser: it’s evolutionary. Think bigger screen, better connectivity, smarter browsing, and small but meaningful workflow upgrades. That’s it. That’s the headline. But those “small” changes are what matter in day-to-day DJing — especially for DJs who aren’t tethered to festival rigs and mountain-of-cables setups. A capacitive 10.1-inch touchscreen replaces the older resistive panel and displays more tracks and metadata at once, which speeds up finding the right record mid-set. Built-in Wi-Fi and NFC sign-in streamline access to cloud libraries and streaming services. For a home DJ who wants to play professional-grade sets without lugging extra routers or worrying about Ethernet runs, this is a huge quality-of-life improvement.

The connectivity story: less cable, more convenience
AlphaTheta’s designers have clearly thought about how DJs actually move: late arrivals, short setup windows, and the necessity of jumping straight into a set. The 3000X has built-in Wi-Fi, a front-panel NFC pad for quick rekordbox login, and USB-C instead of the old SD slot. That all adds up to fewer annoying hardware rituals — no more scrambling for SD cards, no wrestling with Ethernet, and no slow login routines. For a club engineer, that’s a relief. For a home DJ, it means you can bring pro hardware into the living room and not feel like you’ve set up mission control. Reviews consistently point to this as one of the 3000X’s core advantages.
Why home DJs are paying attention
There are three main reasons the 3000X is starting to appear in home rigs:
-
Feature set that actually scales down — The improvements are as useful for club-level performance as they are for a producer practising at home. High-quality audio, responsive jog, professional I/O — these don’t become overkill in a bedroom; they just make practice feel real.
-
Ease of use for streaming and cloud libraries — DJs who built huge Rekordbox libraries now have seamless cloud options. Coupled with OneLibrary and cross-platform initiatives, DJs can hop between setups without recreating libraries — that’s particularly appealing for hobbyists who DJ on different machines or want to practice with the exact record pool they’ll use in a bar or festival. (OneLibrary’s cross-platform promise is a big ecosystem play that makes owning a CDJ feel future-proof.)
-
A status symbol and a learning tool — For some, owning a CDJ is aspirational: it’s the gear the pros use. For others, it’s a practical way to learn industry standards. The 3000X’s incremental upgrades mean home DJs can learn on the same UI, same workflow, same button feel as in the top booths. That parity reduces a friction point when moving from the bedroom to the bar.
What AlphaTheta kept, and why it matters
You won’t find radical new performance modes like multi-layer decks or built-in Stems processing on the 3000X. AlphaTheta doubled down on reliability — tougher play/cue buttons, a redesigned jog with refined feel, and improved audio circuitry. It’s the “if it ain’t broke, improve it” philosophy. For working DJs and club buyers, longevity and predictability outweigh gimmicks. For home buyers, that means the unit won’t suddenly become outdated when a software trend shifts. It’s a long-term investment in a stable, pro-level workflow.
The one potential friction: price and features
Let’s be honest: the CDJ-3000X is a premium product with premium pricing. The street figures sit comfortably above many standalone players and controllers, and there’s still no onboard SSD for massive local storage. AlphaTheta seems to be betting on cloud and connected workflows rather than stuffing more local storage into the chassis. For home DJs with smaller budgets, controllers from other brands still represent value; for players who want authentic club hardware and the feeling of a pro booth in their lounge, the 3000X fills that niche. Reviews have repeatedly noted that the model is the most polished CDJ yet — but not necessarily a must-have upgrade for every CDJ-3000 owner.
The ecosystem shift: OneLibrary and cross-platform freedom
One of the quiet game-changers here isn’t a physical button on the player — it’s the move toward a more open library standard. OneLibrary, which aims to let DJs carry cue points, beatgrids and playlists between rekordbox, Traktor, and djay Pro, reduces vendor lock-in and makes buying high-end hardware less risky. If your collection is portable between systems, owning a CDJ that plays nice with cloud libraries becomes logical even for a hobbyist — suddenly you aren’t buying into a one-brand lifecycle. That kind of ecosystem move encourages more people to invest in higher-end kit because the software and library headaches are eased.
Real talk: reliability and early hiccups
No product is perfect at launch; AlphaTheta’s firmware story this year shows how sensitive the community is to updates that affect library integrity. There have been reports of problematic firmware updates that caused missing tracks and playlists for some users, which is a serious wake-up call for anyone relying on USB sticks or freshly created library formats. It underlines one lesson: pro hardware depends as much on thoughtful software rollout as it does on chassis design. Home users should be cautious with firmware updates and keep backups of their libraries. AlphaTheta’s response to revert and investigate is an important part of maintaining trust.
Who should consider a 3000X for home use?
If you fit any of the following profiles, the 3000X makes strong sense:
-
You want a pro-grade practice environment that translates directly to club performance.
-
You stream often or want to integrate cloud streaming/Beatport/Tidal into your DJing workflow.
-
You care about long-term compatibility with industry standards and prefer the physical media feel over controllers.
-
You’re building a hybrid rig (controller + media player) and want the industry standard jog, audio path, and build quality.
If you’re a casual weekend mixer, teaching yourself to DJ, or on a tight budget, a high-quality controller or a second-hand media player might be the smarter choice — but for the committed bedroom jockey who wants the tactile, “real” booth feel, the 3000X is a compelling option.
What this means for clubs, schools, and the broader scene
Clubs will adopt the 3000X because it reduces setup friction and looks slick in the booth. DJ schools and tutors will want students to learn on the hardware that represents the standard. The interesting consequence? Those students, once they buy gear for home practice, will increasingly buy the same brand hardware. The CDJ, long a professional-only symbol, is slowly migrating into learning studios and living rooms — and that trickle could shift the baseline expectation of what “home DJ gear” looks like in five years. That’s how a refined product becomes ubiquitous.
Bottom line: ubiquity through refinement
The CDJ-3000X proves a design truth that’s easy to forget: ubiquity rarely arrives via bold reinvention — it arrives when a product becomes so well adapted to real-world workflows that it fits everywhere. AlphaTheta didn’t invent a new way to DJ with the 3000X. Instead, it made the everyday work better: faster logins, cleaner browsing, better screens, and more reliable buttons. Those changes may sound incremental, but they lower the barrier for home DJs to buy pro gear and for clubs to keep a consistent booth standard. The result is an unexpected audience — not because AlphaTheta chased home DJs, but because it made a pro tool that also happens to be brilliantly at home in the living room.
If you’re a bedroom DJ who’s been daydreaming about pro gear, the 3000X is a sensible, aspirational pick — provided you’re ready to invest and you respect the importance of firmware discipline and backups. For clubs and rental houses, it’s the kind of incremental polish that turns “good enough” into the new baseline.
And for the DJ community? It’s another step toward a future where the gap between the living room and the main room keeps getting smaller. That’s good for music, good for learning, and — frankly — good for anyone who loves the feeling of cueing up a record on a machine that feels like the heart of the scene.
Rage-bait? Reddit users stunned as Suno user complains about running out of prompt ideas
Written by Sounds SpaceThe post that lit the fuse
A user on r/SunoAI posted something blunt and frustratingly honest: they like what Suno makes, but they’ve hit a wall — they don’t know what to prompt the model with next. They even asked if Suno should add a “generate prompt” button so the tool could spit out new ideas for them. That simple ask — “write my prompts for me” — prompted waves of reactions: amusement, scorn, satire, and a surprisingly earnest debate about what creativity means when AI does the heavy lifting.
Music journalism picked up the thread and framed it as “rage-bait?” — a headline ready to go viral because it’s easy to make fun of someone for “outsourcing creativity.” But beneath the jokes and GIFs on Reddit there are real questions: is this a symptom of prompt fatigue, platform design failing users, or a new kind of creative dependency?
Why people reacted so strongly
-
Expectation vs. craft. Lots of redditors see coming up with prompts as the core intellectual work — the part that makes something uniquely yours. To them, asking an AI to dream up prompts feels like trying to get an assistant to think for you. That rubbed some folks the wrong way, because it looks like trading creative agency for convenience. The comments were brutal but revealing: some users suggested the poster “learn the craft,” others offered to post lists of starter prompts.
-
Saturation & sameness. Over time, people notice that the outputs of generative models can flatten into similar-sounding results. Several long threads in the Suno community discuss “getting the same bland results” after a number of generations. When the outputs feel repetitive, coming up with fresh prompts becomes harder — not because the person has no imagination, but because the model’s space of plausible outputs seems narrower. That fuels frustration and the “what’s the point?” takes.
-
Platform design matters. The OP asked for a “new prompt” button — and that’s a legit UX idea. If your product depends on users continually injecting new creative intent, you should provide ways to lower the barrier. Some redditors had already prototyped prompt-helper prompts (community-built meta-prompts) to generate better prompts for Suno — meaning users solved it themselves before the company did.
-
The performative layer. Social media loves a spectacle. Posts that invite outrage or mockery get traction fast. A simple admission of creative fatigue becomes a perfect target for jokes and hot takes. That’s the “rage-bait” angle: it’s cheap to rile up the crowd by implying laziness or entitlement, and people obliged to defend their craft respond accordingly.
Is the poster actually in the wrong?
Short answer: no — and also yes, depending on what your baseline is.
If you treat AI as a tool that extends your ideas, then needing help to reseed your imagination is totally normal. Creative work has cycles: bursts of inspiration followed by dry spells. Tools that speed up iteration can also remove the parts of the process that spark new directions (the “happy accidents” of manual experimentation). So asking for help — even from the very tool you use — is reasonable.
But if you treat the act of prompting as the creative spark itself, then leaning on the same tool to produce your prompts can feel circular: you’re outsourcing both idea-generation and execution. For critics on Reddit, that’s where the problem lies — they value the human input as the thing that gives an output meaning.
Bigger picture: prompt fatigue and generative tools
What the thread reveals is a broader phenomenon: prompt fatigue. As generative AI becomes more capable, the human role morphs into a new kind of craft: engineering prompts, curating outputs, and post-editing results. That craft can be rewarding, but it can also get exhausting. People who generate dozens of tracks or images per day hit a creative plateau — not because they’re uncreative, but because the interface (a text box) becomes the bottleneck for variety.
Communities have already started to respond: shared prompt libraries, “prompt-of-the-week” challenges, and meta-prompts — prompts that generate prompts — have cropped up on subreddits and Discord servers. Some users advocate for collaborative prompt-sharing, while others build little scripts and tools that randomize elements (genre + instrument + tempo + mood) to give the model something new to chew on.
Could Suno (or any platform) solve this?
Yes. There are direct, practical features a company like Suno could add to reduce friction and keep creators moving:
-
Prompt starter packs. Curated sets of starter prompts (by genre, mood, or production goal) to help users explore new directions.
-
Prompt generator toggle. A one-click “generate a new prompt idea” button that either uses a rule-based template (genre + instrument + hook) or an LLM to suggest prompts.
-
Seed mutation tools. Buttons to “mutate” an existing prompt — change tempo, swap instruments, or twist the mood automatically.
-
Community prompt marketplace. A place where creators can share and rate prompts, making it easier to discover high-quality seeds.
-
Integrated workflows with randomness. Allow users to run A/B batches with slight prompt variation, surfacing serendipitous winners instead of expecting each prompt to be a perfect hit.
The Reddit thread’s OP explicitly suggested the “generate prompt” idea, which is both a UX ask and a growth opportunity for companies. Some users have already created community workarounds: meta-prompts and prompt templates that you can paste into Suno. That suggests the demand exists, and that the community is willing to fill the gap if the company doesn’t.
The ethics & aesthetics of leaning on AI for ideas
This argument isn’t just UX: it’s philosophical. When we shift idea-generation to algorithms, what happens to artistic authorship? Is a song generated by Suno because a user typed “sad indie ballad about a lost bus pass” still art?
There’s a spectrum:
-
Tool-as-accelerant: The human brings the central idea; AI accelerates execution.
-
Tool-as-collaborator: The human and AI co-create; prompts and outputs are interdependent.
-
Tool-as-proxy: The human mostly curates; the AI supplies the spark and shape.
Each mode has aesthetic and ethical implications. Creators may feel devalued when the provenance of an idea becomes murky; listeners may feel disconnected if everything starts to sound like algorithmic furniture. That’s why community policing quality (and complaining when outputs get bland) matters: they’re defending aesthetics, not just gatekeeping.
Practical tips if you’re stuck on prompts (for Suno users)
Bro, if you’re the kind of person who hit a wall and wants to keep making stuff, try these immediate moves — ripped from community wisdom and prompt-hacker playbooks:
-
Genre mashups. Combine two very different genres (e.g., “80s synthwave + mariachi trumpet”) to force the model into unfamiliar territory.
-
Swap constraints. Pick an odd constraint: “write a breakup song using only metaphors about weather” — constraints breed creativity.
-
Seed from media. Use a movie scene, a painting, or a line of poetry as your prompt seed (not to copy but to inspire).
-
Mutate the mood. Take a happy track and regenerate with “mournful” or “aggressive” toggles.
-
Use meta-prompts. Feed Suno (or an LLM) a prompt like: “Give me 10 unique song prompt ideas that blend prog-rock and lofi hip-hop.” Paste the outputs back into Suno and iterate. (Communities already do this.)
-
Curate, don’t expect perfection. Generate 20 small variations and harvest the best 1–2 moments. Treat the model like your sound library generator.
Final thought: what this moment exposes about AI culture
That viral Reddit moment is funny because it’s relatable: creativity is hard. It’s provocative because it surfaces anxieties about over-reliance on tools. And it’s useful because it points to a clear product opportunity: make creativity-sparking features for people using generative tools at scale.
So is it rage-bait? Kinda — it makes for an easy meme. But it’s also a genuine note from a creator saying: “I’m stuck — can my tools help?” And that’s a question we should answer with empathy, not mockery. The funniest thing is that the community largely solved the issue already: shared prompts, templates, and meta-prompts are out there waiting. Suno and other platforms that listen could make this a native feature and turn a moment of mockery into a growth hack.
If you’re a Suno user and you’ve hit the wall, don’t be ashamed — be curious. Join a prompt swap, try a mutator script, or give your machine a machine to brainstorm with. Creativity loves constraints, and sometimes the best prompt is the weirdest one you haven’t tried yet.
Ableton Live 12.3 has arrived: Stem Separation, Splice integration and everything else you need to know
Written by Sounds SpaceAbleton Live 12.3 has arrived, and this one’s got a little bit of everything:
AI-powered stem separation, native Splice integration, better bounce workflows, notable Push 3 improvements, and a clutch of smaller but genuinely useful workflow refinements. The update is free to existing Live 12 users, and Ableton is running a limited-time 25% discount for newcomers and Pack purchases, so whether you’re already deep into Live or thinking about joining the ecosystem, now’s a good moment to take stock.
Below, I break down the features that matter most, what they’ll let you do differently in the studio or on stage, and a few practical tips for getting the most out of the new tools.
The big headline: stem separation built into Live Suite
The flagship addition in 12.3 is built-in stem separation for Live Suite. This isn’t a third-party plug-in or cloud-only service — Ableton has integrated local, offline stem splitting that separates an audio clip into four component stems: Vocals, Drums, Bass, and Others. Practically, that means you can take any sample, loop, or full stereo track and pull it apart inside Live, then remix, rearrange or resample the resulting parts without leaving your DAW.
Why this matters: Stem separation used to be a workflow that required external services or specialized tools that sent audio to the cloud. Having a fast, local option directly inside Live closes the loop: faster iteration, less context switching, and no file juggling. For remixers, producers who build stems for collaborators, or anyone doing sample-based creativity, the ability to isolate a vocal line or a drum bus without leaving your session is huge. Expect creative uses beyond remixing, to think creative sidechaining, subbing in new drums under a vocal stem, or extracting a texture from the “Others” stem to turn into an ambient pad.
A practical tip: stem separation quality varies with the source material. Clear, well-separated recordings (dry vocals, distinct drum hits) give the best results; heavily distorted or extremely dense mixes may produce artifacts. Use the stems as starting material: resample them, run them through effects racks, and don’t be afraid to combine stems back together after processing.
Splice integration: search, audition, and drop — inside Live
Ableton’s Splice integration is more than a link — it brings Splice’s sample library into Live’s Browser so you can search, audition in sync and key, then drag samples into your project without switching apps. The “Search with Sound” feature is particularly neat: you can capture audio from your set (or drag a clip into the Splice panel) and ask Splice to find samples that fit the rhythms and harmonic content of what you already have. That can turn a friction-filled sample hunt into a fast, creative playground.
Why this matters: searching for the right sample used to be a deep rabbit hole — dozens of browser tabs, trawling keyword searches, guessing about tempo and key. Native integration means auditioning is instantaneous and context-aware: the samples are previewed in time with your project, so you hear how they groove before committing. For fast sketching and late-night idea sessions, that’s a serious time-saver.
Bounce Groups and smarter offline workflows
Another workflow-focused upgrade is Bounce Groups: the ability to render an entire group (with processing) to a single audio file. This lets you commit CPU-heavy group chains to audio without losing the option to keep your original tracks for later edits — a clean balance between commit-and-mix efficiency and flexibility. It’s the kind of workflow improvement you don’t notice until you need it, then you wonder how you ever lived without it.
Other audio/bouncing improvements under the hood include faster, more reliable bounce operations and fixes for edge cases on different platforms. If you run large projects or play Live sets that depend on pre-rendered stems, these changes will smooth your workflow and (importantly) reduce last-minute rendering headaches.
Push 3 and hardware improvements
Live 12.3 isn’t just a software update: Push 3 gets form-and-function updates that expand what you can do in standalone and tethered modes. Notably, Push 3 in standalone mode can now work with class-compliant audio interfaces — meaning more ins/outs and a more flexible standalone setup without relying on ADAT tricks. Push’s expressive grid also gets new XY-style control modes and improved step-sequencing with touch-sensitive velocity control, plus a new Rhythm Generator view for drum programming. If you own Push 3, these firmware/software upgrades broaden its standalone studio potential.
Why this matters for live performers: being able to plug a wider range of audio interfaces into Push 3 without complex routing opens up richer live setups. For producers, the XY mode and improved sequencing make Push more tactile and creative for beat design and expressive performance.
Smaller but meaningful updates
12.3 also brings a handful of thoughtful enhancements that will please power users:
-
Auto Pan → Auto Pan-Tremolo: The Auto Pan device gets more modes and dynamic responsiveness, including dedicated tremolo behavior and level-based response shaping — great for rhythmic modulation and dynamic pumping effects.
-
A/B states for devices: You can set A and B states for instruments and effects and flip between them easily. That’s a huge boon for sound design and comparative mixing — quickly audition two radically different device settings without losing your place.
-
New Packs and Max for Live tools: Live Standard and Suite users get new creative devices (Patterns, Sting) and updates to Expressive Chords and Sequencers — useful for generative ideas and getting out of production ruts.
These are the kinds of improvements that might not make the big headlines but end up improving daily workflows — faster experimentation, easier comparisons, and more creative starting points.
Performance, platform fixes, and stability
As with most iterative updates, Ableton has bundled a number of stability and compatibility fixes across platforms. Release notes mention fixes for stem separation failures on certain macOS configurations and improvements to the Splice UI behavior, among other bug fixes. The public beta cycle surfaced issues and Ableton addressed several of them before pushing the stable release, which is reassuring for users who rely on Live in critical sessions.
If you depend on specific third-party plug-ins or unusual workflows, it’s always wise to test 12.3 on a copy of your projects first — don’t overwrite production sessions until you confirm all the plug-ins and setups behave as expected.
Pricing, availability, and the limited-time offer
Ableton is releasing 12.3 as a free update for everyone already on Live 12 (so if you’re on Live 12 Standard or Suite, it’s yours at no extra cost). For newcomers or those upgrading from much older versions, Ableton is running a limited-time promotion: 25% off Live 12 and Packs (and 20% off Push 3 and related hardware) for a short window around the release. If you were sitting on the fence about upgrading or buying in, that’s a practical savings window to consider.
Who should care — and who might want to wait
-
Remixers and sample-based producers: stem separation and Splice integration are direct wins. Faster sample hunting plus local stem extraction changes the way you can build remixes and reworks.
-
Live performers and Push 3 owners: Push 3’s standalone expansions and Bounce Groups make set preparation and standalone jamming more powerful.
-
Sound designers and experimental producers: A/B device states, Patterns, Sting, and Max for Live updates give fresh sound-design workflows.
-
Users running older or very plugin-heavy projects: test first. While Ableton has patched many issues, complex third-party setups sometimes reveal edge cases.
Quick hands-on tips to get started with 12.3
-
Try stem separation on a few different source types — a dry acapella, a full mixed track, and a drum loop — to learn the tool’s strengths and limitations. Resample processed stems to hide artifacts.
-
Open the Splice panel and experiment with Search with Sound: drag a loop or capture a section, let Splice find matches, and audition samples in sync. The faster previews will change how you hunt for sounds.
-
Bounce a group, then compare: use Bounce Groups to commit heavy group processing to audio, then A/B with the original to confirm what you’ve gained or lost.
-
Explore Push 3’s new modes if you own one — class-compliant audio support alone opens up new routing possibilities for standalone rigs.
Final thoughts
Ableton Live 12.3 is a model of how mature DAWs continue to evolve: headline features (stem separation, Splice integration) that grab attention, plus a steady stream of smaller but meaningful workflow and hardware updates. The inclusion of local stem separation is a game-changer for many workflows, while Splice integration tightens the loop between inspiration and production. For Live users, the update is essentially a must-install; for newcomers, the limited-time discount sweetens the deal.
If you use Live in any professional or semi-professional capacity, treat 12.3 as a strong, practical refinement — one that delivers immediate creative tools and sensible performance improvements rather than risky, half-baked experimentation. Fire up a non-critical project, poke around the new panels and devices, and you’ll likely find something that speeds up your process or sparks a new idea.
Suno and Warner Music Group: what the settlement means for artists, AI, and the future of music
Written by Sounds SpaceSuno and Warner Music Group: what the settlement means for artists, AI, and the future of music
This week’s surprising — and, for some, seismic — news that Warner Music Group (WMG) has settled its lawsuit with AI music generator Suno and entered into a commercial partnership marks a turning point in the music industry’s standoff with generative AI. After more than a year of litigation, public debate and fear among creators about the misuse of their work, the two companies announced a deal that aims to balance Suno’s rapid technical progress with protections and revenue for artists. Below, I unpack what happened, what’s in the agreement as reported, the likely consequences for creators and platforms, and what the settlement tells us about how music and AI will coexist going forward.
The headline: settlement + partnership, not a courtroom victory
At the heart of the story is a straightforward pivot: Warner, which had been litigating against Suno for alleged copyright infringement, is no longer pursuing that courtroom route — instead, the companies have struck a licensing and commercial partnership. As part of the agreement, Suno will phase out its current broad-use models in favour of licensed models, implement new restrictions on downloads (including limiting or blocking downloads for free users and capping paid-user downloads), and introduce mechanisms that let Warner artists and songwriters control whether and how their names, voices, likenesses, and compositions are used on Suno’s platform. Suno also announced the acquisition of Songkick — a live-music discovery brand previously under Warner’s control — as part of the broader transaction.
That combination — settling the legal claim while creating a commercial path forward — signals both sides’ strategic thinking. Warner gained contractual safeguards and revenue opportunities without protracted litigation; Suno gained industry legitimacy and access to an enormous catalog of artist assets and marketing channels. But the wrinkle is important: the agreement appears to be opt-in for artists. That means the platform can offer realistic “artist-like” outputs only for those creators who choose to participate, while others are (in principle) protected from unauthorized stylistic or likeness use.
What the agreement reportedly includes (the practical bits)
Reporting across major outlets and Suno’s own announcement gives us a reasonably clear list of concrete changes and commitments:
• Suno will launch new, licensed AI models next year to replace its current open models, designed to operate under licensing terms that compensate rights-holders.
• Downloads of AI-generated audio will be restricted: free accounts will no longer be able to download songs; paid accounts will have download caps and the option to pay more for higher allowances.
• Artists and songwriters signed to Warner will be able to opt in (or not) to have their voices, names, likenesses, and compositions used; participating artists will receive compensation as negotiated.
• Suno acquired Songkick from Warner; Songkick will remain as a fan destination under Suno’s ownership.
Warner and Suno have not publicly disclosed the financial terms of the settlement, nor the exact revenue split for artists — understandably, these are sensitive commercial details. But the structural commitments are what matter most for industry precedent: an existing AI vendor has accepted licensing obligations, and a major label has accepted a commercial route instead of purely legal enforcement.
Why this matters: three immediate implications
-
It normalizes licensing as the dominant path forward. For much of 2024–2025 the dispute between labels and AI companies felt binary: either platforms would be forced to stop using copyrighted content in training and outputs, or labels would license their works. This settlement validates the licensing route. If licensing becomes the industry norm, it means artists (and labels) can directly capture value when their style or likeness is used, rather than relying on uncertain litigation victories.
-
It sets an opt-in model for artist control. The opt-in approach that Warner and Suno are implementing is significant because it preserves artist agency. Rather than an all-or-nothing ban, artists can choose new revenue streams while controlling the use of their voice and other personal rights. That could become a blueprint for other deals — but it also raises questions about bargaining power and transparency in the offers artists receive.
-
It changes product economics for AI platforms. Requiring paid accounts and capping downloads shifts the economics of Suno’s product from an ad/scale-first free model to a hybrid subscription/licensing model. That could slow viral, mass-free proliferation of AI-created tracks, while creating predictable revenue lines that can be shared with rights-holders. It’s a step away from the “wild west” era of unfettered generation.
The broader legal landscape: one settlement, more work to do
It’s important to remember this is not the end of legal friction. Universal Music Group previously settled with AI platform Udio, and both Sony and Universal have ongoing or recent legal entanglements with AI startups. Suno’s deal with Warner doesn’t automatically resolve disputes with other labels or publishers, nor does it fully answer hard questions about training datasets, informed consent, or derivative uses. The settlement shows parties can negotiate, but it does not replace necessary public policy discussions about how copyright law should adapt to generative models.
Moreover, artist advocacy groups and creators have voiced concerns about transparency and fairness in earlier label-AI deals. Opt-in programs look good on paper, but their fairness depends on the visibility of contract terms and the bargaining leverage individual artists — especially less established ones — can realistically exercise. Without standardised transparency practices or industry-wide minimums, a fragmented patchwork of label-specific offers could leave many creators at a disadvantage.
For artists: opportunity and risk
For major, established artists, the Suno–Warner deal represents a new monetization channel and a way to control fan experiences. Imagine paid fan interactions where users can generate licensed remixes or AI-collabs that include an artist’s signature timbre or phrasing — with the artist paid for each use. That can be lucrative.
But for less-visible artists, the risks are real. Labels often negotiate on behalf of roster artists; depending on contract terms, an artist might find their likeness or composition licensed without direct negotiation with them, or receive royalties according to label agreements that artists already critique as opaque. The real question is whether these licensing mechanisms will be transparent and equitable at the artist level, not just at the label level.
For platforms and startups: a new playbook — at a price
For generative-audio startups, the message is clear: build licensing into the product roadmap early. Models trained on copyrighted works without clearance create legal risk and, as we’ve seen, can end with either an injunction or a negotiated settlement that forces product redesign. For investors and founders, the Suno–Warner deal shows a path to scale that includes paying for rights and restructuring user experiences — but it comes with higher operating costs and a need for careful artist relations.
This outcome also tilts the playing field toward platforms that can shoulder licensing costs and administrative complexity: well-funded startups, major tech companies or ventures backed by institutional capital. Smaller open-source projects and hobbyist tools may find it harder to offer competitive functionality without proper rights clearance.
Policy and the public interest: what regulators should watch
The Suno–Warner settlement may reduce the immediate urgency for aggressive legislative fixes, but regulators still have work to do. Key policy areas include: clarifying whether training on copyrighted material qualifies as fair use, mandating transparency around datasets, and ensuring a reasonable framework for attribution and compensation. There’s also a consumer-protection angle — users should understand when they are generating content that directly resembles a living artist and what rights attach to that output.
Finally, antitrust considerations could emerge — as major labels and well-funded AI firms stitch commercial relationships, regulators should watch for exclusivity deals that could lock out competitors or limit creative alternatives for artists and fans.
The cultural question: will AI-made music co-exist with human artistry?
Beyond contracts and code lies a cultural debate. AI music platforms can democratize production, allowing novices to create richer-sounding tracks and fans to experience new kinds of interactivity. But they also risk diluting artistic labor if the market floods with indistinguishable, cheaply generated tracks. The Suno–Warner settlement nudges the ecosystem toward a model where human artistry is recognized as a monetizable input to AI outputs — a compromise that, if implemented fairly, might preserve creative incentives.
That said, commercial deals alone won’t answer aesthetic questions about authenticity and taste. Those will be decided over time by listeners, curators, and creators themselves. If fans value the human story behind songs — the voice, lived experience, performance — then human artists retain a cultural edge. If, instead, novelty and volume dominate streaming economies, artists will need to adapt their business models accordingly.
Final thoughts: a test case, not the final chapter
The Suno–Warner settlement is a high-profile test case in an industry grappling with technological change. It shows litigation and negotiation can coexist: the threat of legal action pushed an AI startup to the bargaining table, and the settlement produced a commercial framework that may become a template for other deals. But the devil is in the details — the fairness of compensation, the transparency of contracts, and the long-term policy framework will determine whether this outcome is a durable solution or merely a stopgap.
For artists, the takeaway is simple but urgent: engage with these developments, seek clarity about contracts, and insist on transparency. For platforms, the lesson is equally practical: if you want to scale in music, plan to pay — financially and reputationally — to licence the human artistry that makes music meaningful.
We’re in the early chapters of the AI-music story. The Suno–Warner deal doesn’t end the debate, but it moves the conversation from purely adversarial litigation to negotiated commerce — and that shift has consequences for how music will be made, shared and valued in the years ahead.
Sources: reporting and company announcements from TechCrunch, Suno’s official blog, the Los Angeles Times, The Guardian and Pitchfork informed this analysis.
If you’d like, I can now:
• Draft a short “artist-facing” explainer summarizing what this deal means for an individual artist (rights, revenue and questions to ask).
• Or create a one-page checklist for independent musicians to protect their work and negotiate AI licensing terms. Which would help you more?
When the majors embraced the machine: what KLAY’s licensing deals with the “big three” mean for music’s AI future
Written by Sounds SpaceWhen the majors embraced the machine: what KLAY’s licensing deals with the “big three” mean for music’s AI future
The music industry has been wrestling with artificial intelligence for the better part of this decade — first in courtrooms and headlines, then in negotiation rooms and press releases. This week’s news that Los Angeles–based AI music startup KLAY Vision Inc. has struck licensing deals with the three major record groups — Universal Music Group, Sony Music (and their publishing arms), and Warner Music Group — marks a clear inflection point. For the first time, the companies that own the lion’s share of recorded and published music have formally licensed their catalogs to an AI-driven streaming product intended to let fans remake songs in new styles. That’s business-as-usual turned on its head: the record industry is no longer only defending copyright against automated imitators — it’s now actively building a commercial framework to monetize AI creativity.
Below, I unpack what the KLAY deals actually are, why they matter, the unanswered questions they raise for artists and listeners, and what they suggest about how the music business will adapt to — and profit from — generative audio technologies.
What KLAY says it does (and what the labels licensed)
KLAY bills itself as a music-technology company building a streaming product with integrated AI tools that let users remix and “remake” songs. The startup says it trained its systems on thousands of licensed recordings and compositions, and that the agreements with Universal, Sony, and Warner grant KLAY the rights to use those catalogs both to power its machine-learning models and to enable listeners to create personalized AI variations of the original songs. In press materials and label statements, the emphasis is on a paid, controlled system that compensates rights-holders and gives participating artists “control” over how their work is used.
Put simply: instead of AI engines scraping the internet for music (and getting sued for it), KLAY has secured the catalogs it needs officially — licensing the underlying recordings and publishing rights so it can legally train models and offer derivative experiences to subscribers. This is the fundamental pivot from conflict to commerce.
Why the labels decided to license, not litigate
The move reflects a broader cooling of the industry’s initial, lawsuit-heavy reaction to AI-generated music. Early skirmishes — including high-profile cases targeting companies accused of training models on copyrighted recordings without permission — exposed how difficult it is to police every synthetic output on major platforms. Licensing gives labels an avenue to recapture value from AI activity rather than leave it to unregulated creators and deepfake projects.
From a business perspective, the labels are treating AI as a new revenue stream: licensed models can be monetized through subscriptions, micro-payments for individual remixes, and potential new sync/derivative licensing arrangements for commercial use. By agreeing on terms now, labels lock in payment frameworks and control mechanisms rather than ceding the field to adversaries. Reuters and other industry outlets reporting on the deals note that this is the first time a single AI company has landed agreements across all three majors — a symbolic and practical milestone.
What this could mean for artists and songwriters
The labels stress payments and artist protection, but practical outcomes will depend on contract details that aren’t yet public. Important questions include:
-
Revenue split mechanics: Will revenues from AI-remixes be shared with performers, songwriters, producers and publishers in the same way as traditional streams, or will there be a bespoke split for AI-derived works? Early statements promise “proper recognition and reward,” but the breakdown — and whether smaller artists get a fair share — is not yet transparent.
-
Opt-in vs opt-out: Some label statements imply artist participation will be managed through label agreements, but will individual artists be able to opt out of allowing their work to be used by AI remixes? Opt-in models (where artists explicitly authorize usage) are friendlier to artist autonomy; blanket label-level licensing is administratively simpler but risks leaving artists surprised by new forms of derivative use.
-
Moral rights and voice cloning: Even when a song is licensed, voice impersonation or highly realistic vocal “clones” of living artists raise unique reputational and ethical concerns. Will artists be able to veto impersonations that sound like them, or require additional consent for likeness-based uses? The deals suggest labels will attempt to govern these uses, but public-facing safeguards remain to be seen.
In short: licensing opens the door to new income streams, but it also creates novel bargaining points and potential pitfalls. Musicians and their managers should read the fine print closely as product specifics unfold.
How KLAY’s product model could work in practice
Based on available reporting, KLAY appears to be positioning itself as a hybrid of a streaming service and an interactive creative tool. Imagine a subscription app where a fan can pull up a favorite track and choose stylistic options — tempo, instrumentation, genre twist, vocal emphasis — and the AI generates a personalized version on the fly. That output would be playable inside the app, shared within the platform, and possibly downloadable or licensed for other uses depending on rights negotiated.
Business models that have been floated by industry commentators include:
-
Subscription-first: paying for access to the AI remixing engine and licensed catalog (similar to an advanced streaming tier).
-
Per-use micropayments: small fees for individual remixes or for the right to download/share an AI-created version.
-
Creator tools and marketplace: artists and producers could sell presets, commercial licenses, or custom remixes created with the platform.
All of these require plumbing to allocate money back to the correct rights-holders and to catalogue derivative uses for accounting — a nontrivial metadata and royalty-engineering challenge.
Risks and friction points
While the deals are a big step toward normalization, several hazards remain:
-
Quality dilution and platform clutter: Generative tools can flood platforms with low-quality, AI-generated tracks masquerading as new music. That trend has already raised concerns on existing streaming services where “AI spam” has pushed down discoverability for human artists. A licensed approach may mitigate some of that, but discovery algorithms and curation will matter.
-
Deepfake vocals and likeness rights: Even with catalogs licensed, using an artist’s vocal likeness without separate consent is a distinct legal and ethical issue. Platforms and labels will need enforceable rules about vocal impersonation and identity-protecting mechanisms.
-
Creative credit and attribution: How will AI-remixes credit the original artist, songwriter, and the machine model? Proper attribution standards are crucial for maintaining transparency and preserving the lineage of creative works.
-
Inequality of bargaining power: Major-label artists and top-tier songwriters are likely to get better negotiated terms than independent or legacy artists signed on older deals. Without industry-wide standards, smaller creators could be left behind.
Why is this also an opportunity
For listeners, platforms like KLAY could unlock new ways to interact with music: personalization, reimagined versions of favorites, and collaborative experiences where fans co-create with AI. For creators, licensed AI can serve as an additional income source and a promotional tool: remixes can introduce songs to new audiences, and interactive experiences can enhance fan engagement.
More broadly, the deals hint at an industry strategy: partner with AI entrepreneurs and formalize revenue pathways rather than attempting to shut down innovation entirely. That approach promises to steer AI development toward mutually beneficial products — if the commercial terms are fair and transparent.
What to watch next
If you’re tracking this story, here are the developments that will matter most in the coming months:
-
Public launch and product UX: How KLAY implements artist controls, attribution, and payment flows in its actual app will show whether the deals were symbolic or substance-driven. A consumer launch will reveal whether listeners embrace remixable subscriptions.
-
Artist reactions and opt-ins: Watch for statements or actions from major artists about whether they authorize voice-cloning-like features for their work, and whether there’s an opt-out movement among musicians uncomfortable with the new model.
-
Regulatory scrutiny and industry standards: Governments and industry bodies may step in to set guidelines around deepfakes, voice rights, and transparent royalty reporting. The deals with the majors could influence those standards.
-
Extensions and competition: Other AI firms and major labels may ink similar deals, or the majors might partner with multiple AI players. Competition could be healthy — pushing platforms to improve artist protections and user experience.
Final take: pragmatic adaptation, with caveats
KLAY’s licensing deals with Universal, Sony and Warner represent a pragmatic shift: the music industry is no longer fighting the generative-AI tide by default. Instead, it’s trying to steer that tide into monetizable channels. For fans and creators, that can be a win — new experiences and revenue streams — but the outcome depends on how equitably and transparently the underlying systems are built.
This moment is a reminder that technology doesn’t just disrupt markets: it forces social and legal frameworks to catch up. Licensing the catalogs was step one. Step two is delivering a platform that truly balances innovation, artist rights, and fair payment. If the majors and KLAY can design that system together — with clear artist consent, transparent accounting, and sensible limits on voice impersonation — the music ecosystem may emerge more diverse and more lucrative for creators. If not, the next chapter will be courtrooms and headlines all over again.
Either way, the industry’s choice to collaborate with — rather than simply contest — AI companies signals that generative audio is here to stay. The crucial question now is who writes the rules, and whether those rules will protect the people who make the music.
“Music producers are rejecting AI”: what the new studies tell us — and what it means for music’s future
Written by Sounds Space“Music producers are rejecting AI”: what the new studies tell us and what it means for music’s future
Brother — the headlines have been loud this week: a new survey says over 80% of music producers are actively against AI-generated songs, while separate listener-focused research shows most people can’t tell AI-made tracks from human ones. That collision — creators rejecting the tech at the same time listeners can’t reliably detect it — is the perfect storm for arguments about ethics, jobs, creativity, and regulation. Below, I’ll walk you through the findings, explain why they matter, and give a blunt take on what could happen next.
The headline findings:
• A global survey of music producers, commissioned by Tracklib, reports that a very large majority of producers are sceptical or hostile to generative-AI use in music: adoption is low and rejection high — the press release frames it as “music producers are rejecting AI,” with only a small percentage actively using generative tools.
• At the same time, Deezer and Ipsos ran listening tests and found that ~97% of people couldn’t reliably tell AI-generated tracks from human-made ones in blind tests — a striking demonstration that AI music quality has reached near-parity for the ear of the average listener. That same study shows big public support for transparency (label AI-made tracks) and concerns about ethics and copyright.
• Major outlets reporting on these themes also point to real-world effects: AI tracks are already rising on streaming platforms and even some charts, which is intensifying industry debate about promotion, royalties, and verification.
What the producers’ rejection actually means
When a study says “80%+ of producers are against AI,” don’t read that as a single emotion — it’s a mix of reasons:
-
Artistic integrity and craft. Many producers see production choices as part of artistic identity. Machines that spit out passable arrangements or vocal lines feel like a shortcut that removes the human voice from the process. Tracklib’s survey captures that sentiment: a lot of rejection is because producers feel AI undermines the craft.
-
Quality concerns — and nuance. Some producers reject full-song generation but still use AI for small tasks (like reference stems, mastering presets, or chord suggestions). The “rejecting AI” headline often conceals that many are open to assistive AI, just not to AI that replaces core creative decisions.
-
Copyright and fairness. Producers worry about training data (AI trained on copyrighted music without consent), displacement, and how royalties should be split if models borrow from identifiable artist outputs. The Deezer/Ipsos work also shows public support for ethical training and labeling — so this is both a creator and consumer issue.
-
Economic risk. If streaming platforms allow AI tracks to surge without clear rules for attribution and payment, session work, ghost production, and publishing could lose income. That’s why producers are defensive — it’s their livelihood, not just ideology.
Why listeners’ inability to tell matters
The Deezer/Ipsos finding that ~97% of listeners failed to detect AI tracks is a game-changer for three reasons:
-
Market-level indistinguishability. If audiences can’t tell, then AI tracks can compete on the same platforms, playlists, and algorithms. That puts human artists in direct competition with low-cost synthetic output unless platforms treat them differently.
-
Transparency demand. The public’s unease in that study shows people want to know. If platforms don’t label AI content, they risk trust erosion and regulators stepping in. Deezer has already started tagging AI submissions — an early institutional response.
-
Policy & chart implications. Charts and awards rely on attribution; if synthetic works crowd charts, the industry must decide whether to treat them equally. We’re already seeing cases of AI songs making chart noise and being removed or disputed. That bubbling conflict is the reason labels, distributors, and streaming services are scrambling.
The contradictions: why both things can be true at once
It seems paradoxical: producers reject AI while listeners don’t care or can’t detect it. But both can be true because they’re answering different questions. Producers are speaking about process, ethics, and livelihoods; listeners are responding to end-product enjoyment. The industry is now forced to reconcile process (how music was made and who benefits) with product (what people enjoy or stream).
The pressure points: where conflict will be decided
Here are the battlegrounds we should watch:
-
Streaming platform policies. Will platforms require AI labels? Will they remove AI tracks from editorial playlists? Deezer’s tagging is a blueprint; other platforms may follow or resist.
-
Copyright litigation & licensing. Lawsuits are already happening and will accelerate. Courts will decide whether training models on copyrighted works is infringement, and that ruling will determine cost structures for AI music firms.
-
Royalties & metadata standards. If AI tracks use samples or stylistic mimicry, who gets paid? The industry needs metadata standards and traceability to assign rights and royalties correctly — otherwise, payments and charts will be gamed.
-
Consumer-facing transparency. Mandatory labelling is likely to be politically popular and technically feasible. Expect regulators to push for clarity so consumers know whether they’re listening to AI.
-
Artist and union pressure. Creators’ associations and unions will push for protections, training data consent, and revenue-sharing terms. If producers maintain unified resistance, they can shape licensing frameworks more than tech companies expect.
The likely near-term scenarios (what I actually expect)
-
Hybrid adoption increases. Most producers will continue using AI as assistants (mixing tools, stems, mastering suggestions) rather than full-song generators. The “pure AI” artist will mostly be an outlier or novelty that occasionally breaks through virally.
-
Platform responses: tags + restrictions. Platforms will adopt transparent tagging and may restrict AI-only tracks from editorial playlists and algorithmic boosts until legal/ethical frameworks are set — Deezer is already an early mover here. Expect other major services to follow to avoid reputational and legal risk.
-
Regulatory and legal clarifications. Courts and regulators will clarify what “training on copyrighted works” means. If rulings favor copyright holders, AI music firms will need licensing deals akin to the way streaming licenses catalogs, which raises costs and reduces the free-for-all.
-
An uneven marketplace. We’ll end up with two flavors of music commerce: human-made content protected, promoted, and monetized under traditional deals; and AI-generated content that lives in a parallel space with different rules, visibility, and monetization models. Chart systems may split or flag AI entries.
What this means for producers, labels, and listeners: practical takeaways
For producers:
-
Lean into what makes you uniquely human: emotional risk-taking, imperfections, personal storytelling, and brand as a creator. These are the qualities AI struggles to authentically replicate.
-
Treat AI as a tool, not a competitor. Learn to use assistive AI for workflow speedups (mastering, stem separation, idea generation) while protecting your signature creative decisions.
For labels & distributors:
-
Invest in metadata standards and provenance systems now. If you don’t capture who did what — and whether a track used AI — you’ll lose control over chart inclusion, payouts, and legal defenses.
-
Think about curated playlists as a premium human-made space; reserve some editorial real estate for verified human artistry to protect brand value.
For listeners:
-
Expect transparency features (filters, labels, opt-outs). If you want only human-made music, the tools to enforce that preference are likely coming.
Final thought, this is a crossroads, not an end
AI music is no longer a futuristic demo or toy. The technology is good enough to matter commercially and culturally, and the industry’s response will shape careers and revenue flows for a generation. Producers’ concern is legitimate — it’s about artistry and survival. Listeners’ reaction is equally important: they show curiosity but also want honesty.
The healthy path is a negotiated one: transparency, licensing, and new business models that protect creators while allowing innovation. If the industry can’t agree, regulation will force the issue — and that could either protect creators or ossify the market. Either way, producers rejecting AI today is a loud signal: the music world wants rules before mass adoption. That moment of negotiation is the real story behind the headline.
Suno raises $250M — what this means for music, makers, and the industry
Written by Sounds SpaceSuno raises $250M — what this means for music, makers, and the industry
Big money is chasing big ideas. AI music startup Suno just closed a $250 million Series C round at a reported $2.45 billion post-money valuation — a jump that underlines investor faith in generative audio, and at the same time stokes every debate happening around AI and creative work right now. This isn't just another VC headline: it has real consequences for artists, labels, startups and anyone who cares about how music gets made, credited, and monetized.
Belo,w I break down the news, why investors are writing huge checks, the tensions and lawsuits now in play, and what musicians and music businesses should actually do next.
The headlines (fast)
-
Suno raised $250M in a Series C led by Menlo Ventures, with participation from groups including NVIDIA’s venture arm and others. The round reportedly values the company at $2.45 billion.
-
Published reporting indicates Suno’s growth has been rapid (user growth and monetization metrics are being cited in industry coverage) — some outlets claim high revenue and quick adoption.
-
At the same time, Suno is entangled in copyright challenges and criticism from musicians and labels about how generative music models were trained and how outputs resemble existing works. The legal and cultural fights are unresolved.
Why investors doubled down — a simple math + product story
Investors put money where they see leverage: a product that can scale rapidly with relatively low marginal cost, a big addressable market, and defensible tech or distribution. Suno checks several boxes:
-
Mass-market potential. Text-to-music and AI-assisted songwriting promise to democratize music production: people who never touched a DAW can generate a song from a prompt. That expands the total market of music creators — and therefore potential paying customers, creators who want tool subscriptions, and businesses that need licensed music.
-
Platform economics. If Suno can combine an easy generator with a marketplace, publishing or licensing deals, it can earn recurring revenue on subscriptions, commercial licenses, and partnerships (e.g., plugin bundles, enterprise licensing for games, ads, content creators). Multiple recent write-ups highlight Suno’s strong monetization trajectory, which helps justify a big valuation.
-
Strategic investors. With participation from deep-tech and media investors, Suno gains both capital and routes into key partners — think distributed compute (NVIDIA), media relationships, and scaling channels. Those backers see AI audio as infrastructure that will be embedded across content pipelines.
Put together, the logic is: invest now to own the interface and distribution layer for AI music before incumbents (DAWs, labels, platform giants) lock it down.
But there’s a huge “but”: copyright, ethics, and cultural pushback
Money buys growth, not goodwill. The same scaling that makes Suno attractive to investors also amplifies the sticky problems everyone warned about.
-
Lawsuits and label pushback. Major labels and some artists have filed or threatened legal action alleging Suno trained models on copyrighted recordings without adequate consent or compensation. Those suits could force licensing deals, statutory damages, or changes in how models are built and monetized — each of which would hit margins and product claims.
-
Creative legitimacy. Critics argue that clicking “generate” is not the same as composing or performing — and that mass-producing music risks diluting cultural value and displacing paid work. Defenders say Suno’s tools augment creators and open new pipelines for musical discovery. That debate isn't just philosophical; it affects regulation, platform policy, and public perception.
-
Platform policy risk. Streaming platforms and distributors are already experimenting with AI content rules. If major streaming platforms restrict AI-generated content or demand specific metadata/labels and licensing, Suno’s commercial playbook may have to adapt quickly.
What will the $250M will likely fund?
Sizable rounds don’t just cover payroll. Watch for:
-
Product R&D at scale. Faster models, higher-quality vocals, longer-form music, and tools that enable customization and stems for mixing. Suno will invest in making outputs sound more polished and controllable — a route to premium tiers.
-
Legal and licensing teams. Expect Suno to beef up legal resources and negotiate licensing deals with publishers/labels — whether proactively or under pressure. That’s expensive but strategically necessary if they want to serve commercial customers.
-
Distribution and enterprise partnerships. Integrations with DAWs, game engines, ad platforms, and content platforms. With partners like NVIDIA (or NVentures’ involvement), cloud and inference costs could be optimized.
-
Creator economy features. Marketplaces, revenue shares, creator profiles, and tools for human+AI collaboration that let musicians add their voice and retain rights — those features can help reduce backlash and create a community that Suno can monetize.
How this affects musicians, producers, and indie labels
Short answer: the world changes, but you can steer your outcome.
-
New income streams, if you adapt. Musicians who learn to use AI as a co-producer, who sell curated stems, or who offer performance/production services on top of AI demos can create new revenue models. Tools that reduce routine work free time for higher-value creative tasks.
-
Competition for stock music and low-end beats. Expect price pressure when it comes to commoditized background music and simple beats. If you sell simple loops or generic beats, you’ll feel it first. Level up your offerings: unique sounds, signature production, bespoke performances.
-
Negotiation leverage with platforms. As AI content grows, musicians and rights-holders should demand better metadata, attribution, and share of revenue when models or platforms use their material. Collective bargaining, unionization, or industry-wide licenses may become necessary.
-
Credibility and branding matter more. Human-driven musical artistry becomes a differentiator. Fans will value authenticity, live performance, and artist narratives. Musicians who emphasize craft and story can maintain premium positioning.
What labels and publishers will do and why it matters:
Labels don’t want to be sidelined. Expect:
-
Licensing deals and partnerships with AI vendors — labels get paid and retain control over how recordings are used. That's a comfortable outcome for rights-holders if the economics work.
-
Litigation as leverage. Suits can force settlement talks or shape regulation. Even if litigation is slow, it creates a bargaining chip and public pressure.
-
New product offerings. Labels might launch their own AI tools or partner with platforms to offer branded, safe AI sounds — turning risk into opportunity.
Regulatory and platform questions to watch
-
Transparency mandates. Will platforms require AI-generated tracks to be labeled? How will attribution be tracked across distribution chains?
-
Compulsory licensing vs. negotiated deals. Regulators could push for clear rules on training data rights and compensation (analogous to sample-clearing regimes). That would change unit economics for every AI music company.
-
Data rights and model audits. Governments or industry bodies might demand audits of training datasets or new standards for demonstrable provenance.
Regulation is messy and slow — but high valuations like this make regulation inevitable, as the stakes are significant.
Practical takeaways for creators and music businesses
If you’re a musician, producer or indie label:
-
Experiment with AI but keep control. Use Suno and other tools to prototype quickly, but always decide which outputs become public and how you protect your brand and rights.
-
Treat AI as a collaborator, not a replacement. Use it to sketch ideas, expand arrangements, or create stems — then add your unique performance, production or vocal signature.
-
Document provenance and contracts. When licensing music, be explicit about AI usage in contracts and distribution metadata.
-
Learn to monetize beyond streaming. Offer stems, bundles, custom production, licensing for indie games/ads, or educational content. Diversify income.
-
Get savvy about IP. If you care about how models use existing works, follow industry litigation, and consider joining collectives lobbying for fair terms.
Final thought: The music ecosystem will adapt
Suno’s $250M raise is a landmark signal: investors expect AI to be central to the next wave of music creation and distribution. That will bring both opportunity and friction. Artists who adapt, insist on fair terms, and leverage the new tools creatively can thrive. The ones who don't will face a reshaped marketplace where commoditized sounds are abundant and differentiation is everything.
This moment is a crossroads. Massive capital accelerates capability; the public, artists and policymakers will decide the boundaries. For musicians: learn the tools, protect your craft, and keep making music that only a human can make — because even in a future of perfect algorithms, humans will still be the ones who turn sound into meaning.
Serato DJ: Host of new Crate management options lead its 4.0 update
Written by Sounds SpaceSerato DJ: A host of new Crate-management options lead its 4.0 update
Serato has just rolled out the biggest library overhaul in years. With the 4.0 update — released as Serato DJ Pro / Lite 4.0 and promoted through a public beta — the company has refocused its attention away from flashy effects and toward something DJs quietly crave: better ways to organize and access music. The result is a smarter, faster, and more flexible library system built around powerful crate management tools that make prepping sets and finding the right track on the fly far less frustrating.
Below, I unpack the most important additions, why they matter, and how DJs of all stripes can use the new library features to speed up workflows and boost creativity.
What’s new? The highlights
While 4.0 includes a variety of UX and performance tweaks, the marquee improvements center on crates — Serato’s core organizational unit. The most important additions are:
-
Crate Search — quickly find the crate you need by name. No more scrolling through long lists. (Pro-only feature).
-
Favorite Crates (Shortcuts / Pinning) — pin go-to crates to the top of your list for instant access during a set.
-
Crate Color—color—code crates using a right-click palette for visual separation and faster scanning.
-
Crate Sort — sort crates by alphabet, date created, or arrange custom orders.
-
Show Track in Crates — right-click any track and instantly see every crate the track lives in. Great for checking overlaps and curating transitions.
-
Streaming Tracks in Crates — add tracks from streaming services into crates alongside local files, reducing the friction of mixed-source sets.
-
Track Rating (emoji or stars) — rate tracks in the library using stars or playful emojis (watermelons, ghosts, etc.) for fast recall.
-
Crate Info — see crate metadata like total track count, total runtime, and file size in the status bar. Useful for timing sets.
-
Resizable Album Artwork & UI polishing — resize artwork in the library view, improved file relocation, safer USB ejection, and automatic analysis on import.
These features combine to make Serato’s library both more informative and more manipulable — two qualities that add up to real time saved during prep and performance.
Why crate management matters
Organizing music isn’t glamorous, but it’s foundational. A DJ’s library is the toolbox for every set; when tools are poorly organized, the performance suffers. The new crate system attacks several long-standing pain points:
-
Speed of access: Crate Search + Favorites removes the need to dig through nested folders or remember obscure crate names. In a busy club or festival, shaving seconds from a search can change the course of a set.
-
Visual scanning: Color coding and resizable album art let DJs build a visual language for their library — e.g., blue crates for house, red for peak-time weapons, green for vocal edits — speeding recognition when adrenaline’s high.
-
Context awareness: “Show Track in Crates” and Crate Info remove guesswork. Want to know if that remix is already in another playlist or how long a crate will last on the dancefloor? Now you can know instantly.
-
Bridging streaming + local files: Being able to store streaming tracks in crates reduces the friction when you mix paid downloads and on-demand tracks. It simplifies set building in hybrid workflows.
Put succinctly: better crate tools mean less cognitive load during prep and performance, and more headspace for creative decisions.
Practical workflows — how to use the new feature immediately.
1. Build a “go-to” rack: Create favorite crates for different parts of your set (warmup, peak, halftime, closer) and pin them. During a set, jump straight to the correct vibe without hunting.
2. Color code by energy & key: Use crate colors to encode energy levels or harmonic compatibility. For instance, green = compatible in key, orange = high-energy, blue = vocal tracks. Visual cues speed selection under pressure.
3. Use Track Rating as a preflight checklist: During practice, tag tracks with an emoji or star rating to indicate readiness: 5-star = performance-ready, 3-star = needs edits or cueing, 1-star = archive. Filter by rating when compiling a set.
4. Merge streaming & local crates for gig-specific builds: Add your subscription service tracks into crates with your local edits — then use Crate Info to verify runtime and track counts for set length planning.
5. Prep faster with Analyze on Import: Switch on automatic analysis so newly added tracks are ready to beatmatch and use stems/features without manual scanning. This is handy when hot-swapping files from USB drives.
Comparison: how Serato’s changes stack up
Competitors like Rekordbox and Traktor have offered various library conveniences for years. What’s notable about Serato 4.0 is the breadth of small, practical quality-of-life improvements delivered together — the kind of iterative polish that DJs notice in the field. Features like “Show Track in Crates” are reminiscent of functions some rivals implemented earlier, but Serato's approach to bringing multiple improvements at once (color, ratings, pinning, crate metadata) means you get a cohesive, modern library experience rather than a single standout tool.
Potential fixes & what to watch
No big software release is flawless. Early public beta threads point to a few things DJs should watch:
-
Crate Search in Lite: Some features (like crate search) are Pro-only. If you’re on Lite, check which tools you’ll miss before updating.
-
Library merging & external drives: while file relocation and hot-swap behavior have improved, DJs with complex external drive setups should test the new automatic merging carefully before relying on it in live shows.
-
Workflow muscle memory: any library overhaul may temporarily slow DJs who are used to older shortcuts — but the tradeoff is faster, clearer workflows after a short adjustment period.
Serato has emphasized community feedback during the public beta, so many small issues will likely be patched quickly. If you rely on Serato for gigs, don’t upgrade mid-gig season without confirming stability in your specific setup.
Pricing and availability
Serato released 4.0 as a public beta initially, with the company listing the update and downloadable builds on its site. Serato DJ Pro remains a paid product with purchase and subscription options, while Serato DJ Lite continues to be free — though some of the new features are Pro-only. Check Serato’s official downloads and release notes for the exact availability and the latest stable build.
Final thoughts — a library update that matters
Serato DJ 4.0 isn’t headline-grabbing because it adds a new effect or a flashy performance feature. It matters because it addresses the mundane, everyday work of DJs — organizing, finding, and prepping music — with practical tools that shave time off routine tasks and reduce friction in the creative process. For working DJs, producers who moonlight as selectors, mobile DJs, and club residents, those minutes reclaimed from searching and guessing add up to better sets and less stress.
If you’re a Serato user: test the public beta in a safe environment, try the crate color and favorites system, and begin using the rating system to build a library taxonomy that works for you. If you’re still on another platform, Serato 4.0 narrows the gap in library ergonomics and gives you one more good reason to reevaluate your DJ software workflow.
More...
SoundCloud's Game-Changing Leap: Artists Claim 100% of Royalties in New Era of Creator Empowerment
Written by Sounds SpaceSoundCloud's Game-Changing Leap: Artists Claim 100% of Royalties in New Era of Creator Empowerment
By Elena Vasquez, Music Business Daily, November 14, 2025
In a seismic shift that's sending ripples through the music industry, SoundCloud has unveiled a subscription model that hands artists the full reins on their earnings. Announced on October 30, 2025, the platform's new "All-In-One Artist Subscription" allows independent creators to retain 100% of royalties from music distributed across more than 60 streaming services, including Spotify, Apple Music, TikTok, and YouTube Music. No more 20% cuts from the distributor—every penny from streams, downloads, and syncs goes straight to the artist. This isn't just a tweak; it's a declaration of war on the opaque royalty structures that have long plagued emerging talent.
For years, SoundCloud has been the digital sandbox where bedroom producers and viral hopefuls test their sound, amassing over 175 million tracks and a fiercely loyal community of "music obsessives." But monetization? That was the rub. Under the previous SoundCloud for Artists Premier plan, creators kept about 80% of distribution royalties after the platform's slice, a model that mirrored competitors like DistroKid and TuneCore but still left many feeling shortchanged in an era where a single stream nets less than a tenth of a cent. Now, with this update effective for payouts starting November 2025, SoundCloud is betting big on transparency and artist-first economics to reclaim its throne as the go-to hub for indie music.
The timing couldn't be more poignant. As major labels consolidate power amid AI-driven disruptions and TikTok's algorithm-fueled virality, independent artists are squeezed harder than ever. A 2024 MIDiA Research report pegged the average indie artist's annual streaming revenue at just $1,200—barely enough for a month's rent in Brooklyn. SoundCloud's move feels like a lifeline, especially for the platform's core demographic: Gen Z and millennial creators who upload demos, remixes, and EPs without the safety net of a label advance.
From Uploads to Uplifts: SoundCloud's Evolution
SoundCloud's story is one of scrappy innovation turned industry staple. Founded in 2007 by Swedish entrepreneurs Alex Ljung and Eric Wahlforss, it started as a simple audio-sharing site for sound designers and DJs. By 2013, it had exploded into a discovery engine, launching careers like those of Billie Eilish (whose early tracks racked up millions of plays) and Post Malone. But growth brought growing pains: a 2017 funding crunch nearly sank the company, forcing it to pivot toward monetization tools like repost chains and basic royalties.
Enter SoundCloud for Artists in 2018, which bundled distribution, analytics, and promo tools into tiered plans. The Premier tier, at $99 annually (or about $8.25 monthly), promised 80% royalties on distributed tracks—a step up from free uploaders but still trailing the "zero-cut" dreams peddled by upstarts like Amuse. Artists grumbled about hidden fees, payout delays (net 60 days), and the platform's infamous "repayable advances" that could claw back earnings if streams tanked.
Fast-forward to 2025, and SoundCloud's under new stewardship after a 2023 acquisition by SiriusXM, which injected $200 million into creator tools. CEO Michael Herd, a former Spotify exec, has steered the ship toward "sustainable independence," citing data that 70% of SoundCloud's 40 million monthly users are undiscovered artists seeking direct fan connections. The All-In-One plan is the culmination of this vision: for that same $8.25 monthly fee, artists get unlimited distribution, built-in monetization, and zero platform cuts on royalties, Fan Support tips, or merch sales.
Digging into the fine print reveals a model built for scale. Tracks uploaded via the plan beam out to 60+ DSPs (digital service providers) instantly, tapping SoundCloud's 76 million global users for immediate plays. Royalties? 100% artist-retained, with only a nominal payout processing fee (as low as $0.50 per transaction via partners like PayPal or bank transfer) once earnings hit $25. No upfront costs for vinyl-on-demand releases—artists can press limited-edition runs without bulk minimums—or for linking merch stores, where SoundCloud takes zilch from sales.
It's not without caveats. The 100% share applies only to new distributions post-November 2025; legacy earnings might still reflect the old 80/20 split. And while SoundCloud touts "no hidden fees," tax forms and two-factor authentication are mandatory for payouts, a nod to fraud prevention amid rising cyber threats in creator economies. Still, for a Nashville trap producer like DXRKNOVA, who credits the platform for doubling her monthly earnings through targeted promo, it's a no-brainer upgrade.
The All-In-One Blueprint: Features That Fuel Futures
At its heart, the All-In-One Subscription is a Swiss Army knife for the solo artist hustler. Distribution is the star: upload once, and your track hits Spotify playlists, Apple Music's algorithmic feeds, and even niche spots like Deezer or Tidal. SoundCloud's edge? Its proprietary "Instant Reach" algorithm, which funnels new releases to superfans and tastemakers within hours, potentially spiking plays before competitors even index the file.
Monetization layers on seamlessly. Fan Support lets listeners tip directly—every cent to the artist, displayed as a "top supporter" badge on profiles to foster loyalty. Merch integration is plug-and-play: link your Teespring or Bandcamp store, and SoundCloud's storefront widget handles discovery with zero fees. For physical die-hards, the vinyl-on-demand tool turns digital hits into 7-inch pressings, shipped globally without inventory risks. "It's like having a label without the middleman," says Brooklyn DJ nextdimensional, whose EP Portal saw a 40% revenue bump after bundling streams with fan tips and limited-edition merch.
Analytics round out the package: real-time insights on listener demographics, stream sources, and even AI-powered "mood mapping" to suggest collab opportunities. At $99/year, it's competitively priced against DistroKid's $22.99 annual unlimited plan (which takes 0% but lacks built-in audience tools) or TuneCore's $29.99/single (with varying cuts). SoundCloud's pitch: Why fragment your workflow when one dashboard handles discovery, distribution, and dollars?
Critics might quibble over the subscription lock-in—free users still face upload limits and no distribution—but for pros, it's a steal. Early adopters report 15-20% earnings lifts, per internal SoundCloud data, thanks to the platform's 320 million tracks-strong library acting as a perpetual promo engine.
Voices from the Upload Trenches: Artist Reactions Pour In
The announcement lit up X like a viral remix. "SoundCloud the MVP baby 100% Royalties >," tweeted electronic artist Siren Diamant, capturing the euphoria among bedroom producers. South African DJ Tezz echoed the sentiment: "Major Update for Independent Artists... keep 100% of their royalties, not just on SoundCloud, but also on Spotify, Apple Music, TikTok and more." NME amplified the buzz, headlining it as a "100 per cent of distribution royalties" win, while Mixmag hailed it for dance music creators.
Not everyone's popping champagne uncorked. Your EDM's analysis cut through the hype: "When each stream is still worth a fraction of a cent, this indicates that artist's 'keeping' their royalties may not mean much at all." Indie vocal coach Randy Morano noted the fee: "100% royalties for creators (minus an $8.25/mo fee). A huge shift toward artist-first monetization." And Charm Cracker, a self-produced multi-instrumentalist, plugged referral codes, underscoring the grassroots uptake: "They are bypassing the competition... Spotify is not working for me, really."
Interviews with early switchers paint a nuanced picture. "I've been grinding SoundCloud since 2019, scraping by on 80% scraps," shares LA-based rapper J. Mira, who dropped her debut mixtape Echo Chamber last week. "Now? That extra 20% covers my plug-ins. It's not riches, but it's respect." For global acts like Romio, an Italian Web3 musician blending NFTs with beats, the plan aligns with blockchain's decentralization ethos: "Big news for indie artists! 100% of distribution royalties for tracks released via SoundCloud."
Even skeptics concede the momentum. Rolling Stone Brasil covered it in Portuguese, emphasizing accessibility for non-English markets, while Exclaim! touted it as a North American boon. As one anonymous SoundCloud vet put it: "It's the push we needed to treat music like a business, not a hobby."
Ripples in the Royalty Pool: Broader Industry Shifts
This isn't happening in a vacuum. The music biz is in flux: Universal Music Group's 2024 TikTok standoff exposed streaming's fragility, while AI tools like Suno democratize production but dilute per-stream value. Platforms like Bandcamp thrive on direct-to-fan models (100% artist cuts, minus 10-15% fees), but lack distribution scale. SoundCloud bridges that gap, potentially siphoning users from Spotify for Artists (70% royalties post-label cuts) or Apple's $9.99 pro tier.
Economically, it's a calculated risk. SoundCloud's revenue—pegged at $300 million annually—leans on ads and subs, not distribution vig. By waiving the 20% (estimated $50 million yearly), they're banking on volume: more uploads mean more engagement, data sales, and ecosystem lock-in. Analysts at Billboard predict a 25% subscriber bump by Q2 2026, especially if viral hits like the platform's past (e.g., Lil Nas X's "Old Town Road") recur.
For marginalized creators—women, POC, LGBTQ+ artists who comprise 60% of SoundCloud's base per a 2025 USC Annenberg study—this levels the field. No more gatekept advances or predatory deals; just tools to build empires. "We're not just streaming; we're stewarding," Herd said in a company blog, though exec quotes remain sparse.
A Symphony of Change: What's Next?
As November's payouts roll out—first 100% checks hitting December—watch for copycats. Will DistroKid drop its unlimited model to zero cuts? Could Spotify counter with fan-direct royalties? SoundCloud's bet is bold: in a $28 billion streaming market dominated by three majors, empowering the underdogs might just rewrite the score.
For now, artists like J. Mira are tuning up. "This feels like the remix we deserved," she says. In an industry often accused of lip-syncing equity, SoundCloud is finally letting creators take the mic—and keep the mic drop.
The previously leaked Behringer BQ-10 analogue sequencer has officially launched — here’s everything you need to know
Written by Sounds SpaceThe previously leaked Behringer BQ-10 analogue sequencer has officially launched — here’s everything you need to know
Behringer has quietly moved another long-rumoured product from the “leak” pile to the shop window: the BQ-10 analogue sequencer. If the name and styling look familiar, that’s because the BQ-10 is a modern reimagining of the classic Korg SQ-10 design — rebuilt with contemporary connectivity and Eurorack friendliness in mind. Whether you’re a vintage-sequencing obsessive, a modular tinkerer, or a bedroom producer who loves hands-on pattern creation, the BQ-10 is designed to slot into many setups without demanding a PhD in patching.
What the BQ-10 actually is
At its heart, the BQ-10 is an analogue, voltage-controlled step sequencer giving you 24 total steps arranged as three 8-step lanes. Channels A and B can operate independently as two 8-step sequences or be combined into a single 16-step melodic lane; channel C remains dedicated as an 8-step control lane for filter, amplitude or other timbral modulations. That flexible three-lane architecture is what gives the BQ-10 a lot of creative mileage: you can have two melodic lines and a separate automation lane running at once, or stitch lanes together for longer phrases.
Physically, it’s presented as a desktop unit that can be removed and mounted into Eurorack — useful if you want both a tidy desktop workstation and rack integration. The layout is classic: knobs and switches for each step, portamento and duty-cycle/gate controls, dedicated sync and CV jacks, plus MIDI for integrating with modern gear.
Key specs and features (quick list)
-
24 analogue steps arranged as 3 × 8-step lanes (A, B, C). A+B = 16 steps when combined.
-
CV and gate outputs for full voltage control; per-lane voltage range switches for compatibility with different synth standards.
-
MIDI & USB — clock sync and MIDI note/Gate output for DAW/hardware integration.
-
Clock generator (internal), adjustable roughly from very slow to very fast — and CV-controllable, plus external clock input for sync.
-
Per-channel portamento (glide) and a global duty-cycle control to shape gate length.
-
Eight playback modes for A/B, including series, parallel, alternating and random options to vary sequence behaviour.
-
Desktop & Eurorack capable — the unit detaches from its desktop case for rack mounting and lists a modest HP size on modular databases.
Those bullet points are the essence of what delivers the classic SQ-10 vibe with modern usability. The BQ-10 keeps the tactile, real-time programming workflow that analogue sequencers are loved for, while making them far less awkward to integrate with modern setups.
How it differs from the Korg SQ-10 (and why that matters)
Behringer’s take is recognisably inspired by the Korg SQ-10, but with a few practical updates. The addition of MIDI/USB and CV-friendly voltage range switches is the biggest functional upgrade — it means the BQ-10 can talk to vintage Hz/V synths, modern 1V/oct modular gear, and computer setups without a lot of external conversion. The dedicated third lane (C) being explicitly pitched for modulation tasks is also a modern convenience: rather than patching an external sequencer to a VCF or VCA, you can build timbral changes into your patterns from the same interface.
There are aesthetic and ergonomic tweaks too: Behringer has chosen a darker control surface and wood end-cheeks for a retro but contemporary look, and the ability to run as a standalone desktop box or snap it into Eurorack opens it up to more workflows than the original hardware typically supported. For many users that “plug and play” modular compatibility is a decisive factor.
Real-world usability: what it’s good at
-
Hands-on compositional jamming. The step pots and physical transport controls invite experimentation — perfect for quickly sketching motifs or improvising live sequences.
-
Modulation-driven patches. Use channel C to twist filter cutoff, VCA level, or any CV-controllable parameter on your synth to create evolving grooves without external CV routing.
-
Integration with DAWs/hardware. MIDI/USB means the BQ-10 can either be the central timing source for hardware rigs or slotted into a software workflow; conversely, the CV outputs let it remain a pure-analogue controller when you want that sound.
-
Modular-friendly sequencing. If you build with Eurorack, the detachable form factor and full CV outputs let the BQ-10 live as a compact step engine in someone’s case.
It’s not trying to be a sequencer DAW replacement; its charm and strength are immediacy, character and hands-on performance — the traits that made step sequencers indispensable in classic electro and electronic music.
Pricing and availability
Behringer has listed the BQ-10 at an accessible price point — around US$149 / £149 in many retailers’ listings — which positions it as an affordable entry into hardware sequencing and modular-adjacent setups. Preorders have appeared across multiple dealers, and Behringer’s own product page and press materials indicate the unit is now officially launched and shipping through retailers in batches. Expect initial demand to outpace supply for a while; popular Behringer launches tend to sell quickly.
Who should consider buying one?
-
Beginners to hands-on sequencing who want a quick, low-cost way to learn analogue sequencing concepts without a full modular investment.
-
Modular users seeking an inexpensive, compact step engine to run melodic and modulation lanes in their rack.
-
Live performers and jam musicians who want a tactile pattern machine that responds well to on-the-fly changes.
-
Producers after character — if you like the slightly unpredictable, squelchy charm of analogue CV and manual knob tweaks, the BQ-10 brings that to low cost.
If your main studio workflow is tightly DAW-centric and you need deep pattern memory, song mode, step automation recall and long-form sequencing, you may find the BQ-10’s analogue, patch-centric workflow less convenient than a dedicated sequencer with full CV-to-MIDI mapping and presets. But for raw hands-on sequencing, it’s hard to beat for the price.
A few practical caveats and community reactions
-
Purists will notice differences between the original Korg SQ-10 and Behringer’s implementation — some step counts and nuanced behaviours vary. That said, many users welcome the modern connectivity, even if small purist gripes appear on synth forums.
-
Supply timing. As with many new Behringer launches, availability may vary regionally, and some retailers show preorder lead times — so if you need one immediately, double-check local stock and shipping windows.
Final thoughts — is the BQ-10 worth the buzz?
The BQ-10 is exactly what it promises to be: a relatively inexpensive, hands-on analogue sequencer that wears its heritage on its sleeve while adding pragmatic modern features. For anyone who’s ever loved the immediacy of knob-per-step sequencing and wanted to plug it directly into modular or MIDI setups without extra interface boxes, the BQ-10 is a compelling option. At its price point and with its combination of CV, gate, MIDI and Eurorack compatibility, it looks set to be a popular choice for curious beginners and seasoned hardware lovers alike.
If you want to dive deeper before buying, check the Behringer product page for full specs and manuals, watch hands-on demo videos from retailers and reviewers to hear how it behaves with your kind of synth, and read user threads for early impressions on build quality and integration tips. Happy sequencing!
Pandora Launches 2 AM: A Bold New Home for DJ Set Culture
Written by Sounds SpacePandora Launches 2 AM: A Bold New Home for DJ Set Culture
In a streaming market saturated with algorithm-driven playlists and on-demand single tracks, Pandora is staking a new claim in the world of electronic music with 2 AM, its brand-new station and DJ takeover series. The idea is simple yet ambitious: to elevate DJ sets as curation, offering exclusive mixes and artist-led programming rather than relying purely on automated playlists. The official positioning? 2 AM is being billed as “the home for DJ set culture on Pandora.” In this post, we’ll dig into what 2 AM is, why it matters, how it’s different from existing DJ/dance streaming options, and what challenges and opportunities lie ahead.
What Exactly Is 2 AM?
At its core, 2 AM is a curated music channel and a rotating “DJ takeover” format in which established DJs handpick tracks, mix them into playlists, and deliver hosted intros or commentary. The first wave of takeovers comes from big names in electronic music — such as Above & Beyond, John Summit, and Don Diablo — who will lend their own sensibilities to the station.
Pandora intends for 2 AM to cover the full spectrum of dance music: from EDM and tech house to nu-disco, chillout, and beyond. The playlists will be refreshed regularly, ensuring a steady stream of new and evolving content.
What sets 2 AM apart from Pandora’s existing dance stations is the emphasis on the DJ as curator, not an algorithm. Rather than simply mirroring what listener behavior suggests, 2 AM intends to reflect what DJs are actually spinning in clubs and festival sets — the tracks that move crowds, not just the ones data suggests you might like.
Why It Matters — For Listeners, Artists & DJs
1. Reconnecting Listeners to the Community and Energy of DJ Culture
One criticism of much digital music consumption is that it strips away context. You might listen to a track, but you don’t hear why it was chosen or how it works in a larger DJ set. With 2 AM, Pandora is attempting to re-embed that context: the intros, the track transitions, the DJ’s narrative flow.
For passionate fans of electronic music, 2 AM offers something more immersive than a static playlist. It’s closer to being at a club, but in your ears. It gives listeners an inside track on what’s trending in dancefloors globally — not just what an algorithm infers from your history.
2. Spotlighting DJs as Creative Curators
Many streaming services downplay the role of the DJ, reducing them to a single “official artist” credit or dropping them from playlists entirely. 2 AM flips that paradigm. The DJ becomes the central creative voice: you listen through them, not just to them.
For those DJs whose work centers on live sets, remixes, or hybrid genre explorations, 2 AM offers a new canvas: curated radio-style shows, unique intros, and perhaps even a way to test unreleased edits or blends.
3. An Avenue for Deeper Discovery
Another benefit of 2 AM is its potential as a discovery engine. Instead of being guided by algorithmic similarity suggestions, listeners might follow a DJ’s personal tracklist and stumble into underground or emerging artists they wouldn’t have found otherwise. This gives Pandora another differentiator in a market awash with “similar to what you already listened to” stale recommendations.
What Makes 2 AM Different from Other Platforms?
DJ-centered content isn’t new — platforms like Mixcloud, SoundCloud, or YouTube host DJ sets and live mixes. But 2 AM offers something that many of those platforms lack:
-
Native integration with Pandora’s streaming infrastructure. No need to bounce between apps; the station lives within Pandora’s ecosystem.
-
High production standards: These aren’t user-uploaded club sets; they’re artist-approved, hosted, and styled for streaming consumption.
-
Scheduled rotation + freshness: Rather than one static mix, 2 AM will continuously refresh with new DJ takeovers, shifting styles, and seasonal changes.
-
Curator identity built in: The DJ is not anonymous. Their personality, voice, and taste become a part of the listening experience.
In short, 2 AM is not just “DJ sets in Pandora” — it’s DJ programming inside a streaming radio experience.

Risks, Hurdles & Considerations
As bold as 2 AM sounds, Pandora will face challenges in making it work sustainably:
Metadata & Rights Management
DJ sets often include a blend of tracks, edits, remixes, or bootlegs. Licensing that material for streaming — especially when included in continuous, curated mixes — can be tricky. Pandora will need robust metadata, royalty accounting, and legal agreements to ensure rights are properly cleared.
Balancing Mass Appeal with Niche Integrity
The broader Pandora audience might not always gravitate toward extended DJ sets or deeper cuts. Pandora must strike a balance between accessibility (making 2 AM inviting for casual listeners) and authenticity (sustaining the respect of die-hard EDM/dance fans). If the mixes lean too “safe,” they risk alienating the core audience; lean too underground, and they might lose mainstream appeal.
Scheduling & Rotation
To keep listeners coming back, Pandora will need to manage scheduling carefully — rotating DJ takeovers, refreshing playlists, and introducing new voices regularly. Stagnation is a risk if the same DJs or subgenres dominate for too long.
Measuring ROI & User Engagement
How will Pandora measure success? Will listener metrics (hours streamed, listener retention) justify the investment in curated DJ content versus algorithmic programming? Pandora will likely monitor how 2AM engagement compares with its existing stations.
What to Watch Going Forward
Here are a few signals I’ll be keeping an eye on:
-
New DJ lineups — will Pandora bring in more underground or rising talent, not just established names?
-
Listener data — how long do users stay tuned? Is 2 AM driving new registrations or retention?
-
Community features — will Pandora add the ability to “follow DJs,” get notifications about upcoming takeovers, or explore past sets?
-
Cross-promotion with events — could Pandora link 2 AM DJ sets to live festival streams or event tie-ins?
-
Artist exclusives & premieres — the ability to debut exclusive remixes or edits through 2 AM could become a talking point and competitive edge.
Conclusion: Why 2 AM Could Be a Turning Point
Pandora’s 2 AM is more than just another streaming channel — it’s a bet on returning agency to the human curators in music. In an era where algorithmic discovery dominates, giving DJs a prominent voice is a statement: dance music is about more than data; it’s about taste, context, and narrative flow.
If Pandora can get the licensing, rotation, and listener onboarding right, 2 AM could become a meaningful bridge between club culture and everyday listening. It offers fans a new way to engage, DJs a new platform to express, and Pandora a fresh point of differentiation in a crowded streaming market.
As 2 AM rolls out and evolves, it could redefine how we experience DJ sets — not just as occasional live events, but as serialized audio journeys you can tune into nightly, weekly, or whenever the mood strikes.
Spotify’s “SongDNA”: A New Era of Music Discovery Through Credits
Written by Sounds SpaceSpotify’s “SongDNA”: A New Era of Music Discovery Through Credits
In a world where music streaming is dominated by algorithms, playlists, and endless recommendations, Spotify seems to be taking a fresh, human-centric approach. Recently, tech insiders discovered that Spotify is developing a new feature called “SongDNA.” This potential update could change how listeners discover music — not just by sound or genre, but by the people behind the songs.
Although the company hasn’t officially announced it yet, clues buried in Spotify’s app code hint at an exciting new direction for music discovery — one that highlights the unsung heroes of every track: the writers, producers, and engineers who make the magic happen.
What Is SongDNA?
SongDNA appears to be a feature in development that will give users the ability to explore music through its credits. Instead of focusing purely on the artist or the song’s title, Spotify’s new approach aims to let users dive into the DNA of a track — the creative contributors who shaped it.
This discovery came from Jane Manchun Wong, a well-known app researcher who frequently uncovers unreleased features in popular apps. Wong found references to “SongDNA” inside Spotify’s code, along with screenshots of what appears to be a new interface that lists each contributor on a track and connects their work across other songs.
For example, if you were listening to a song produced by Mark Ronson, SongDNA might let you tap his name to see all the tracks he’s ever produced — across artists, genres, and eras. Imagine discovering connections between your favorite songs and realizing they share the same songwriter, engineer, or session musician. That’s the kind of experience SongDNA could deliver.
A Shift from Algorithms to Humans
For years, Spotify’s success has been driven by its algorithmic curation — playlists like Discover Weekly or Release Radar use data and behavior patterns to serve you the next best track. But SongDNA hints at a more organic and human discovery path, one rooted in creativity, connection, and collaboration.
It represents a philosophical shift. Music discovery has long been about what you like. SongDNA introduces the idea of who you like — the creative fingerprints behind your taste. Instead of browsing by genre, you could browse by creator network.
This approach could redefine how fans think about their favorite songs. Many listeners don’t realize that some of today’s biggest pop hits were written or produced by the same handful of creative masterminds. For example, Max Martin has written dozens of chart-topping hits for Britney Spears, The Weeknd, Taylor Swift, and Katy Perry. SongDNA could make those invisible connections instantly visible — revealing the invisible threads that tie the music industry together.

Why This Matters for Artists and Producers
The potential benefits of SongDNA go far beyond fans. For songwriters, producers, and engineers, this could be a game-changer.
One of the biggest frustrations in the modern music industry is the lack of visibility for the people who work behind the scenes. While major artists get the spotlight, songwriters and producers often remain anonymous to the average listener — even though their contributions shape the sound of entire generations.
SongDNA could finally bridge that gap. By integrating credits directly into the discovery experience, Spotify would give recognition where it’s due — making it easier for industry professionals to build a public portfolio within the world’s largest streaming platform.
This transparency could also create new networking opportunities. A producer might find a new vocalist to collaborate with. A songwriter could discover others who share a similar creative style. Even music fans could use this feature to better understand the craftsmanship behind their favorite hits.
How SongDNA Could Work
While Spotify hasn’t confirmed any details, leaks suggest that SongDNA might include:
-
A detailed credit breakdown: listing every person involved in the song’s creation — from writers and producers to mixers and mastering engineers.
-
Clickable profiles: allowing users to tap a contributor’s name to see all the other songs they’ve worked on.
-
Connected discovery: recommendations based on shared contributors rather than listening habits — e.g., “Other tracks produced by Metro Boomin.”
-
Interactive visual design: perhaps showcasing “creative webs” connecting different artists and their teams.
Essentially, Spotify could evolve into an IMDB for music — a place where you can trace every creative connection in the industry.
The Competition: Tidal and Beyond
Spotify isn’t the first streaming platform to explore song credits. Tidal, for example, has long championed detailed credit listings, allowing users to view who contributed to a track. However, SongDNA seems to go beyond simple listings — it’s designed to make those credits discoverable.
Instead of credits being buried in a menu, they could become the foundation for new recommendations. This would give Spotify a fresh discovery edge — a more narrative, human layer to its data-driven ecosystem.
It also aligns with a broader industry trend: recognizing creators. From TikTok tagging original sound creators to YouTube’s music metadata updates, platforms are realizing the importance of crediting every contributor. SongDNA fits perfectly into this cultural moment.
Challenges Spotify May Face
As exciting as SongDNA sounds, it won’t be easy to implement. There are a few major hurdles Spotify must overcome:
-
Incomplete metadata – Many songs, especially older ones or indie releases, lack accurate or complete credit data. This could make the feature inconsistent at first.
-
Licensing and data integration – Spotify may need to partner with publishers, PROs, and metadata providers to gather and verify this information.
-
User interface complexity – Displaying all this data clearly and attractively will require careful UX design to avoid overwhelming users.
-
Adoption and awareness – Listeners need to care about credits. While hardcore music fans will love this, casual users may not explore it unless Spotify promotes it effectively.
If the company pulls it off, however, SongDNA could become one of Spotify’s most valuable discovery tools in years.
The Bigger Picture: Rediscovering the Human Side of Music
In a time when AI and automation dominate the headlines, SongDNA feels refreshingly human. It reminds us that music is not just content; it’s a collaboration — a product of emotion, creativity, and teamwork.
By surfacing the people behind the songs, Spotify could spark a cultural shift: encouraging fans to value who made the music as much as who performed it. For upcoming artists, that visibility could translate into recognition, respect, and new opportunities.
If Spotify launches SongDNA, it could redefine how the next generation interacts with music — not just consuming it, but exploring its creative lineage.
Final Thoughts
At the moment, SongDNA remains in development, and Spotify has yet to make an official statement. However, the discovery of this feature in Spotify’s app code — confirmed by multiple sources like TechCrunch and Digital Trends — indicates that the company is serious about exploring this direction.
Whether it launches next month or next year, one thing is clear: Spotify wants to make music discovery more personal, more connected, and more human.
If SongDNA becomes a reality, it could mark the next great evolution in streaming — turning every play into a journey through the creative universe of the artists, writers, and producers who make music come alive.
