General Knowledge

General Knowledge (93)

Ableton Live 12.3 has arrived, and this one’s got a little bit of everything:

AI-powered stem separation, native Splice integration, better bounce workflows, notable Push 3 improvements, and a clutch of smaller but genuinely useful workflow refinements. The update is free to existing Live 12 users, and Ableton is running a limited-time 25% discount for newcomers and Pack purchases, so whether you’re already deep into Live or thinking about joining the ecosystem, now’s a good moment to take stock. 

Below, I break down the features that matter most, what they’ll let you do differently in the studio or on stage, and a few practical tips for getting the most out of the new tools.

The big headline: stem separation built into Live Suite

The flagship addition in 12.3 is built-in stem separation for Live Suite. This isn’t a third-party plug-in or cloud-only service — Ableton has integrated local, offline stem splitting that separates an audio clip into four component stems: Vocals, Drums, Bass, and Others. Practically, that means you can take any sample, loop, or full stereo track and pull it apart inside Live, then remix, rearrange or resample the resulting parts without leaving your DAW. 

Why this matters: Stem separation used to be a workflow that required external services or specialized tools that sent audio to the cloud. Having a fast, local option directly inside Live closes the loop: faster iteration, less context switching, and no file juggling. For remixers, producers who build stems for collaborators, or anyone doing sample-based creativity, the ability to isolate a vocal line or a drum bus without leaving your session is huge. Expect creative uses beyond remixing, to think creative sidechaining, subbing in new drums under a vocal stem, or extracting a texture from the “Others” stem to turn into an ambient pad.

A practical tip: stem separation quality varies with the source material. Clear, well-separated recordings (dry vocals, distinct drum hits) give the best results; heavily distorted or extremely dense mixes may produce artifacts. Use the stems as starting material: resample them, run them through effects racks, and don’t be afraid to combine stems back together after processing.

Splice integration: search, audition, and drop — inside Live

Ableton’s Splice integration is more than a link — it brings Splice’s sample library into Live’s Browser so you can search, audition in sync and key, then drag samples into your project without switching apps. The “Search with Sound” feature is particularly neat: you can capture audio from your set (or drag a clip into the Splice panel) and ask Splice to find samples that fit the rhythms and harmonic content of what you already have. That can turn a friction-filled sample hunt into a fast, creative playground. 

Why this matters: searching for the right sample used to be a deep rabbit hole — dozens of browser tabs, trawling keyword searches, guessing about tempo and key. Native integration means auditioning is instantaneous and context-aware: the samples are previewed in time with your project, so you hear how they groove before committing. For fast sketching and late-night idea sessions, that’s a serious time-saver.



Bounce Groups and smarter offline workflows

Another workflow-focused upgrade is Bounce Groups: the ability to render an entire group (with processing) to a single audio file. This lets you commit CPU-heavy group chains to audio without losing the option to keep your original tracks for later edits — a clean balance between commit-and-mix efficiency and flexibility. It’s the kind of workflow improvement you don’t notice until you need it, then you wonder how you ever lived without it. 

Other audio/bouncing improvements under the hood include faster, more reliable bounce operations and fixes for edge cases on different platforms. If you run large projects or play Live sets that depend on pre-rendered stems, these changes will smooth your workflow and (importantly) reduce last-minute rendering headaches.

Push 3 and hardware improvements

Live 12.3 isn’t just a software update: Push 3 gets form-and-function updates that expand what you can do in standalone and tethered modes. Notably, Push 3 in standalone mode can now work with class-compliant audio interfaces — meaning more ins/outs and a more flexible standalone setup without relying on ADAT tricks. Push’s expressive grid also gets new XY-style control modes and improved step-sequencing with touch-sensitive velocity control, plus a new Rhythm Generator view for drum programming. If you own Push 3, these firmware/software upgrades broaden its standalone studio potential. 

Why this matters for live performers: being able to plug a wider range of audio interfaces into Push 3 without complex routing opens up richer live setups. For producers, the XY mode and improved sequencing make Push more tactile and creative for beat design and expressive performance.

Smaller but meaningful updates

12.3 also brings a handful of thoughtful enhancements that will please power users:

  • Auto Pan → Auto Pan-Tremolo: The Auto Pan device gets more modes and dynamic responsiveness, including dedicated tremolo behavior and level-based response shaping — great for rhythmic modulation and dynamic pumping effects. 

  • A/B states for devices: You can set A and B states for instruments and effects and flip between them easily. That’s a huge boon for sound design and comparative mixing — quickly audition two radically different device settings without losing your place. 

  • New Packs and Max for Live tools: Live Standard and Suite users get new creative devices (Patterns, Sting) and updates to Expressive Chords and Sequencers — useful for generative ideas and getting out of production ruts.

 

These are the kinds of improvements that might not make the big headlines but end up improving daily workflows — faster experimentation, easier comparisons, and more creative starting points.

Performance, platform fixes, and stability

As with most iterative updates, Ableton has bundled a number of stability and compatibility fixes across platforms. Release notes mention fixes for stem separation failures on certain macOS configurations and improvements to the Splice UI behavior, among other bug fixes. The public beta cycle surfaced issues and Ableton addressed several of them before pushing the stable release, which is reassuring for users who rely on Live in critical sessions. 

If you depend on specific third-party plug-ins or unusual workflows, it’s always wise to test 12.3 on a copy of your projects first — don’t overwrite production sessions until you confirm all the plug-ins and setups behave as expected.

Pricing, availability, and the limited-time offer

Ableton is releasing 12.3 as a free update for everyone already on Live 12 (so if you’re on Live 12 Standard or Suite, it’s yours at no extra cost). For newcomers or those upgrading from much older versions, Ableton is running a limited-time promotion: 25% off Live 12 and Packs (and 20% off Push 3 and related hardware) for a short window around the release. If you were sitting on the fence about upgrading or buying in, that’s a practical savings window to consider. 

Who should care — and who might want to wait

  • Remixers and sample-based producers: stem separation and Splice integration are direct wins. Faster sample hunting plus local stem extraction changes the way you can build remixes and reworks.

  • Live performers and Push 3 owners: Push 3’s standalone expansions and Bounce Groups make set preparation and standalone jamming more powerful.

  • Sound designers and experimental producers: A/B device states, Patterns, Sting, and Max for Live updates give fresh sound-design workflows.

  • Users running older or very plugin-heavy projects: test first. While Ableton has patched many issues, complex third-party setups sometimes reveal edge cases.

Quick hands-on tips to get started with 12.3

  1. Try stem separation on a few different source types — a dry acapella, a full mixed track, and a drum loop — to learn the tool’s strengths and limitations. Resample processed stems to hide artifacts. 

  2. Open the Splice panel and experiment with Search with Sound: drag a loop or capture a section, let Splice find matches, and audition samples in sync. The faster previews will change how you hunt for sounds. 

  3. Bounce a group, then compare: use Bounce Groups to commit heavy group processing to audio, then A/B with the original to confirm what you’ve gained or lost. 

  4. Explore Push 3’s new modes if you own one — class-compliant audio support alone opens up new routing possibilities for standalone rigs. 

Final thoughts

Ableton Live 12.3 is a model of how mature DAWs continue to evolve: headline features (stem separation, Splice integration) that grab attention, plus a steady stream of smaller but meaningful workflow and hardware updates. The inclusion of local stem separation is a game-changer for many workflows, while Splice integration tightens the loop between inspiration and production. For Live users, the update is essentially a must-install; for newcomers, the limited-time discount sweetens the deal.

If you use Live in any professional or semi-professional capacity, treat 12.3 as a strong, practical refinement — one that delivers immediate creative tools and sensible performance improvements rather than risky, half-baked experimentation. Fire up a non-critical project, poke around the new panels and devices, and you’ll likely find something that speeds up your process or sparks a new idea.

Suno and Warner Music Group: what the settlement means for artists, AI, and the future of music

This week’s surprising — and, for some, seismic — news that Warner Music Group (WMG) has settled its lawsuit with AI music generator Suno and entered into a commercial partnership marks a turning point in the music industry’s standoff with generative AI. After more than a year of litigation, public debate and fear among creators about the misuse of their work, the two companies announced a deal that aims to balance Suno’s rapid technical progress with protections and revenue for artists. Below, I unpack what happened, what’s in the agreement as reported, the likely consequences for creators and platforms, and what the settlement tells us about how music and AI will coexist going forward. 

The headline: settlement + partnership, not a courtroom victory

At the heart of the story is a straightforward pivot: Warner, which had been litigating against Suno for alleged copyright infringement, is no longer pursuing that courtroom route — instead, the companies have struck a licensing and commercial partnership. As part of the agreement, Suno will phase out its current broad-use models in favour of licensed models, implement new restrictions on downloads (including limiting or blocking downloads for free users and capping paid-user downloads), and introduce mechanisms that let Warner artists and songwriters control whether and how their names, voices, likenesses, and compositions are used on Suno’s platform. Suno also announced the acquisition of Songkick — a live-music discovery brand previously under Warner’s control — as part of the broader transaction. 

That combination — settling the legal claim while creating a commercial path forward — signals both sides’ strategic thinking. Warner gained contractual safeguards and revenue opportunities without protracted litigation; Suno gained industry legitimacy and access to an enormous catalog of artist assets and marketing channels. But the wrinkle is important: the agreement appears to be opt-in for artists. That means the platform can offer realistic “artist-like” outputs only for those creators who choose to participate, while others are (in principle) protected from unauthorized stylistic or likeness use. 

What the agreement reportedly includes (the practical bits)

Reporting across major outlets and Suno’s own announcement gives us a reasonably clear list of concrete changes and commitments:

• Suno will launch new, licensed AI models next year to replace its current open models, designed to operate under licensing terms that compensate rights-holders. 
• Downloads of AI-generated audio will be restricted: free accounts will no longer be able to download songs; paid accounts will have download caps and the option to pay more for higher allowances. 
• Artists and songwriters signed to Warner will be able to opt in (or not) to have their voices, names, likenesses, and compositions used; participating artists will receive compensation as negotiated. 
• Suno acquired Songkick from Warner; Songkick will remain as a fan destination under Suno’s ownership. 

Warner and Suno have not publicly disclosed the financial terms of the settlement, nor the exact revenue split for artists — understandably, these are sensitive commercial details. But the structural commitments are what matter most for industry precedent: an existing AI vendor has accepted licensing obligations, and a major label has accepted a commercial route instead of purely legal enforcement. 



Why this matters: three immediate implications

  1. It normalizes licensing as the dominant path forward. For much of 2024–2025 the dispute between labels and AI companies felt binary: either platforms would be forced to stop using copyrighted content in training and outputs, or labels would license their works. This settlement validates the licensing route. If licensing becomes the industry norm, it means artists (and labels) can directly capture value when their style or likeness is used, rather than relying on uncertain litigation victories.

  2. It sets an opt-in model for artist control. The opt-in approach that Warner and Suno are implementing is significant because it preserves artist agency. Rather than an all-or-nothing ban, artists can choose new revenue streams while controlling the use of their voice and other personal rights. That could become a blueprint for other deals — but it also raises questions about bargaining power and transparency in the offers artists receive. 

  3. It changes product economics for AI platforms. Requiring paid accounts and capping downloads shifts the economics of Suno’s product from an ad/scale-first free model to a hybrid subscription/licensing model. That could slow viral, mass-free proliferation of AI-created tracks, while creating predictable revenue lines that can be shared with rights-holders. It’s a step away from the “wild west” era of unfettered generation. 

The broader legal landscape: one settlement, more work to do

It’s important to remember this is not the end of legal friction. Universal Music Group previously settled with AI platform Udio, and both Sony and Universal have ongoing or recent legal entanglements with AI startups. Suno’s deal with Warner doesn’t automatically resolve disputes with other labels or publishers, nor does it fully answer hard questions about training datasets, informed consent, or derivative uses. The settlement shows parties can negotiate, but it does not replace necessary public policy discussions about how copyright law should adapt to generative models. 

Moreover, artist advocacy groups and creators have voiced concerns about transparency and fairness in earlier label-AI deals. Opt-in programs look good on paper, but their fairness depends on the visibility of contract terms and the bargaining leverage individual artists — especially less established ones — can realistically exercise. Without standardised transparency practices or industry-wide minimums, a fragmented patchwork of label-specific offers could leave many creators at a disadvantage. 

For artists: opportunity and risk

For major, established artists, the Suno–Warner deal represents a new monetization channel and a way to control fan experiences. Imagine paid fan interactions where users can generate licensed remixes or AI-collabs that include an artist’s signature timbre or phrasing — with the artist paid for each use. That can be lucrative.

But for less-visible artists, the risks are real. Labels often negotiate on behalf of roster artists; depending on contract terms, an artist might find their likeness or composition licensed without direct negotiation with them, or receive royalties according to label agreements that artists already critique as opaque. The real question is whether these licensing mechanisms will be transparent and equitable at the artist level, not just at the label level. 

For platforms and startups: a new playbook — at a price

For generative-audio startups, the message is clear: build licensing into the product roadmap early. Models trained on copyrighted works without clearance create legal risk and, as we’ve seen, can end with either an injunction or a negotiated settlement that forces product redesign. For investors and founders, the Suno–Warner deal shows a path to scale that includes paying for rights and restructuring user experiences — but it comes with higher operating costs and a need for careful artist relations.

This outcome also tilts the playing field toward platforms that can shoulder licensing costs and administrative complexity: well-funded startups, major tech companies or ventures backed by institutional capital. Smaller open-source projects and hobbyist tools may find it harder to offer competitive functionality without proper rights clearance. 

Policy and the public interest: what regulators should watch

The Suno–Warner settlement may reduce the immediate urgency for aggressive legislative fixes, but regulators still have work to do. Key policy areas include: clarifying whether training on copyrighted material qualifies as fair use, mandating transparency around datasets, and ensuring a reasonable framework for attribution and compensation. There’s also a consumer-protection angle — users should understand when they are generating content that directly resembles a living artist and what rights attach to that output.

Finally, antitrust considerations could emerge — as major labels and well-funded AI firms stitch commercial relationships, regulators should watch for exclusivity deals that could lock out competitors or limit creative alternatives for artists and fans. 

The cultural question: will AI-made music co-exist with human artistry?

Beyond contracts and code lies a cultural debate. AI music platforms can democratize production, allowing novices to create richer-sounding tracks and fans to experience new kinds of interactivity. But they also risk diluting artistic labor if the market floods with indistinguishable, cheaply generated tracks. The Suno–Warner settlement nudges the ecosystem toward a model where human artistry is recognized as a monetizable input to AI outputs — a compromise that, if implemented fairly, might preserve creative incentives.

That said, commercial deals alone won’t answer aesthetic questions about authenticity and taste. Those will be decided over time by listeners, curators, and creators themselves. If fans value the human story behind songs — the voice, lived experience, performance — then human artists retain a cultural edge. If, instead, novelty and volume dominate streaming economies, artists will need to adapt their business models accordingly. 

Final thoughts: a test case, not the final chapter

The Suno–Warner settlement is a high-profile test case in an industry grappling with technological change. It shows litigation and negotiation can coexist: the threat of legal action pushed an AI startup to the bargaining table, and the settlement produced a commercial framework that may become a template for other deals. But the devil is in the details — the fairness of compensation, the transparency of contracts, and the long-term policy framework will determine whether this outcome is a durable solution or merely a stopgap.

For artists, the takeaway is simple but urgent: engage with these developments, seek clarity about contracts, and insist on transparency. For platforms, the lesson is equally practical: if you want to scale in music, plan to pay — financially and reputationally — to licence the human artistry that makes music meaningful.

We’re in the early chapters of the AI-music story. The Suno–Warner deal doesn’t end the debate, but it moves the conversation from purely adversarial litigation to negotiated commerce — and that shift has consequences for how music will be made, shared and valued in the years ahead. 


Sources: reporting and company announcements from TechCrunch, Suno’s official blog, the Los Angeles Times, The Guardian and Pitchfork informed this analysis. 

If you’d like, I can now:
• Draft a short “artist-facing” explainer summarizing what this deal means for an individual artist (rights, revenue and questions to ask).
• Or create a one-page checklist for independent musicians to protect their work and negotiate AI licensing terms. Which would help you more?

When the majors embraced the machine: what KLAY’s licensing deals with the “big three” mean for music’s AI future

The music industry has been wrestling with artificial intelligence for the better part of this decade — first in courtrooms and headlines, then in negotiation rooms and press releases. This week’s news that Los Angeles–based AI music startup KLAY Vision Inc. has struck licensing deals with the three major record groups — Universal Music Group, Sony Music (and their publishing arms), and Warner Music Group — marks a clear inflection point. For the first time, the companies that own the lion’s share of recorded and published music have formally licensed their catalogs to an AI-driven streaming product intended to let fans remake songs in new styles. That’s business-as-usual turned on its head: the record industry is no longer only defending copyright against automated imitators — it’s now actively building a commercial framework to monetize AI creativity. 

Below, I unpack what the KLAY deals actually are, why they matter, the unanswered questions they raise for artists and listeners, and what they suggest about how the music business will adapt to — and profit from — generative audio technologies.

What KLAY says it does (and what the labels licensed)

KLAY bills itself as a music-technology company building a streaming product with integrated AI tools that let users remix and “remake” songs. The startup says it trained its systems on thousands of licensed recordings and compositions, and that the agreements with Universal, Sony, and Warner grant KLAY the rights to use those catalogs both to power its machine-learning models and to enable listeners to create personalized AI variations of the original songs. In press materials and label statements, the emphasis is on a paid, controlled system that compensates rights-holders and gives participating artists “control” over how their work is used. 

Put simply: instead of AI engines scraping the internet for music (and getting sued for it), KLAY has secured the catalogs it needs officially — licensing the underlying recordings and publishing rights so it can legally train models and offer derivative experiences to subscribers. This is the fundamental pivot from conflict to commerce.

Why the labels decided to license, not litigate

The move reflects a broader cooling of the industry’s initial, lawsuit-heavy reaction to AI-generated music. Early skirmishes — including high-profile cases targeting companies accused of training models on copyrighted recordings without permission — exposed how difficult it is to police every synthetic output on major platforms. Licensing gives labels an avenue to recapture value from AI activity rather than leave it to unregulated creators and deepfake projects.

From a business perspective, the labels are treating AI as a new revenue stream: licensed models can be monetized through subscriptions, micro-payments for individual remixes, and potential new sync/derivative licensing arrangements for commercial use. By agreeing on terms now, labels lock in payment frameworks and control mechanisms rather than ceding the field to adversaries. Reuters and other industry outlets reporting on the deals note that this is the first time a single AI company has landed agreements across all three majors — a symbolic and practical milestone. 

What this could mean for artists and songwriters

The labels stress payments and artist protection, but practical outcomes will depend on contract details that aren’t yet public. Important questions include:

  • Revenue split mechanics: Will revenues from AI-remixes be shared with performers, songwriters, producers and publishers in the same way as traditional streams, or will there be a bespoke split for AI-derived works? Early statements promise “proper recognition and reward,” but the breakdown — and whether smaller artists get a fair share — is not yet transparent.

  • Opt-in vs opt-out: Some label statements imply artist participation will be managed through label agreements, but will individual artists be able to opt out of allowing their work to be used by AI remixes? Opt-in models (where artists explicitly authorize usage) are friendlier to artist autonomy; blanket label-level licensing is administratively simpler but risks leaving artists surprised by new forms of derivative use.

  • Moral rights and voice cloning: Even when a song is licensed, voice impersonation or highly realistic vocal “clones” of living artists raise unique reputational and ethical concerns. Will artists be able to veto impersonations that sound like them, or require additional consent for likeness-based uses? The deals suggest labels will attempt to govern these uses, but public-facing safeguards remain to be seen. 

In short: licensing opens the door to new income streams, but it also creates novel bargaining points and potential pitfalls. Musicians and their managers should read the fine print closely as product specifics unfold.

How KLAY’s product model could work in practice

Based on available reporting, KLAY appears to be positioning itself as a hybrid of a streaming service and an interactive creative tool. Imagine a subscription app where a fan can pull up a favorite track and choose stylistic options — tempo, instrumentation, genre twist, vocal emphasis — and the AI generates a personalized version on the fly. That output would be playable inside the app, shared within the platform, and possibly downloadable or licensed for other uses depending on rights negotiated.

Business models that have been floated by industry commentators include:

  • Subscription-first: paying for access to the AI remixing engine and licensed catalog (similar to an advanced streaming tier).

  • Per-use micropayments: small fees for individual remixes or for the right to download/share an AI-created version.

  • Creator tools and marketplace: artists and producers could sell presets, commercial licenses, or custom remixes created with the platform.

All of these require plumbing to allocate money back to the correct rights-holders and to catalogue derivative uses for accounting — a nontrivial metadata and royalty-engineering challenge. 

Risks and friction points

While the deals are a big step toward normalization, several hazards remain:

  1. Quality dilution and platform clutter: Generative tools can flood platforms with low-quality, AI-generated tracks masquerading as new music. That trend has already raised concerns on existing streaming services where “AI spam” has pushed down discoverability for human artists. A licensed approach may mitigate some of that, but discovery algorithms and curation will matter. 

  2. Deepfake vocals and likeness rights: Even with catalogs licensed, using an artist’s vocal likeness without separate consent is a distinct legal and ethical issue. Platforms and labels will need enforceable rules about vocal impersonation and identity-protecting mechanisms.

  3. Creative credit and attribution: How will AI-remixes credit the original artist, songwriter, and the machine model? Proper attribution standards are crucial for maintaining transparency and preserving the lineage of creative works.

  4. Inequality of bargaining power: Major-label artists and top-tier songwriters are likely to get better negotiated terms than independent or legacy artists signed on older deals. Without industry-wide standards, smaller creators could be left behind.

Why is this also an opportunity

For listeners, platforms like KLAY could unlock new ways to interact with music: personalization, reimagined versions of favorites, and collaborative experiences where fans co-create with AI. For creators, licensed AI can serve as an additional income source and a promotional tool: remixes can introduce songs to new audiences, and interactive experiences can enhance fan engagement.

More broadly, the deals hint at an industry strategy: partner with AI entrepreneurs and formalize revenue pathways rather than attempting to shut down innovation entirely. That approach promises to steer AI development toward mutually beneficial products — if the commercial terms are fair and transparent. 

What to watch next

If you’re tracking this story, here are the developments that will matter most in the coming months:

  • Public launch and product UX: How KLAY implements artist controls, attribution, and payment flows in its actual app will show whether the deals were symbolic or substance-driven. A consumer launch will reveal whether listeners embrace remixable subscriptions. 

  • Artist reactions and opt-ins: Watch for statements or actions from major artists about whether they authorize voice-cloning-like features for their work, and whether there’s an opt-out movement among musicians uncomfortable with the new model.

  • Regulatory scrutiny and industry standards: Governments and industry bodies may step in to set guidelines around deepfakes, voice rights, and transparent royalty reporting. The deals with the majors could influence those standards.

  • Extensions and competition: Other AI firms and major labels may ink similar deals, or the majors might partner with multiple AI players. Competition could be healthy — pushing platforms to improve artist protections and user experience.

Final take: pragmatic adaptation, with caveats

KLAY’s licensing deals with Universal, Sony and Warner represent a pragmatic shift: the music industry is no longer fighting the generative-AI tide by default. Instead, it’s trying to steer that tide into monetizable channels. For fans and creators, that can be a win — new experiences and revenue streams — but the outcome depends on how equitably and transparently the underlying systems are built.

This moment is a reminder that technology doesn’t just disrupt markets: it forces social and legal frameworks to catch up. Licensing the catalogs was step one. Step two is delivering a platform that truly balances innovation, artist rights, and fair payment. If the majors and KLAY can design that system together — with clear artist consent, transparent accounting, and sensible limits on voice impersonation — the music ecosystem may emerge more diverse and more lucrative for creators. If not, the next chapter will be courtrooms and headlines all over again.

Either way, the industry’s choice to collaborate with — rather than simply contest — AI companies signals that generative audio is here to stay. The crucial question now is who writes the rules, and whether those rules will protect the people who make the music.

“Music producers are rejecting AI”: what the new studies tell us and what it means for music’s future

Brother — the headlines have been loud this week: a new survey says over 80% of music producers are actively against AI-generated songs, while separate listener-focused research shows most people can’t tell AI-made tracks from human ones. That collision — creators rejecting the tech at the same time listeners can’t reliably detect it — is the perfect storm for arguments about ethics, jobs, creativity, and regulation. Below, I’ll walk you through the findings, explain why they matter, and give a blunt take on what could happen next. 


The headline findings:

• A global survey of music producers, commissioned by Tracklib, reports that a very large majority of producers are sceptical or hostile to generative-AI use in music: adoption is low and rejection high — the press release frames it as “music producers are rejecting AI,” with only a small percentage actively using generative tools. 

• At the same time, Deezer and Ipsos ran listening tests and found that ~97% of people couldn’t reliably tell AI-generated tracks from human-made ones in blind tests — a striking demonstration that AI music quality has reached near-parity for the ear of the average listener. That same study shows big public support for transparency (label AI-made tracks) and concerns about ethics and copyright. 

• Major outlets reporting on these themes also point to real-world effects: AI tracks are already rising on streaming platforms and even some charts, which is intensifying industry debate about promotion, royalties, and verification. 


What the producers’ rejection actually means

When a study says “80%+ of producers are against AI,” don’t read that as a single emotion — it’s a mix of reasons:

  1. Artistic integrity and craft. Many producers see production choices as part of artistic identity. Machines that spit out passable arrangements or vocal lines feel like a shortcut that removes the human voice from the process. Tracklib’s survey captures that sentiment: a lot of rejection is because producers feel AI undermines the craft. 

  2. Quality concerns — and nuance. Some producers reject full-song generation but still use AI for small tasks (like reference stems, mastering presets, or chord suggestions). The “rejecting AI” headline often conceals that many are open to assistive AI, just not to AI that replaces core creative decisions. 

  3. Copyright and fairness. Producers worry about training data (AI trained on copyrighted music without consent), displacement, and how royalties should be split if models borrow from identifiable artist outputs. The Deezer/Ipsos work also shows public support for ethical training and labeling — so this is both a creator and consumer issue. 

  4. Economic risk. If streaming platforms allow AI tracks to surge without clear rules for attribution and payment, session work, ghost production, and publishing could lose income. That’s why producers are defensive — it’s their livelihood, not just ideology. 


Why listeners’ inability to tell matters

The Deezer/Ipsos finding that ~97% of listeners failed to detect AI tracks is a game-changer for three reasons:

  1. Market-level indistinguishability. If audiences can’t tell, then AI tracks can compete on the same platforms, playlists, and algorithms. That puts human artists in direct competition with low-cost synthetic output unless platforms treat them differently. 

  2. Transparency demand. The public’s unease in that study shows people want to know. If platforms don’t label AI content, they risk trust erosion and regulators stepping in. Deezer has already started tagging AI submissions — an early institutional response.

  3. Policy & chart implications. Charts and awards rely on attribution; if synthetic works crowd charts, the industry must decide whether to treat them equally. We’re already seeing cases of AI songs making chart noise and being removed or disputed. That bubbling conflict is the reason labels, distributors, and streaming services are scrambling. 


The contradictions: why both things can be true at once

It seems paradoxical: producers reject AI while listeners don’t care or can’t detect it. But both can be true because they’re answering different questions. Producers are speaking about process, ethics, and livelihoods; listeners are responding to end-product enjoyment. The industry is now forced to reconcile process (how music was made and who benefits) with product (what people enjoy or stream).


The pressure points: where conflict will be decided

Here are the battlegrounds we should watch:

  1. Streaming platform policies. Will platforms require AI labels? Will they remove AI tracks from editorial playlists? Deezer’s tagging is a blueprint; other platforms may follow or resist.

  2. Copyright litigation & licensing. Lawsuits are already happening and will accelerate. Courts will decide whether training models on copyrighted works is infringement, and that ruling will determine cost structures for AI music firms. 

  3. Royalties & metadata standards. If AI tracks use samples or stylistic mimicry, who gets paid? The industry needs metadata standards and traceability to assign rights and royalties correctly — otherwise, payments and charts will be gamed. 

  4. Consumer-facing transparency. Mandatory labelling is likely to be politically popular and technically feasible. Expect regulators to push for clarity so consumers know whether they’re listening to AI. 

  5. Artist and union pressure. Creators’ associations and unions will push for protections, training data consent, and revenue-sharing terms. If producers maintain unified resistance, they can shape licensing frameworks more than tech companies expect. 


The likely near-term scenarios (what I actually expect)

  1. Hybrid adoption increases. Most producers will continue using AI as assistants (mixing tools, stems, mastering suggestions) rather than full-song generators. The “pure AI” artist will mostly be an outlier or novelty that occasionally breaks through virally. 

  2. Platform responses: tags + restrictions. Platforms will adopt transparent tagging and may restrict AI-only tracks from editorial playlists and algorithmic boosts until legal/ethical frameworks are set — Deezer is already an early mover here. Expect other major services to follow to avoid reputational and legal risk. 

  3. Regulatory and legal clarifications. Courts and regulators will clarify what “training on copyrighted works” means. If rulings favor copyright holders, AI music firms will need licensing deals akin to the way streaming licenses catalogs, which raises costs and reduces the free-for-all. 

  4. An uneven marketplace. We’ll end up with two flavors of music commerce: human-made content protected, promoted, and monetized under traditional deals; and AI-generated content that lives in a parallel space with different rules, visibility, and monetization models. Chart systems may split or flag AI entries. 


What this means for producers, labels, and listeners: practical takeaways

For producers:

  • Lean into what makes you uniquely human: emotional risk-taking, imperfections, personal storytelling, and brand as a creator. These are the qualities AI struggles to authentically replicate.

  • Treat AI as a tool, not a competitor. Learn to use assistive AI for workflow speedups (mastering, stem separation, idea generation) while protecting your signature creative decisions. 

For labels & distributors:

  • Invest in metadata standards and provenance systems now. If you don’t capture who did what — and whether a track used AI — you’ll lose control over chart inclusion, payouts, and legal defenses.

  • Think about curated playlists as a premium human-made space; reserve some editorial real estate for verified human artistry to protect brand value.

For listeners:

  • Expect transparency features (filters, labels, opt-outs). If you want only human-made music, the tools to enforce that preference are likely coming. 


Final thought, this is a crossroads, not an end

AI music is no longer a futuristic demo or toy. The technology is good enough to matter commercially and culturally, and the industry’s response will shape careers and revenue flows for a generation. Producers’ concern is legitimate — it’s about artistry and survival. Listeners’ reaction is equally important: they show curiosity but also want honesty.

The healthy path is a negotiated one: transparency, licensing, and new business models that protect creators while allowing innovation. If the industry can’t agree, regulation will force the issue — and that could either protect creators or ossify the market. Either way, producers rejecting AI today is a loud signal: the music world wants rules before mass adoption. That moment of negotiation is the real story behind the headline.

Suno raises $250M — what this means for music, makers, and the industry

Big money is chasing big ideas. AI music startup Suno just closed a $250 million Series C round at a reported $2.45 billion post-money valuation — a jump that underlines investor faith in generative audio, and at the same time stokes every debate happening around AI and creative work right now. This isn't just another VC headline: it has real consequences for artists, labels, startups and anyone who cares about how music gets made, credited, and monetized. 

Belo,w I break down the news, why investors are writing huge checks, the tensions and lawsuits now in play, and what musicians and music businesses should actually do next.

The headlines (fast)

  • Suno raised $250M in a Series C led by Menlo Ventures, with participation from groups including NVIDIA’s venture arm and others. The round reportedly values the company at $2.45 billion

  • Published reporting indicates Suno’s growth has been rapid (user growth and monetization metrics are being cited in industry coverage) — some outlets claim high revenue and quick adoption. 

  • At the same time, Suno is entangled in copyright challenges and criticism from musicians and labels about how generative music models were trained and how outputs resemble existing works. The legal and cultural fights are unresolved. 

Why investors doubled down — a simple math + product story

Investors put money where they see leverage: a product that can scale rapidly with relatively low marginal cost, a big addressable market, and defensible tech or distribution. Suno checks several boxes:

  1. Mass-market potential. Text-to-music and AI-assisted songwriting promise to democratize music production: people who never touched a DAW can generate a song from a prompt. That expands the total market of music creators — and therefore potential paying customers, creators who want tool subscriptions, and businesses that need licensed music. 

  2. Platform economics. If Suno can combine an easy generator with a marketplace, publishing or licensing deals, it can earn recurring revenue on subscriptions, commercial licenses, and partnerships (e.g., plugin bundles, enterprise licensing for games, ads, content creators). Multiple recent write-ups highlight Suno’s strong monetization trajectory, which helps justify a big valuation. 

  3. Strategic investors. With participation from deep-tech and media investors, Suno gains both capital and routes into key partners — think distributed compute (NVIDIA), media relationships, and scaling channels. Those backers see AI audio as infrastructure that will be embedded across content pipelines. 

Put together, the logic is: invest now to own the interface and distribution layer for AI music before incumbents (DAWs, labels, platform giants) lock it down.

But there’s a huge “but”: copyright, ethics, and cultural pushback

Money buys growth, not goodwill. The same scaling that makes Suno attractive to investors also amplifies the sticky problems everyone warned about.

  • Lawsuits and label pushback. Major labels and some artists have filed or threatened legal action alleging Suno trained models on copyrighted recordings without adequate consent or compensation. Those suits could force licensing deals, statutory damages, or changes in how models are built and monetized — each of which would hit margins and product claims. 

  • Creative legitimacy. Critics argue that clicking “generate” is not the same as composing or performing — and that mass-producing music risks diluting cultural value and displacing paid work. Defenders say Suno’s tools augment creators and open new pipelines for musical discovery. That debate isn't just philosophical; it affects regulation, platform policy, and public perception.

  • Platform policy risk. Streaming platforms and distributors are already experimenting with AI content rules. If major streaming platforms restrict AI-generated content or demand specific metadata/labels and licensing, Suno’s commercial playbook may have to adapt quickly. 

 

What will the $250M will likely fund?

Sizable rounds don’t just cover payroll. Watch for:

  • Product R&D at scale. Faster models, higher-quality vocals, longer-form music, and tools that enable customization and stems for mixing. Suno will invest in making outputs sound more polished and controllable — a route to premium tiers. 

  • Legal and licensing teams. Expect Suno to beef up legal resources and negotiate licensing deals with publishers/labels — whether proactively or under pressure. That’s expensive but strategically necessary if they want to serve commercial customers.

  • Distribution and enterprise partnerships. Integrations with DAWs, game engines, ad platforms, and content platforms. With partners like NVIDIA (or NVentures’ involvement), cloud and inference costs could be optimized. 

  • Creator economy features. Marketplaces, revenue shares, creator profiles, and tools for human+AI collaboration that let musicians add their voice and retain rights — those features can help reduce backlash and create a community that Suno can monetize. 


How this affects musicians, producers, and indie labels

Short answer: the world changes, but you can steer your outcome.

  • New income streams, if you adapt. Musicians who learn to use AI as a co-producer, who sell curated stems, or who offer performance/production services on top of AI demos can create new revenue models. Tools that reduce routine work free time for higher-value creative tasks. 

  • Competition for stock music and low-end beats. Expect price pressure when it comes to commoditized background music and simple beats. If you sell simple loops or generic beats, you’ll feel it first. Level up your offerings: unique sounds, signature production, bespoke performances. 

  • Negotiation leverage with platforms. As AI content grows, musicians and rights-holders should demand better metadata, attribution, and share of revenue when models or platforms use their material. Collective bargaining, unionization, or industry-wide licenses may become necessary. 

  • Credibility and branding matter more. Human-driven musical artistry becomes a differentiator. Fans will value authenticity, live performance, and artist narratives. Musicians who emphasize craft and story can maintain premium positioning.

What labels and publishers will do and why it matters:

Labels don’t want to be sidelined. Expect:

  • Licensing deals and partnerships with AI vendors — labels get paid and retain control over how recordings are used. That's a comfortable outcome for rights-holders if the economics work. 

  • Litigation as leverage. Suits can force settlement talks or shape regulation. Even if litigation is slow, it creates a bargaining chip and public pressure. 

  • New product offerings. Labels might launch their own AI tools or partner with platforms to offer branded, safe AI sounds — turning risk into opportunity.

Regulatory and platform questions to watch

  • Transparency mandates. Will platforms require AI-generated tracks to be labeled? How will attribution be tracked across distribution chains?

  • Compulsory licensing vs. negotiated deals. Regulators could push for clear rules on training data rights and compensation (analogous to sample-clearing regimes). That would change unit economics for every AI music company.

  • Data rights and model audits. Governments or industry bodies might demand audits of training datasets or new standards for demonstrable provenance.

Regulation is messy and slow — but high valuations like this make regulation inevitable, as the stakes are significant.

Practical takeaways for creators and music businesses

If you’re a musician, producer or indie label:

  1. Experiment with AI but keep control. Use Suno and other tools to prototype quickly, but always decide which outputs become public and how you protect your brand and rights.

  2. Treat AI as a collaborator, not a replacement. Use it to sketch ideas, expand arrangements, or create stems — then add your unique performance, production or vocal signature.

  3. Document provenance and contracts. When licensing music, be explicit about AI usage in contracts and distribution metadata.

  4. Learn to monetize beyond streaming. Offer stems, bundles, custom production, licensing for indie games/ads, or educational content. Diversify income.

  5. Get savvy about IP. If you care about how models use existing works, follow industry litigation, and consider joining collectives lobbying for fair terms.


Final thought: The music ecosystem will adapt

Suno’s $250M raise is a landmark signal: investors expect AI to be central to the next wave of music creation and distribution. That will bring both opportunity and friction. Artists who adapt, insist on fair terms, and leverage the new tools creatively can thrive. The ones who don't will face a reshaped marketplace where commoditized sounds are abundant and differentiation is everything.

This moment is a crossroads. Massive capital accelerates capability; the public, artists and policymakers will decide the boundaries. For musicians: learn the tools, protect your craft, and keep making music that only a human can make — because even in a future of perfect algorithms, humans will still be the ones who turn sound into meaning.

Serato DJ: A host of new Crate-management options lead its 4.0 update

Serato has just rolled out the biggest library overhaul in years. With the 4.0 update — released as Serato DJ Pro / Lite 4.0 and promoted through a public beta — the company has refocused its attention away from flashy effects and toward something DJs quietly crave: better ways to organize and access music. The result is a smarter, faster, and more flexible library system built around powerful crate management tools that make prepping sets and finding the right track on the fly far less frustrating. 

Below, I unpack the most important additions, why they matter, and how DJs of all stripes can use the new library features to speed up workflows and boost creativity.


What’s new? The highlights

While 4.0 includes a variety of UX and performance tweaks, the marquee improvements center on crates — Serato’s core organizational unit. The most important additions are:

  • Crate Search — quickly find the crate you need by name. No more scrolling through long lists. (Pro-only feature). 

  • Favorite Crates (Shortcuts / Pinning) — pin go-to crates to the top of your list for instant access during a set. 

  • Crate Color—color—code crates using a right-click palette for visual separation and faster scanning. 

  • Crate Sort — sort crates by alphabet, date created, or arrange custom orders. 

  • Show Track in Crates — right-click any track and instantly see every crate the track lives in. Great for checking overlaps and curating transitions. 

  • Streaming Tracks in Crates — add tracks from streaming services into crates alongside local files, reducing the friction of mixed-source sets. 

  • Track Rating (emoji or stars) — rate tracks in the library using stars or playful emojis (watermelons, ghosts, etc.) for fast recall. 

  • Crate Info — see crate metadata like total track count, total runtime, and file size in the status bar. Useful for timing sets. 

  • Resizable Album Artwork & UI polishing — resize artwork in the library view, improved file relocation, safer USB ejection, and automatic analysis on import. 

These features combine to make Serato’s library both more informative and more manipulable — two qualities that add up to real time saved during prep and performance. 


Why crate management matters

Organizing music isn’t glamorous, but it’s foundational. A DJ’s library is the toolbox for every set; when tools are poorly organized, the performance suffers. The new crate system attacks several long-standing pain points:

  • Speed of access: Crate Search + Favorites removes the need to dig through nested folders or remember obscure crate names. In a busy club or festival, shaving seconds from a search can change the course of a set. 

  • Visual scanning: Color coding and resizable album art let DJs build a visual language for their library — e.g., blue crates for house, red for peak-time weapons, green for vocal edits — speeding recognition when adrenaline’s high. 

  • Context awareness: “Show Track in Crates” and Crate Info remove guesswork. Want to know if that remix is already in another playlist or how long a crate will last on the dancefloor? Now you can know instantly. 

  • Bridging streaming + local files: Being able to store streaming tracks in crates reduces the friction when you mix paid downloads and on-demand tracks. It simplifies set building in hybrid workflows. 

Put succinctly: better crate tools mean less cognitive load during prep and performance, and more headspace for creative decisions.





Practical workflows — how to use the new feature 
immediately.

1. Build a “go-to” rack: Create favorite crates for different parts of your set (warmup, peak, halftime, closer) and pin them. During a set, jump straight to the correct vibe without hunting. 

2. Color code by energy & key: Use crate colors to encode energy levels or harmonic compatibility. For instance, green = compatible in key, orange = high-energy, blue = vocal tracks. Visual cues speed selection under pressure. 

3. Use Track Rating as a preflight checklist: During practice, tag tracks with an emoji or star rating to indicate readiness: 5-star = performance-ready, 3-star = needs edits or cueing, 1-star = archive. Filter by rating when compiling a set. 

4. Merge streaming & local crates for gig-specific builds: Add your subscription service tracks into crates with your local edits — then use Crate Info to verify runtime and track counts for set length planning. 

5. Prep faster with Analyze on Import: Switch on automatic analysis so newly added tracks are ready to beatmatch and use stems/features without manual scanning. This is handy when hot-swapping files from USB drives. 


Comparison: how Serato’s changes stack up

Competitors like Rekordbox and Traktor have offered various library conveniences for years. What’s notable about Serato 4.0 is the breadth of small, practical quality-of-life improvements delivered together — the kind of iterative polish that DJs notice in the field. Features like “Show Track in Crates” are reminiscent of functions some rivals implemented earlier, but Serato's approach to bringing multiple improvements at once (color, ratings, pinning, crate metadata) means you get a cohesive, modern library experience rather than a single standout tool. 


Potential fixes & what to watch

No big software release is flawless. Early public beta threads point to a few things DJs should watch:

  • Crate Search in Lite: Some features (like crate search) are Pro-only. If you’re on Lite, check which tools you’ll miss before updating. 

  • Library merging & external drives: while file relocation and hot-swap behavior have improved, DJs with complex external drive setups should test the new automatic merging carefully before relying on it in live shows. 

  • Workflow muscle memory: any library overhaul may temporarily slow DJs who are used to older shortcuts — but the tradeoff is faster, clearer workflows after a short adjustment period.

Serato has emphasized community feedback during the public beta, so many small issues will likely be patched quickly. If you rely on Serato for gigs, don’t upgrade mid-gig season without confirming stability in your specific setup. 


Pricing and availability

Serato released 4.0 as a public beta initially, with the company listing the update and downloadable builds on its site. Serato DJ Pro remains a paid product with purchase and subscription options, while Serato DJ Lite continues to be free — though some of the new features are Pro-only. Check Serato’s official downloads and release notes for the exact availability and the latest stable build. 


Final thoughts — a library update that matters

Serato DJ 4.0 isn’t headline-grabbing because it adds a new effect or a flashy performance feature. It matters because it addresses the mundane, everyday work of DJs — organizing, finding, and prepping music — with practical tools that shave time off routine tasks and reduce friction in the creative process. For working DJs, producers who moonlight as selectors, mobile DJs, and club residents, those minutes reclaimed from searching and guessing add up to better sets and less stress.

If you’re a Serato user: test the public beta in a safe environment, try the crate color and favorites system, and begin using the rating system to build a library taxonomy that works for you. If you’re still on another platform, Serato 4.0 narrows the gap in library ergonomics and gives you one more good reason to reevaluate your DJ software workflow.

SoundCloud's Game-Changing Leap: Artists Claim 100% of Royalties in New Era of Creator Empowerment

By Elena Vasquez, Music Business Daily, November 14, 2025

In a seismic shift that's sending ripples through the music industry, SoundCloud has unveiled a subscription model that hands artists the full reins on their earnings. Announced on October 30, 2025, the platform's new "All-In-One Artist Subscription" allows independent creators to retain 100% of royalties from music distributed across more than 60 streaming services, including Spotify, Apple Music, TikTok, and YouTube Music. No more 20% cuts from the distributor—every penny from streams, downloads, and syncs goes straight to the artist. This isn't just a tweak; it's a declaration of war on the opaque royalty structures that have long plagued emerging talent.

For years, SoundCloud has been the digital sandbox where bedroom producers and viral hopefuls test their sound, amassing over 175 million tracks and a fiercely loyal community of "music obsessives." But monetization? That was the rub. Under the previous SoundCloud for Artists Premier plan, creators kept about 80% of distribution royalties after the platform's slice, a model that mirrored competitors like DistroKid and TuneCore but still left many feeling shortchanged in an era where a single stream nets less than a tenth of a cent. Now, with this update effective for payouts starting November 2025, SoundCloud is betting big on transparency and artist-first economics to reclaim its throne as the go-to hub for indie music.

The timing couldn't be more poignant. As major labels consolidate power amid AI-driven disruptions and TikTok's algorithm-fueled virality, independent artists are squeezed harder than ever. A 2024 MIDiA Research report pegged the average indie artist's annual streaming revenue at just $1,200—barely enough for a month's rent in Brooklyn. SoundCloud's move feels like a lifeline, especially for the platform's core demographic: Gen Z and millennial creators who upload demos, remixes, and EPs without the safety net of a label advance.


From Uploads to Uplifts: SoundCloud's Evolution

SoundCloud's story is one of scrappy innovation turned industry staple. Founded in 2007 by Swedish entrepreneurs Alex Ljung and Eric Wahlforss, it started as a simple audio-sharing site for sound designers and DJs. By 2013, it had exploded into a discovery engine, launching careers like those of Billie Eilish (whose early tracks racked up millions of plays) and Post Malone. But growth brought growing pains: a 2017 funding crunch nearly sank the company, forcing it to pivot toward monetization tools like repost chains and basic royalties.

Enter SoundCloud for Artists in 2018, which bundled distribution, analytics, and promo tools into tiered plans. The Premier tier, at $99 annually (or about $8.25 monthly), promised 80% royalties on distributed tracks—a step up from free uploaders but still trailing the "zero-cut" dreams peddled by upstarts like Amuse. Artists grumbled about hidden fees, payout delays (net 60 days), and the platform's infamous "repayable advances" that could claw back earnings if streams tanked.

Fast-forward to 2025, and SoundCloud's under new stewardship after a 2023 acquisition by SiriusXM, which injected $200 million into creator tools. CEO Michael Herd, a former Spotify exec, has steered the ship toward "sustainable independence," citing data that 70% of SoundCloud's 40 million monthly users are undiscovered artists seeking direct fan connections. The All-In-One plan is the culmination of this vision: for that same $8.25 monthly fee, artists get unlimited distribution, built-in monetization, and zero platform cuts on royalties, Fan Support tips, or merch sales.

Digging into the fine print reveals a model built for scale. Tracks uploaded via the plan beam out to 60+ DSPs (digital service providers) instantly, tapping SoundCloud's 76 million global users for immediate plays. Royalties? 100% artist-retained, with only a nominal payout processing fee (as low as $0.50 per transaction via partners like PayPal or bank transfer) once earnings hit $25. No upfront costs for vinyl-on-demand releases—artists can press limited-edition runs without bulk minimums—or for linking merch stores, where SoundCloud takes zilch from sales.

It's not without caveats. The 100% share applies only to new distributions post-November 2025; legacy earnings might still reflect the old 80/20 split. And while SoundCloud touts "no hidden fees," tax forms and two-factor authentication are mandatory for payouts, a nod to fraud prevention amid rising cyber threats in creator economies. Still, for a Nashville trap producer like DXRKNOVA, who credits the platform for doubling her monthly earnings through targeted promo, it's a no-brainer upgrade.


The All-In-One Blueprint: Features That Fuel Futures

At its heart, the All-In-One Subscription is a Swiss Army knife for the solo artist hustler. Distribution is the star: upload once, and your track hits Spotify playlists, Apple Music's algorithmic feeds, and even niche spots like Deezer or Tidal. SoundCloud's edge? Its proprietary "Instant Reach" algorithm, which funnels new releases to superfans and tastemakers within hours, potentially spiking plays before competitors even index the file.

Monetization layers on seamlessly. Fan Support lets listeners tip directly—every cent to the artist, displayed as a "top supporter" badge on profiles to foster loyalty. Merch integration is plug-and-play: link your Teespring or Bandcamp store, and SoundCloud's storefront widget handles discovery with zero fees. For physical die-hards, the vinyl-on-demand tool turns digital hits into 7-inch pressings, shipped globally without inventory risks. "It's like having a label without the middleman," says Brooklyn DJ nextdimensional, whose EP Portal saw a 40% revenue bump after bundling streams with fan tips and limited-edition merch.

Analytics round out the package: real-time insights on listener demographics, stream sources, and even AI-powered "mood mapping" to suggest collab opportunities. At $99/year, it's competitively priced against DistroKid's $22.99 annual unlimited plan (which takes 0% but lacks built-in audience tools) or TuneCore's $29.99/single (with varying cuts). SoundCloud's pitch: Why fragment your workflow when one dashboard handles discovery, distribution, and dollars?

Critics might quibble over the subscription lock-in—free users still face upload limits and no distribution—but for pros, it's a steal. Early adopters report 15-20% earnings lifts, per internal SoundCloud data, thanks to the platform's 320 million tracks-strong library acting as a perpetual promo engine.


Voices from the Upload Trenches: Artist Reactions Pour In

The announcement lit up X like a viral remix. "SoundCloud the MVP baby 100% Royalties >," tweeted electronic artist Siren Diamant, capturing the euphoria among bedroom producers. South African DJ Tezz echoed the sentiment: "Major Update for Independent Artists... keep 100% of their royalties, not just on SoundCloud, but also on Spotify, Apple Music, TikTok and more." NME amplified the buzz, headlining it as a "100 per cent of distribution royalties" win, while Mixmag hailed it for dance music creators.

Not everyone's popping champagne uncorked. Your EDM's analysis cut through the hype: "When each stream is still worth a fraction of a cent, this indicates that artist's 'keeping' their royalties may not mean much at all." Indie vocal coach Randy Morano noted the fee: "100% royalties for creators (minus an $8.25/mo fee). A huge shift toward artist-first monetization." And Charm Cracker, a self-produced multi-instrumentalist, plugged referral codes, underscoring the grassroots uptake: "They are bypassing the competition... Spotify is not working for me, really."

Interviews with early switchers paint a nuanced picture. "I've been grinding SoundCloud since 2019, scraping by on 80% scraps," shares LA-based rapper J. Mira, who dropped her debut mixtape Echo Chamber last week. "Now? That extra 20% covers my plug-ins. It's not riches, but it's respect." For global acts like Romio, an Italian Web3 musician blending NFTs with beats, the plan aligns with blockchain's decentralization ethos: "Big news for indie artists! 100% of distribution royalties for tracks released via SoundCloud."

Even skeptics concede the momentum. Rolling Stone Brasil covered it in Portuguese, emphasizing accessibility for non-English markets, while Exclaim! touted it as a North American boon. As one anonymous SoundCloud vet put it: "It's the push we needed to treat music like a business, not a hobby."


Ripples in the Royalty Pool: Broader Industry Shifts

This isn't happening in a vacuum. The music biz is in flux: Universal Music Group's 2024 TikTok standoff exposed streaming's fragility, while AI tools like Suno democratize production but dilute per-stream value. Platforms like Bandcamp thrive on direct-to-fan models (100% artist cuts, minus 10-15% fees), but lack distribution scale. SoundCloud bridges that gap, potentially siphoning users from Spotify for Artists (70% royalties post-label cuts) or Apple's $9.99 pro tier.

Economically, it's a calculated risk. SoundCloud's revenue—pegged at $300 million annually—leans on ads and subs, not distribution vig. By waiving the 20% (estimated $50 million yearly), they're banking on volume: more uploads mean more engagement, data sales, and ecosystem lock-in. Analysts at Billboard predict a 25% subscriber bump by Q2 2026, especially if viral hits like the platform's past (e.g., Lil Nas X's "Old Town Road") recur.

For marginalized creators—women, POC, LGBTQ+ artists who comprise 60% of SoundCloud's base per a 2025 USC Annenberg study—this levels the field. No more gatekept advances or predatory deals; just tools to build empires. "We're not just streaming; we're stewarding," Herd said in a company blog, though exec quotes remain sparse.


A Symphony of Change: What's Next?

As November's payouts roll out—first 100% checks hitting December—watch for copycats. Will DistroKid drop its unlimited model to zero cuts? Could Spotify counter with fan-direct royalties? SoundCloud's bet is bold: in a $28 billion streaming market dominated by three majors, empowering the underdogs might just rewrite the score.

For now, artists like J. Mira are tuning up. "This feels like the remix we deserved," she says. In an industry often accused of lip-syncing equity, SoundCloud is finally letting creators take the mic—and keep the mic drop.

The previously leaked Behringer BQ-10 analogue sequencer has officially launched — here’s everything you need to know

Behringer has quietly moved another long-rumoured product from the “leak” pile to the shop window: the BQ-10 analogue sequencer. If the name and styling look familiar, that’s because the BQ-10 is a modern reimagining of the classic Korg SQ-10 design — rebuilt with contemporary connectivity and Eurorack friendliness in mind. Whether you’re a vintage-sequencing obsessive, a modular tinkerer, or a bedroom producer who loves hands-on pattern creation, the BQ-10 is designed to slot into many setups without demanding a PhD in patching. 


What the BQ-10 actually is

At its heart, the BQ-10 is an analogue, voltage-controlled step sequencer giving you 24 total steps arranged as three 8-step lanes. Channels A and B can operate independently as two 8-step sequences or be combined into a single 16-step melodic lane; channel C remains dedicated as an 8-step control lane for filter, amplitude or other timbral modulations. That flexible three-lane architecture is what gives the BQ-10 a lot of creative mileage: you can have two melodic lines and a separate automation lane running at once, or stitch lanes together for longer phrases. 

Physically, it’s presented as a desktop unit that can be removed and mounted into Eurorack — useful if you want both a tidy desktop workstation and rack integration. The layout is classic: knobs and switches for each step, portamento and duty-cycle/gate controls, dedicated sync and CV jacks, plus MIDI for integrating with modern gear. 


Key specs and features (quick list)

  • 24 analogue steps arranged as 3 × 8-step lanes (A, B, C). A+B = 16 steps when combined. 

  • CV and gate outputs for full voltage control; per-lane voltage range switches for compatibility with different synth standards. 

  • MIDI & USB — clock sync and MIDI note/Gate output for DAW/hardware integration. 

  • Clock generator (internal), adjustable roughly from very slow to very fast — and CV-controllable, plus external clock input for sync. 

  • Per-channel portamento (glide) and a global duty-cycle control to shape gate length. 

  • Eight playback modes for A/B, including series, parallel, alternating and random options to vary sequence behaviour. 

  • Desktop & Eurorack capable — the unit detaches from its desktop case for rack mounting and lists a modest HP size on modular databases. 

Those bullet points are the essence of what delivers the classic SQ-10 vibe with modern usability. The BQ-10 keeps the tactile, real-time programming workflow that analogue sequencers are loved for, while making them far less awkward to integrate with modern setups. 


How it differs from the Korg SQ-10 (and why that matters)

Behringer’s take is recognisably inspired by the Korg SQ-10, but with a few practical updates. The addition of MIDI/USB and CV-friendly voltage range switches is the biggest functional upgrade — it means the BQ-10 can talk to vintage Hz/V synths, modern 1V/oct modular gear, and computer setups without a lot of external conversion. The dedicated third lane (C) being explicitly pitched for modulation tasks is also a modern convenience: rather than patching an external sequencer to a VCF or VCA, you can build timbral changes into your patterns from the same interface. 

There are aesthetic and ergonomic tweaks too: Behringer has chosen a darker control surface and wood end-cheeks for a retro but contemporary look, and the ability to run as a standalone desktop box or snap it into Eurorack opens it up to more workflows than the original hardware typically supported. For many users that “plug and play” modular compatibility is a decisive factor. 




Real-world usability: what it’s good at

  • Hands-on compositional jamming. The step pots and physical transport controls invite experimentation — perfect for quickly sketching motifs or improvising live sequences. 

  • Modulation-driven patches. Use channel C to twist filter cutoff, VCA level, or any CV-controllable parameter on your synth to create evolving grooves without external CV routing. 

  • Integration with DAWs/hardware. MIDI/USB means the BQ-10 can either be the central timing source for hardware rigs or slotted into a software workflow; conversely, the CV outputs let it remain a pure-analogue controller when you want that sound. 

  • Modular-friendly sequencing. If you build with Eurorack, the detachable form factor and full CV outputs let the BQ-10 live as a compact step engine in someone’s case. 

It’s not trying to be a sequencer DAW replacement; its charm and strength are immediacy, character and hands-on performance — the traits that made step sequencers indispensable in classic electro and electronic music. 


Pricing and availability

Behringer has listed the BQ-10 at an accessible price point — around US$149 / £149 in many retailers’ listings — which positions it as an affordable entry into hardware sequencing and modular-adjacent setups. Preorders have appeared across multiple dealers, and Behringer’s own product page and press materials indicate the unit is now officially launched and shipping through retailers in batches. Expect initial demand to outpace supply for a while; popular Behringer launches tend to sell quickly. 


Who should consider buying one?

  • Beginners to hands-on sequencing who want a quick, low-cost way to learn analogue sequencing concepts without a full modular investment.

  • Modular users seeking an inexpensive, compact step engine to run melodic and modulation lanes in their rack.

  • Live performers and jam musicians who want a tactile pattern machine that responds well to on-the-fly changes.

  • Producers after character — if you like the slightly unpredictable, squelchy charm of analogue CV and manual knob tweaks, the BQ-10 brings that to low cost. 

If your main studio workflow is tightly DAW-centric and you need deep pattern memory, song mode, step automation recall and long-form sequencing, you may find the BQ-10’s analogue, patch-centric workflow less convenient than a dedicated sequencer with full CV-to-MIDI mapping and presets. But for raw hands-on sequencing, it’s hard to beat for the price. 


A few practical caveats and community reactions

  • Purists will notice differences between the original Korg SQ-10 and Behringer’s implementation — some step counts and nuanced behaviours vary. That said, many users welcome the modern connectivity, even if small purist gripes appear on synth forums.

  • Supply timing. As with many new Behringer launches, availability may vary regionally, and some retailers show preorder lead times — so if you need one immediately, double-check local stock and shipping windows.


Final thoughts — is the BQ-10 worth the buzz?

The BQ-10 is exactly what it promises to be: a relatively inexpensive, hands-on analogue sequencer that wears its heritage on its sleeve while adding pragmatic modern features. For anyone who’s ever loved the immediacy of knob-per-step sequencing and wanted to plug it directly into modular or MIDI setups without extra interface boxes, the BQ-10 is a compelling option. At its price point and with its combination of CV, gate, MIDI and Eurorack compatibility, it looks set to be a popular choice for curious beginners and seasoned hardware lovers alike. 

If you want to dive deeper before buying, check the Behringer product page for full specs and manuals, watch hands-on demo videos from retailers and reviewers to hear how it behaves with your kind of synth, and read user threads for early impressions on build quality and integration tips. Happy sequencing!

Pandora Launches 2 AM: A Bold New Home for DJ Set Culture

In a streaming market saturated with algorithm-driven playlists and on-demand single tracks, Pandora is staking a new claim in the world of electronic music with 2 AM, its brand-new station and DJ takeover series. The idea is simple yet ambitious: to elevate DJ sets as curation, offering exclusive mixes and artist-led programming rather than relying purely on automated playlists. The official positioning? 2 AM is being billed as “the home for DJ set culture on Pandora.” In this post, we’ll dig into what 2 AM is, why it matters, how it’s different from existing DJ/dance streaming options, and what challenges and opportunities lie ahead.


What Exactly Is 2 AM?

At its core, 2 AM is a curated music channel and a rotating “DJ takeover” format in which established DJs handpick tracks, mix them into playlists, and deliver hosted intros or commentary. The first wave of takeovers comes from big names in electronic music — such as Above & Beyond, John Summit, and Don Diablo — who will lend their own sensibilities to the station. 

Pandora intends for 2 AM to cover the full spectrum of dance music: from EDM and tech house to nu-disco, chillout, and beyond. The playlists will be refreshed regularly, ensuring a steady stream of new and evolving content. 

What sets 2 AM apart from Pandora’s existing dance stations is the emphasis on the DJ as curator, not an algorithm. Rather than simply mirroring what listener behavior suggests, 2 AM intends to reflect what DJs are actually spinning in clubs and festival sets — the tracks that move crowds, not just the ones data suggests you might like. 


Why It Matters — For Listeners, Artists & DJs

1. Reconnecting Listeners to the Community and Energy of DJ Culture

One criticism of much digital music consumption is that it strips away context. You might listen to a track, but you don’t hear why it was chosen or how it works in a larger DJ set. With 2 AM, Pandora is attempting to re-embed that context: the intros, the track transitions, the DJ’s narrative flow.

For passionate fans of electronic music, 2 AM offers something more immersive than a static playlist. It’s closer to being at a club, but in your ears. It gives listeners an inside track on what’s trending in dancefloors globally — not just what an algorithm infers from your history.

2. Spotlighting DJs as Creative Curators

Many streaming services downplay the role of the DJ, reducing them to a single “official artist” credit or dropping them from playlists entirely. 2 AM flips that paradigm. The DJ becomes the central creative voice: you listen through them, not just to them.

For those DJs whose work centers on live sets, remixes, or hybrid genre explorations, 2 AM offers a new canvas: curated radio-style shows, unique intros, and perhaps even a way to test unreleased edits or blends.

3. An Avenue for Deeper Discovery

Another benefit of 2 AM is its potential as a discovery engine. Instead of being guided by algorithmic similarity suggestions, listeners might follow a DJ’s personal tracklist and stumble into underground or emerging artists they wouldn’t have found otherwise. This gives Pandora another differentiator in a market awash with “similar to what you already listened to” stale recommendations.


What Makes 2 AM Different from Other Platforms?

DJ-centered content isn’t new — platforms like Mixcloud, SoundCloud, or YouTube host DJ sets and live mixes. But 2 AM offers something that many of those platforms lack:

  • Native integration with Pandora’s streaming infrastructure. No need to bounce between apps; the station lives within Pandora’s ecosystem.

  • High production standards: These aren’t user-uploaded club sets; they’re artist-approved, hosted, and styled for streaming consumption.

  • Scheduled rotation + freshness: Rather than one static mix, 2 AM will continuously refresh with new DJ takeovers, shifting styles, and seasonal changes. 

  • Curator identity built in: The DJ is not anonymous. Their personality, voice, and taste become a part of the listening experience.

In short, 2 AM is not just “DJ sets in Pandora” — it’s DJ programming inside a streaming radio experience.




Risks, Hurdles & Considerations

As bold as 2 AM sounds, Pandora will face challenges in making it work sustainably:

Metadata & Rights Management

DJ sets often include a blend of tracks, edits, remixes, or bootlegs. Licensing that material for streaming — especially when included in continuous, curated mixes — can be tricky. Pandora will need robust metadata, royalty accounting, and legal agreements to ensure rights are properly cleared.

  Balancing Mass Appeal with Niche Integrity

The broader Pandora audience might not always gravitate toward extended DJ sets or deeper cuts. Pandora must strike a balance between accessibility (making 2 AM inviting for casual listeners) and authenticity (sustaining the respect of die-hard EDM/dance fans). If the mixes lean too “safe,” they risk alienating the core audience; lean too underground, and they might lose mainstream appeal.

Scheduling & Rotation

To keep listeners coming back, Pandora will need to manage scheduling carefully — rotating DJ takeovers, refreshing playlists, and introducing new voices regularly. Stagnation is a risk if the same DJs or subgenres dominate for too long.

Measuring ROI & User Engagement

How will Pandora measure success? Will listener metrics (hours streamed, listener retention) justify the investment in curated DJ content versus algorithmic programming? Pandora will likely monitor how 2AM engagement compares with its existing stations.


What to Watch Going Forward

Here are a few signals I’ll be keeping an eye on:

  1. New DJ lineups — will Pandora bring in more underground or rising talent, not just established names?

  2. Listener data — how long do users stay tuned? Is 2 AM driving new registrations or retention?

  3. Community features — will Pandora add the ability to “follow DJs,” get notifications about upcoming takeovers, or explore past sets?

  4. Cross-promotion with events — could Pandora link 2 AM DJ sets to live festival streams or event tie-ins?

  5. Artist exclusives & premieres — the ability to debut exclusive remixes or edits through 2 AM could become a talking point and competitive edge.


Conclusion: Why 2 AM Could Be a Turning Point

Pandora’s 2 AM is more than just another streaming channel — it’s a bet on returning agency to the human curators in music. In an era where algorithmic discovery dominates, giving DJs a prominent voice is a statement: dance music is about more than data; it’s about taste, context, and narrative flow.

If Pandora can get the licensing, rotation, and listener onboarding right, 2 AM could become a meaningful bridge between club culture and everyday listening. It offers fans a new way to engage, DJs a new platform to express, and Pandora a fresh point of differentiation in a crowded streaming market.

As 2 AM rolls out and evolves, it could redefine how we experience DJ sets — not just as occasional live events, but as serialized audio journeys you can tune into nightly, weekly, or whenever the mood strikes.

Spotify’s “SongDNA”: A New Era of Music Discovery Through Credits

In a world where music streaming is dominated by algorithms, playlists, and endless recommendations, Spotify seems to be taking a fresh, human-centric approach. Recently, tech insiders discovered that Spotify is developing a new feature called “SongDNA.” This potential update could change how listeners discover music — not just by sound or genre, but by the people behind the songs.

Although the company hasn’t officially announced it yet, clues buried in Spotify’s app code hint at an exciting new direction for music discovery — one that highlights the unsung heroes of every track: the writers, producers, and engineers who make the magic happen.


What Is SongDNA?

SongDNA appears to be a feature in development that will give users the ability to explore music through its credits. Instead of focusing purely on the artist or the song’s title, Spotify’s new approach aims to let users dive into the DNA of a track — the creative contributors who shaped it.

This discovery came from Jane Manchun Wong, a well-known app researcher who frequently uncovers unreleased features in popular apps. Wong found references to “SongDNA” inside Spotify’s code, along with screenshots of what appears to be a new interface that lists each contributor on a track and connects their work across other songs.

For example, if you were listening to a song produced by Mark Ronson, SongDNA might let you tap his name to see all the tracks he’s ever produced — across artists, genres, and eras. Imagine discovering connections between your favorite songs and realizing they share the same songwriter, engineer, or session musician. That’s the kind of experience SongDNA could deliver.


A Shift from Algorithms to Humans

For years, Spotify’s success has been driven by its algorithmic curation — playlists like Discover Weekly or Release Radar use data and behavior patterns to serve you the next best track. But SongDNA hints at a more organic and human discovery path, one rooted in creativity, connection, and collaboration.

It represents a philosophical shift. Music discovery has long been about what you like. SongDNA introduces the idea of who you like — the creative fingerprints behind your taste. Instead of browsing by genre, you could browse by creator network.

This approach could redefine how fans think about their favorite songs. Many listeners don’t realize that some of today’s biggest pop hits were written or produced by the same handful of creative masterminds. For example, Max Martin has written dozens of chart-topping hits for Britney Spears, The Weeknd, Taylor Swift, and Katy Perry. SongDNA could make those invisible connections instantly visible — revealing the invisible threads that tie the music industry together.




Why This Matters for Artists and Producers

The potential benefits of SongDNA go far beyond fans. For songwriters, producers, and engineers, this could be a game-changer.

One of the biggest frustrations in the modern music industry is the lack of visibility for the people who work behind the scenes. While major artists get the spotlight, songwriters and producers often remain anonymous to the average listener — even though their contributions shape the sound of entire generations.

SongDNA could finally bridge that gap. By integrating credits directly into the discovery experience, Spotify would give recognition where it’s due — making it easier for industry professionals to build a public portfolio within the world’s largest streaming platform.

This transparency could also create new networking opportunities. A producer might find a new vocalist to collaborate with. A songwriter could discover others who share a similar creative style. Even music fans could use this feature to better understand the craftsmanship behind their favorite hits.


How SongDNA Could Work

While Spotify hasn’t confirmed any details, leaks suggest that SongDNA might include:

  • A detailed credit breakdown: listing every person involved in the song’s creation — from writers and producers to mixers and mastering engineers.

  • Clickable profiles: allowing users to tap a contributor’s name to see all the other songs they’ve worked on.

  • Connected discovery: recommendations based on shared contributors rather than listening habits — e.g., “Other tracks produced by Metro Boomin.”

  • Interactive visual design: perhaps showcasing “creative webs” connecting different artists and their teams.

Essentially, Spotify could evolve into an IMDB for music — a place where you can trace every creative connection in the industry.


The Competition: Tidal and Beyond

Spotify isn’t the first streaming platform to explore song credits. Tidal, for example, has long championed detailed credit listings, allowing users to view who contributed to a track. However, SongDNA seems to go beyond simple listings — it’s designed to make those credits discoverable.

Instead of credits being buried in a menu, they could become the foundation for new recommendations. This would give Spotify a fresh discovery edge — a more narrative, human layer to its data-driven ecosystem.

It also aligns with a broader industry trend: recognizing creators. From TikTok tagging original sound creators to YouTube’s music metadata updates, platforms are realizing the importance of crediting every contributor. SongDNA fits perfectly into this cultural moment.


Challenges Spotify May Face

As exciting as SongDNA sounds, it won’t be easy to implement. There are a few major hurdles Spotify must overcome:

  1. Incomplete metadata – Many songs, especially older ones or indie releases, lack accurate or complete credit data. This could make the feature inconsistent at first.

  2. Licensing and data integration – Spotify may need to partner with publishers, PROs, and metadata providers to gather and verify this information.

  3. User interface complexity – Displaying all this data clearly and attractively will require careful UX design to avoid overwhelming users.

  4. Adoption and awareness – Listeners need to care about credits. While hardcore music fans will love this, casual users may not explore it unless Spotify promotes it effectively.

If the company pulls it off, however, SongDNA could become one of Spotify’s most valuable discovery tools in years.


The Bigger Picture: Rediscovering the Human Side of Music

In a time when AI and automation dominate the headlines, SongDNA feels refreshingly human. It reminds us that music is not just content; it’s a collaboration — a product of emotion, creativity, and teamwork.

By surfacing the people behind the songs, Spotify could spark a cultural shift: encouraging fans to value who made the music as much as who performed it. For upcoming artists, that visibility could translate into recognition, respect, and new opportunities.

If Spotify launches SongDNA, it could redefine how the next generation interacts with music — not just consuming it, but exploring its creative lineage.


Final Thoughts

At the moment, SongDNA remains in development, and Spotify has yet to make an official statement. However, the discovery of this feature in Spotify’s app code — confirmed by multiple sources like TechCrunch and Digital Trends — indicates that the company is serious about exploring this direction.

Whether it launches next month or next year, one thing is clear: Spotify wants to make music discovery more personal, more connected, and more human.

If SongDNA becomes a reality, it could mark the next great evolution in streaming — turning every play into a journey through the creative universe of the artists, writers, and producers who make music come alive.

Page 1 of 7
The music world is always moving forward: new instruments, fresh sounds and unexpected solutions appear that inspire artists to create unique tracks. The SoundsSpace blog often raises topics related to creativity, recording and modern technologies that help musicians find new ways of expression. The industry is changing rapidly, and along with it, new areas appear where art and technology meet on the same wavelength. One of the interesting areas is digital entertainment, which uses similar technologies to create vivid impressions. Modern online casinos, for example, are introducing innovative programs that improve graphics, sound and the general atmosphere of virtual games. An overview of such software for 2025 is presented on the websitehttps://citeulike.org/en-ch/online-casinos/software/. These solutions are in many ways similar to how music platforms use digital effects and plugins to give the listener a more lively and rich perception. In both music and the entertainment industry, high-quality software comes to the forefront, setting the level of impressions. The artist cares about sound, the player cares about visuals and dynamics, but in both cases technology becomes an invisible mediator between the idea and its implementation. This approach unites creative industries and opens new horizons for musicians and developers, shaping a future where the digital environment becomes part of real art.