Skip to content
Lyric Scoring Standard
Prior art

Lyric scoring before the open standard.

We are not the first to score lyrics. We are the first to publish a complete rubric under CC BY 4.0 with a public calibration corpus, an ed25519 reproducibility seal, and an installable npm package. The tradition we extend; the gap we fill.

This page exists because category-creation claims that ignore prior art lose credibility on contact with anyone who reads the literature. The honest claim is "first OPEN rubric in a tradition with prior art," not "first ever." Naming the prior art is the difference between a confident claim and a wishful one.

Conservatory + craft tradition

Music schools have evaluated lyric craft for decades. The methodologies are real and rigorous; they predate any AI-music tool. What they have not done — and what an open standard now provides — is package the rubric in a form independent implementations can install and cite.

Berklee Songwriting Department craft rubrics

Berklee College of Music, Songwriting Department (decades, ongoing)

What it scores: Multi-metric craft evaluation across prosody, meter, rhyme scheme, narrative arc, and vocal register. Used in classroom evaluation and the Berklee online songwriting curriculum. Pat Pattison's "Writing Better Lyrics" + "Songwriting Without Boundaries" formalize key components.

Gap our open standard fills: Course-internal; not published as a versioned, citable specification. Adoption requires enrolling or buying the books; no npm package, no machine-readable rubric, no reproducibility seal.

NYU Tisch Clive Davis Institute lyric workshops

NYU Tisch Clive Davis Institute (ongoing)

What it scores: Workshop-based critique using rubrics covering specificity, voice consistency, hook strength, and structural integrity. The institute's workshop format scales the methodology to small cohorts of working songwriters.

Gap our open standard fills: Tacit knowledge in the workshop tradition; not a published, externally-implementable rubric.

Sheila Davis songwriting methodology

Sheila Davis, "The Craft of Lyric Writing" (1985) and successor volumes

What it scores: Rigorous craft framework covering imagery, conversational authenticity, rhyme types, prosody, structural patterns. Cited across the conservatory tradition; influential on Berklee + NYU curricula.

Gap our open standard fills: Book-form methodology; not a versioned specification with a calibration corpus.

Music information retrieval (MIR) research

Academic music informatics has scored lyrics in narrow dimensions for years — readability, rhyme density, sentiment trajectory, lyric-melody alignment. The rigor is real; the gap is integration. Each paper measures one thing; an open standard measures the full craft surface.

MIREX Lyric Evaluation tasks

Music Information Retrieval Evaluation eXchange (annual, ongoing since 2005)

What it scores: Annual evaluation tasks covering lyric retrieval, lyric alignment, lyric-language detection, and (sporadically) lyric quality scoring. Provides shared datasets for academic teams to benchmark lyric-related models.

Gap our open standard fills: Task-specific; not a unified craft-quality rubric. Different MIREX years measure different dimensions; cumulative coverage of lyric craft is partial.

Lyric complexity + readability metrics

Various papers (Mahedero et al. 2005; Logan et al. 2004; ongoing work in computational stylistics)

What it scores: Computational measures of lyric vocabulary diversity, sentence-length distribution, parts-of-speech balance, and Flesch-Kincaid-style readability scores adapted to lyrics.

Gap our open standard fills: Surface-level measures (what the lyric LOOKS like statistically); does not score craft moves like prosody, voice consistency, or arc.

Lyric-melody alignment + prosody alignment research

Various papers (Nichols et al.; Kreyssig et al.; ongoing)

What it scores: Computational measures of how a lyric line's stress pattern aligns to a melody's phrase boundaries. Strong on the prosody-musicality dimension; the rubric metric M1 (Prosody) builds on this tradition.

Gap our open standard fills: Single-metric; does not extend to the other 11 dimensions of craft quality.

Music theory + aesthetics scholarship

Academic music theory has developed sophisticated frameworks for analyzing lyric-music interplay, prosody, and pop-song aesthetics for decades. The Lyric Scoring Standard does not invent these categories; it operationalizes them. Naming the tradition prevents the credibility-trap of "we discovered specificity matters" — specificity has been theorized in academic music aesthetics since the 1980s.

Lerdahl & Jackendoff — A Generative Theory of Tonal Music

Fred Lerdahl & Ray Jackendoff (MIT Press, 1983)

What it scores: Foundational text establishing how rhythmic + metrical hierarchies in lyrics align with musical structure. The "metrical preference rules" framework formalized how stressed syllables interact with strong beats — directly informs metric M1 (Prosody & Musicality) in the open standard. Cited across decades of subsequent music-theory + computational-musicology research.

Gap our open standard fills: Theoretical framework, not an applied rubric. The standard operationalizes the theory into a 0-100 score with concrete anti-inflation anchors; Lerdahl & Jackendoff give the theoretical grounding the score relies on.

Philip Tagg's musemes + multi-parameter song analysis

Philip Tagg, multiple volumes (1979-2013), most notably "Music's Meanings" (2013)

What it scores: Established the methodology of analyzing pop songs across multiple expressive parameters simultaneously: tempo, register, instrumentation, lyric content, vocal delivery, structural patterns. Tagg's "interobjective comparison" technique anchors how the open standard's 12 metrics work together rather than in isolation.

Gap our open standard fills: Analytical methodology rather than a normative rubric. Tagg names the parameters; the open standard scores them. The two are complementary — Tagg gives the vocabulary, the standard gives the calibration.

Simon Frith — Performing Rites: On the Value of Popular Music

Simon Frith (Harvard University Press, 1996)

What it scores: Seminal academic work establishing the aesthetics of pop-song evaluation. Frith's framework distinguishes between formal craft (mechanics), semantic content (meaning), and performative authenticity (voice). These three dimensions map onto the open standard's tier structure (Craft 25% / Expression 40% / Impact 35%) — though the weighting is the standard's contribution, not Frith's.

Gap our open standard fills: Aesthetic argument, not an applied scoring instrument. Frith argues for HOW pop songs should be evaluated; the open standard implements WHICH metrics + weights to use.

Industry rubrics + A&R frameworks

Music publishing, A&R review, and sync licensing all use internal rubrics for evaluating lyrics. They share a tradition and a vocabulary, but each company's rubric is private. An open standard externalizes what the industry already does internally.

Publishing house staff scorecards

Industry-internal at major + indie publishers (private; ongoing)

What it scores: Multi-axis scorecards used by song-pluggers and creative staff to evaluate incoming demos. Common axes include: hook strength, lyric specificity, vocal castability, sync-eligibility, market-fit. Published rubric structure varies per company.

Gap our open standard fills: Private; not citable; not externally implementable. The methodology exists across the industry; no public artifact captures it.

A&R critique frameworks

Industry-internal at major labels (ongoing)

What it scores: Evaluation methodologies used in A&R demo review covering vocal performance, lyric craft, hook strength, marketability. Often combined with audio-side critique into a single demo verdict.

Gap our open standard fills: Private label-internal scoring; not a citable specification.

Sync-licensing review templates

Music supervisors + sync agencies (industry-internal; ongoing)

What it scores: Rubrics for evaluating whether a lyric will pair with a brief's emotional register, whether it carries problematic lyric content (clearance flags), and whether the hook is strong enough to anchor the cue. Used across film/TV music supervision.

Gap our open standard fills: Brief-specific; private; tied to the supervisor's personal taste and the production's clearance requirements.

What the Lyric Scoring Standard adds to this tradition

  • Versioned + citable: Every rubric version has a date, a number, and a changelog. External implementers cite by version, not by approximation.
  • Machine-readable: The rubric ships as JSON at /scoring-standard.json and as the npm package @songforgeai/scoring-rubric.
  • Open license: CC BY 4.0. Any tool, course, or paper can implement it; attribution by name + version is the full license requirement.
  • Calibration corpus: A public hand-scored corpus (/scoring/corpus) gives independent implementations a ground-truth to calibrate against.
  • Reproducibility seal: Every score response carries an ed25519 signature binding the score to its rubric version, model, temperature, and build SHA. Verification is independent of SongForgeAI infrastructure.
  • Self-audit posture: Bias findings in the rubric are disclosed publicly before they're patched. The maintainers’ posture is the trust infrastructure no private rubric can replicate.

Missing prior art? Email support@songforgeai.com. Each entry must be verifiable (book, paper, university program); we add new entries as they’re named. The page accumulates rather than curates — the more prior art on it, the more honest the positioning.

More from the standard