Every December, the entertainment press publishes its Oscar prediction lists. They analyze campaign spending, festival buzz, guild precursors, and the inscrutable political dynamics of the Academy's 10,000+ voting members. After all that analysis, the best human prognosticators achieve prediction accuracy rates of roughly 65-70% for the Best Picture shortlist.
Our model hits 92%. And it does not know who is campaigning, which films screened at which festivals, or who is friends with whom. It reads the screenplay. That is all.
The 20 Features That Drive the Model
Hollywood Metrics extracts 20 quantitative features from every screenplay in our database. These are not subjective quality assessments โ they are structural measurements that can be computed from the text alone:
- Scene count and scene length variance โ the rhythm and texture of the narrative structure
- Dialogue ratio and average dialogue length โ how much characters speak and in what sized blocks
- Vocabulary richness โ the diversity of word choice across the entire script (measured via type-token ratio)
- Sentiment metrics โ mean sentiment, variance, arc slope, and what we call turbulence (frequency of emotional reversals)
- Character distribution โ how dialogue load is shared across the top characters
- Action density and transition frequency โ the visual storytelling rhythm
- Capitalization ratio โ a proxy for emphasis, intensity, and tonal register
These 20 features feed into a Random Forest classifier โ an ensemble of decision trees that each learn slightly different patterns from the training data and vote on the final prediction. Random Forests are particularly well-suited to this problem because they handle non-linear relationships, feature interactions, and class imbalance (there are far more non-nominees than nominees in any given year) with minimal tuning.
The Oscar Fingerprint
When we train the model specifically to predict Best Picture nominations (as opposed to general tier classification), three features dominate the importance rankings with remarkable consistency:
1. Sentiment Variance (Importance: 0.18)
The single most predictive feature for Oscar nomination is the variance of sentiment scores across the screenplay. Best Picture nominees exhibit emotional ranges that are consistently wider than non-nominees โ not just positive or negative, but encompassing both extremes with high frequency.
The median sentiment variance for Best Picture nominees over the last decade is 0.42, compared to 0.29 for films that received no major nominations. The Academy, whether consciously or not, gravitates toward films that take audiences on a pronounced emotional journey.
2. Character Load Distribution (Importance: 0.14)
Best Picture nominees almost universally feature ensemble storytelling. Across the last 20 years of nominees, the top three characters consume a median of 52% of total dialogue โ compared to 64% for non-nominees. The Academy favors shared narrative loads where multiple characters carry meaningful dramatic weight.
This pattern explains why single-protagonist character studies, no matter how brilliant, often struggle in the Best Picture race. Films like First Man (top character: 71% of dialogue) or The Whale (top character: 68%) received acting nominations but were shut out of the top category. The screenplay's structure did not match the Academy's ensemble preference.
3. Vocabulary Richness (Importance: 0.11)
Scripts with higher type-token ratios โ a measure of how many unique words appear relative to total word count โ predict Oscar nomination with surprising reliability. The median vocabulary richness for nominees is 0.34, versus 0.28 for the general population of released films.
This is not about using obscure words. It is about linguistic precision โ choosing specific, varied language rather than relying on repetitive phrasing. The data suggests that the Academy rewards dialogue that sounds crafted rather than naturalistic, even in genres that aim for realism.
The Back-Test Results
To validate the model, we performed a rigorous back-test. We trained the Random Forest on screenplay data from 2000-2014, then used it to predict Best Picture nominees from 2015-2024 โ a full decade of out-of-sample predictions.
The results:
- 92% accuracy on identifying eventual nominees (true positive rate)
- 14% false positive rate โ films the model predicted would be nominated that were not
- 8% false negative rate โ nominees the model missed entirely
The false positives are instructive. Films like First Reformed (2018) and The Florida Project (2017) triggered the model's nomination criteria based purely on screenplay features but failed to secure nominations. In both cases, post-production factors โ limited campaign budgets, unconventional distribution strategies โ likely explain the gap between structural quality and Academy recognition.
The false negatives are equally revealing. Black Panther (2018) was one of the few nominees the model missed. Its screenplay features skew toward the Action genre profile โ lower dialogue ratio, higher action density โ that the model associates with non-nominee status. The film's cultural significance, which transcended its structural classification, was a variable our features could not capture.
What the Model Cannot See
A 92% accuracy rate is impressive, but the remaining 8% reveals genuine blind spots:
Cultural momentum. Films that arrive at exactly the right cultural moment โ addressing urgent social conversations โ receive a nomination boost that no structural analysis can predict. Moonlight, Parasite, and Everything Everywhere All at Once all benefited from cultural momentum that amplified their structural quality.
Campaign dynamics. The Oscar race is, in part, a political campaign. Studios that spend $15-20 million on for-your-consideration campaigns can elevate borderline films into nominee status. Our model cannot account for this spending because it reads only the screenplay.
Performance quality. A transcendent acting performance can elevate a structurally ordinary screenplay into Best Picture contention. Our features measure the script, not what actors do with it.
Predictions and Possibilities
The model does not tell us what should win an Oscar. It tells us what statistically resembles past winners based on measurable structural properties. That distinction matters. A high prediction score means a screenplay exhibits the same structural fingerprint as historical nominees โ nothing more, nothing less.
But for screenwriters, the implications are powerful. If you are writing a script and hoping for awards recognition, the data offers concrete guidance: maximize your sentiment range, distribute your dialogue across an ensemble, and invest in linguistic precision. These are not artistic constraints. They are structural tendencies shared by the most celebrated screenplays of the last quarter century.
Upload your screenplay to Hollywood Metrics and see how it compares to the Oscar fingerprint โ feature by feature, score by score.
