To start, we should first examine the
previous study, published by Rotten Tomatoes (RT) in 2005.
Here is part of the report:
When “Star Wars Episode I: The Phantom Menace” was released
in 1999, a group of us actually went to our local library and dug up
a sampling of available sources that reviewed the original trilogy
during the time of their respective release dates in the late 1970s
and early 1980s. [...] The results are actually quite
surprising.
Tomatometer Scores for Original Trilogy During
Original Release Dates:
(Click on the links for the archived
quotes from Archive.org)
31% - Return of the
Jedi
52% - The Empire
Strikes Back
79% - Star
Wars
Average Tomatometer: 54%
As
one can see, only “Star Wars” managed to be Fresh, with a
respectable 79% on the Tomatometer, while the other two sequels got
successively worse. Most of the critics thought the first film was
an inventive, fun, and entertaining summer popcorn movie. It’s
interesting that they complain about the dialogue back then too.
“Empire,” which is regarded as the best of the series nowadays, only
managed to score a mixed 52%. It received great technical grades,
but critics had problems with the plot, one way or other, and
thought it was just “minor entertainment.” It got worse with “Jedi”
– uneven pacing, no character development, tired acting, and hollow
and junky filmmaking. It scored a moldy 30% on the Tomatometer.
Prequels were probably the last thing critics wanted back then after
the thrashing of the last film.
Ironically, if you compare
the average Tomatometer of the prequels and the original trilogies
during the time of their respective original release dates, the
Prequels are actually better reviewed by 16% -- 70% to 54%,
respectively!
Tomatometer Ranking of Star Wars Series
Based on Critical Reaction During Original Release Dates:
83%
- Star Wars Episode III: Revenge of the Sith
79% - Star
Wars
65% - Star Wars Episode II: Attack of the Clones
62% -
Star wars Episode I: The Phantom Menace
52% - The Empire Strikes
Back
31% - Return of the Jedi
This study, while noteworthy for the effort, is
fundamentally flawed in many ways. To start, the comparison between prequels and originals is
unfair--the prequel trilogy takes its rating from websites and the
like, while the original trilogy reviews are sources like New
Yorker and Newsweek, very different sets of reviewers;
RT has a solution to this problem, which is the "top critic" filter,
which only counts legitimate publications. When this is selected, we
see that the prequels' ratings fall drastically, on average being
comparable to its ratings for the originals. Phantom Menace
goes from 62% to 39%, Attack of the Clones from 65% to an
equally moldy 38%, and Revenge of the Sith's proud 83%
becomes 69%. Now, onto the problems with those original trilogy
ratings.
First is that their source sampling is rather low--you
can browse the sources they used, and for the most part they
don't exceed fourteen or fifteen per film; I counted 48 in
total. This is impressive in its own right, but a study on
the films' reception cannot be accurate when accounting for so few
sources. My own study has the same limitation, however as I have
included about twice the amount of sources, amounting to 82 popular
and [ed: amount forthcoming] academic, it is at least useful as
an indicator or guideline. Secondly, and this will be dealt with in
the second part of this study, is the context when one is comparing
the films to the prequels--the films came out in different eras, and
were viewed in different ways. Thirdly is that, because there was
no star-rating system in the OT era, perception of the reviews
is subjective--an issue I hoped to minimize in my own study. I found
that RT's assignment of "rotten" or "fresh" for these reviews
to be very inconsistent as well; for instance, I felt that New
York and New Leader's review of Empire, while
not overtly positive by any means, merited about a 3/5, while RT
tallied them as "rotten" (2.5/5 or less), which to me is not
representative of these two reviews--that their conclusions are
either "fresh" or "rotten" with no sliding scale makes measurement
even more impercise. Their sources are also a bit random, with
a high proportion of political magazines such as New
Republic. Also problematic is the "rating" RT assigns--the
percentage isn't actually a rating. It is a measure from the
"tomatometer", which is a ratio of "fresh" reviews (positive)
to total, which isn't always helpful in determining anything
specific.
Now, I wish to explain the methodology of my own
study, to demonstrate why it is more valid. To start, academic
publications are, in many ways, their own world, with their own
standards, tastes and trends--I have included them in a separate
study, but I chose to stick mainly to reviews from popular sources.
In general, I tried to include as many daily newspapers as possible,
but the availability of archives has made this restricted to a large
degree; in the summation section below I include information from
reviews in local papers that are known to exist but were
not available to me. In total, I surveyed 18 daily newspapers from
the U.S. and Canada, and 1 from the U.K. However, I included other
weekly and monthly popular publications, such as The New
Yorker and People, which ran their own reviews. I also
was sure to include all of the major national publications--in
accounting for media reception, a publication's status and
readership matters. For instance, a good review in an obscure
small-town paper matters much less than if The New Yorker,
with readership in the millions and national prestige, runs a bad
review. As such, I feel as though I have accounted for
nearly every major review of all three films, even if there are
many smaller sources that are nodoubt absent. My source list goes
well beyond any review index available at any rate, and it also goes
well beyond the "top critics" shortlist that RT employs on its site,
being close to the amount that Metacritic.com uses in its
modern-day measurements. I regret that I was unable to obtain
Village Voice or Hollywood Reporter, but
barring these the major critics are accounted for. USA
Today didn't exist in 1977, and it seems Rolling Stone
didn't review the films.
An equally important issue is how to interpret the
reviews. The star rating system did not spring up until the 1980s,
and didn't become standard for many sources until the 1990s. In the
sources we are dealing with, almost all of them must be judged on
the tone and content of the review--which means that any study done
is to some degree subjective. To combat this, I tried to be as
conservative as possible--for instance, if I felt that a
review was leaning towards the positive side but somewhat
ambiguously, I played it safe and assigned it the more conservative
scale, so that my conclusions wouldn't be overstated. I did this
because I knew my ratings would be much higher than the RT study,
and didn't want to be accused of favouritism. When possible, I also
checked my own ratings against those assigned by sites like
metacritic, who rate some vintage reviews themselves; sometimes they
were similar, but sometimes there was some discrepancy. In general,
I tried to split the difference, but I also didn't want to assign a
review a rating I felt was unfair, so this was a fast and loose
rule. For example, Rotten Tomatoes assigns Gene Siskel's Chicago
Tribune review of Jedi as "rotten," and plunders
one of the few bad things he says about it--in fact he loved the
film! He also gave it an enthusiastic thumbs up on his TV show.
There were a few other instances on both RT and Metacritic where I
found some large unfairness.
There were a few reviews, however, that did have a
star rating, such as Gene Siskel (Chicago Tribune). He
rated the films out of 4, which meant I had to translate it to a
rating out of 5. His rating for Star Wars, for instance,
was 3.5/4, which I translated to 4.5/5, based in part on the tone
and content of his review, as I felt that 4/5 was a less accurate
representative.
Before I head into the data, I should also mention
sources and where they came from. Most archives available online,
either through public internet or scholars' portals, do not go
back before the 1990s, unfortunately, which restricted many of the
newspaper sources I could draw from. Some papers and magazines
went back to 1983 or 1981, but it is rare to find online archives
for daily papers that go back to 1977--which means that Star
Wars has far less sources than Return of the
Jedi. Jedi, for example, has about three dozen
sources overall, while Star Wars and Empire
have only about two dozen each. However, I feel that at
approximately 24 sources these first two films get enough
representation to make an approximate rating. Because I am located
in Canada, there was only limited American newspaper archives
available on microfilm (such as L.A. Times and
Washington Post), though this was to some degree balanced
out by the availability of Canadian newspaper
microfilms.
Finally,
I should also mention a word of caution about basing conclusions
solely on the average totals. These don't necessarily tell the full
picture. The trilogy had a very strange pattern where many viewers
would give the films 4/5 (80) or 5/5 (100), but then the odd 1/5
(20) and 2/5 (40) would crop up. Did these people see
the same movie? They are sometimes incredibly devisive. As a
result, even if five people rate the film as a 4 or 5, the one
or two 1/5 and 2/5 ratings drives down the average. You can see this
in all the films, but especially in Star Wars--most initial reviews
gave it 4 and especially 5, but with all the 1s and 2, it's average
of 84% doesn't give the impression of how much wide enthusiasm there
was. There is also a specific context to both the negative and
positive reviews that were posted, as temporal forces (in the case
of some critics who reviewed the film later, and explicitly
responded to the hype by giving it a low rating) and political ones
(such as, arguably, Pauline Kael's New Yorker review) must
also be accounted for. Also, I want to mention this now: there is
more to a movie's "reception" than just the reviews. There are
promotional articles and interviews with the cast and crew that
frequently appear, that go a long way to being part of the film's
reception, that aren't included--reviews do not give a picture of a
film's cultural impact or popularity, and this must also be
acknowledged.
Below, I present multiple sets of data. These reflect
the many ways one can measure the films' reception. For instance, a
review from Variety or Newsweek is much more meaningful in terms of
cultural influence and readership than a review from the Calgary
Herald. As such, I have the raw data, without adjustment, and then
alternate interpretations that attempt to adjust for weighted
influence, inclusion of academic sources, and other various schemes.
In part two of this study I will attempt to explain and interpret
this data, as well as tracking how the films' reception changed or
didn't change in the years between 1983 and 1997, and into 1997 and
beyond.
Overview and the
Data
To begin, I will present a brief overview. A more
thorough version of this is included in Part II of this
study.
Star Wars:
Star Wars by far had the best reception. Its
most enthusiastic reviews came from the earliest ones, which gave it
4/5 and 5/5 unanimously for the ten samples I have studied from this
first-wave period. Negative reviews do not begin to turn up until
June, however, they are still very infrequent in terms of their
appearance--only three reviews out of eight overall gave the film
unfavourable ratings in this second wave. J.W. Rinzler's The
Making of Star Wars affords us a much greater indication of the
national trends. Rinzler writes:
"Dozens of other reviewers were also won over by the
film, in papers throughout the country--in Santa Ana and Sacramento,
California; Salt Lake City, Utah; Evansville, Indiana; Camden, New
Jersey; Cincinnati, Ohio; and Louisville, Kentucky, to name a few.
Negative reviews ran in the papers of Augusta, Maine; Pascagoula,
Mississippi; and Statesville, North Carolina." (p.
296)
It should be noted that even here, the negative reviews
are in similar proportion to the positive ones, which he writes the
film received in "dozens" of newspapers. Of the major publications
writing on the film, only New Yorker gave the film a
negative review--this negative review in September, by influential
critic Pauline Kael, seems to be more a reaction against the cult
status the film had earned by that time.
Empire:
Empire reviews were very often tinged with
disappointment. Particularly, critics felt that the film lacked the
humour and sense of fun that the original had. However, even these
"disappointed" reviews gave the film a relatively good rating,
usually 3/5 (60) or 3.5/5 (70). On the score results, this means the
film overall has a much lower rating than Star Wars,
but a better tomatometer rating--which means, basically, that there
was less ultra-low reviews and more general recommendations (hence
the tomatometer score), but less overall enthusiasm (hence the
overall score).
Jedi:
Jedi has often been noted as the disappointing
film of the trilogy, and the one that earned the lowest ratings.
This is true, but the picture is not so simple. Many felt it was
actually the strongest of the three. On the other hand, many
vehemently despised it. The true picture that emerges of
Jedi is this: many disliked the film, but many liked the
film as well. Most who disliked Empire did not find
anything positive to say about Jedi, and some who
liked Jedi were nonetheless disappointed it wasn't as good
as the others, but a great many also felt that it lived up to, if
not surpassed, any of the previous films. The result is that the
film was often very divisive, which is why it scores the way it
does. There were more positive reviews than bad ones, but the bad
reviews were more frequent than the previous films and frequent
enough that they brought all the positive reviews down in the score,
and it had less 4/5 and 5/5 per capita. Its tomatometer and overall
scores are not nearly as good as the other films because of this,
but the film was critically successful by these measurements all the
same. The difference was that it wasn't as unanimous as the first
two films. A final observation of note is that even among many
enthusiasts the sentiment is that the series had run its
course--Jedi was either the series' entertaining swansong
or its final, bloated groan, but the series was finally over and the
world can move on.
Here is a useful calculator for the
ratings:
1/5=20
1.5/5=30 2/5=40
2.5/5=50
3/5=60
3.5/5=70 4/5=80
4.5/5=90 5/5=100
The raw data:
|
Star Wars |
Empire |
Jedi |
Boston Globe |
|
3.5/5 my 21 |
4.5/5 my 25 |
Calgary Herald |
|
3.5/5 my 23 |
4/5 my 26 |
Chicago Sun Times |
4.5/5 my 25 |
|
4/5 my 25 |
Chicago Tribune |
3.5/4= 4.5/5 my 27 |
3.5/4=4.5/5 my 21 |
5/5 my 25 |
Christian Science Monitor |
4.5/5 jn 2 |
4/5 my 21 |
4/5 my 19 |
Cineaste |
|
|
3/5 |
Globe and Mail |
|
2.5/5 my 22 |
3.5/5 my 27 |
Films and Filming |
4/5 dc |
|
2/4?=3/5 july |
Films in Review |
4/5 ag-sp |
4/5 ag-sp |
3/5 jun-july |
Hartford Courant |
|
4/5 my 23 |
3/5 my 26 |
Los Angeles Times |
5/5 my 22 |
4/5 my 18 |
4/5 my 25 |
Macleans |
4/5 jn 27 |
3.5/5 my 26 |
3.5/5 my 30 |
Miami Herald |
|
|
3/4=4/5 my 25 |
Monthly Film Bulletin |
|
|
4/5 july |
Nation |
4/5 jn 25 |
3.5/5 jn 21 |
2/5 jn 18 |
National Review |
|
|
2.5/5 jn 24 |
New Leader |
|
|
1/5 my 30 |
New Republic |
2.5/5 jn 18 |
3/5 my 31 |
N/A |
New Statesman |
|
|
1/5 jn 3 |
New York |
2/5 jn 20 |
3/5 my 26 |
3/5 my 30 |
New York Daily News |
|
3.5/4 = 4/5 |
4/4 = 5/5 |
New York Post |
|
|
2/5 my 25 |
New York Times |
4/5 my 26 |
4/5 my 21 |
2/5 my 25 |
New Yorker (other) |
4/5 jn 13 |
N/A |
N/A |
New Yorker (kael) |
1.5/5 sp 26 |
4/5 my 26 kael? |
1/5 my 30 |
Newsday |
|
|
3/5 my 25 |
Newsweek |
5/5 my 30 |
4/5 my 19 |
3/5 my 30 |
People Weekly |
N/A |
|
3/5 my 30 |
Sarasota Herald-Tribune |
5/5 jn 2 |
|
|
Time |
5/5 my 25 |
4.5/5 my 19 |
5/5 my 23 |
The Times |
5/5 dc 16 |
2/5 my 23 |
3.5/5 jn 3 |
Toronto Star |
5/5 jn 11 |
3.5/5 my 22 |
4/5 my 25 |
Vancouver Sun |
5/5 jn 25 |
3.5/5 my 16 |
3/5 my 25 |
Variety |
5/5 my 25 |
4/5 my 7 |
3.5/5 my 25 |
Washington Post (other) |
NA |
3/5 my 23 |
NA |
Washington Post (G.Arnold) |
4.5/5 my 25 |
5/5 my 18 |
3.5/5 my 22 |
Washington Spokesman-Review |
4/5 july 2 |
|
|
Washingtonian |
|
2.5/5 july |
2/5 aug |
Winnipeg Free Press |
2.5/5 jn 24 |
5/5 jn 18 |
4/5 my 25 |
TOTAL
AVERAGE |
82 |
73 |
64 |
Converted
to RT's "tomatometer", which is calculated by the number of
"fresh" reviews (60% rating or more) divided by total reviews, this
results in:
Star
Wars: 83%
Empire
Strikes Back: 92%
Return
of the Jedi: 76%
Part
II of this study has more detailed analysis of what the
above scores can tell us.
Below
are weighted scores, based on stature, influence
and readership of the publication. Again, this is based on my
own interpretation. Sources such as Time, Variety, New Yorker,
New York Times, Chicago Tribune and Sun-Times, L.A. Times,
Toronto Star, Newsweek, Macleans, and The Times were
given five times the weight of all other sources.
|
Star Wars |
Empire |
Jedi |
Weighted Total
Average |
86 |
75 |
68 |
Based
on this interpretation, the
weighted "tomatometer" scores would be
equivalent to:
Star
Wars: 89%
Empire
Strikes Back: 89%
Return
of the Jedi: 79%
Below
are scores based on date of review for Star
Wars. I have grouped this into three waves. First is
the reviews based on the 32 theatres that the film first opened on
during May. The second is the expanded release in June--for
instance, the film did not play in Canada, even in Toronto, until
the second week of June. The third wave of reviews is based on
anything written in July and later, which includes London's The
Times as the film did not open in the U.K. until the end of the
year. The sequels are largely exempt from this, as they more or less
opened widely at the same time (there are limited exceptions--for
instance, Empire mandated that all theatres it opened on
had to be 70mm equipped, which prevented it from playing in Winnipeg
until June; however, this is the only exception in the
sampling). Some monthly and weekly publications reviewed the
sequels in their June and July issues, but this is more due to
the deadlines of monthly/weekly printing, which must have copy ready
weeks before the shipment date. For Star Wars, however, I
have included all monthly/weekly reviews from July and later in wave
three. I also included the Sarasota Herald-Tribune and
Christian Science Monitor as wave one, as they were printed
on June 2 and reflect the first wave. These stats are of slightly
limited use, however, due to the smaller sampling rate for the
first film, especially in wave three, where most of the U.K. is not
represented, however I felt it would be an interesting experiment
nonetheless.
WAVE ONE |
|
Chicago Sun Times |
4.5/5 my 25 |
Chicago Tribune |
3.5/4= 4.5/5 my 27 |
Christian Science Monitor |
4.5/5 jn 2 |
Los Angeles Times |
5/5 my 22 |
New York Times |
4/5 my 26 |
Newsweek |
5/5 my 30 |
Sarasota Herald-Tribune |
5/5 jn 2 |
Time |
5/5 my 25 |
Variety |
5/5 my 25 |
Washington Post (G.Arnold) |
4.5/5 my 25 |
WAVE ONE
AVERAGE: |
94 |
|
|
WAVE TWO |
|
Macleans |
4/5 jn 27 |
Nation |
4/5 jn 25 |
New Republic |
2.5/5 jn 18 |
New York |
2/5 jn 20 |
New Yorker (other) |
4/5 jn 13 |
Toronto Star |
5/5 jn 11 |
Vancouver Sun |
5/5 jn 25 |
Winnipeg Free Press |
2.5/5 jn 24 |
WAVE TWO
AVERAGE |
73 |
|
|
WAVE THREE |
|
Films in Review |
4/5 ag-sp |
New Yorker (kael) |
1.5/5 sp 26 |
The Times |
5/5 dc 16 |
Washington Spokesman-Review |
4/5 july 2 |
WAVE THREE
AVERAGE |
73 |
Finally,
I present below the reviews taken from more academic sources during
1977-1983. Some of these are included in the master "popular" list
because they skirt the line--Films in Review and Films
and Filming might be considered academic by some, but their
approach and style is often the same as the popular sources (for
that matter, New Yorker might be considered an academic
source due to its self-esteemed "high brow" approach to art), and
skirt the line between "insider" industry publications as well
(Variety, for example, is not normally read by the public,
as it is an "insider" trade magazine, but appeals to
non-academics). I include these sources in their own category for a
number of reasons. First, their obscurity relative to the popular
sources. Second is that the academic world is in many ways its own
bubble, with its own sets of tastes, values and politics, and it
seemed unfair to group these together with the popular sources. For
the sake of completism, I do however offer an adjusted score
accounting for both popular and academic together.
One
bit of commentary I should offer here is that academic sources
largely paid no or little attention to the film--just based on the
film journals, you would hardly know that such a cinematic cultural
landmark had landed. There are few reviews, and even in the years
following, very little analytical articles focusing on the
films.
-ACADEMIC ARTICLES WILL BE POSTED
SOON-
Conclusion and
Part II
Altogether,
these stats show us many things; because of potential limitations in
sampling rate, they should not be taken as gospel, but at the very
least considered an indicator. I attempt to analyze and interpret
them in part two, and also offer more detailed probing into what the
reviewers actually said, and why. I also attempt to contextualize
them. Following that, I offer further study into how the reviews and
receptions of the films shifted (or didn't) in the two time periods
that followed--1984-1996 (between the theatrical releases), and
1997-present (during and after the theatrical re-release and prequel
trilogy). With that, turn to where the real information is,
in Part II.