So far in the examination of top prospect lists, we’ve introduced our Baseball America dataset and examined prospect performance by ranking, role, and age. This time, we’ll be looking at a trait that is less inherent to the player than either of the last two breakdowns. The question at hand: has the player appeared in a previous top 100 list?
A couple of notes to start off. First and most obviously,
this question inherently limits our data even more than usual; we can’t
use the 1990 list to find the answer, since there was no previous top prospect list
that could have included those players. That leaves us with 2200 player-entries
rather than 2300. Second, I’m not technically looking at whether the player has
appeared on any prior list, I’m looking at whether he was on last year’s list
specifically. This should make a minimal difference; there are some dropouts
who later return, but likely less than 100 in the sample. In exchange for this small loss of information, we get two benefits: this approach is easier to implement, and it opens up an extra check we can run.
For starters, let’s do the simple yes-or-no question: How
many players in our sample were repeaters? Of the 2200 entries we’re
considering, 909 had been on last year’s list (about 41%); 1291 had not. By
ranking bucket, here’s the breakdown:
Rank |
Repeater |
Newbie |
1-10 |
163 |
57 |
11-25 |
172 |
158 |
26-50 |
253 |
297 |
51-75 |
179 |
371 |
76-100 |
142 |
408 |
Unsurprisingly, it’s far more common for a top-10 player to
be a returner than for someone toward the bottom of the list. The other 26% of
the top 10 will be predominantly players who weren’t eligible for the list in
prior seasons; it would be a rare prospect indeed who made the leap from
afterthought to future star in one year.
Usually, we would jump into percentile tables here – but
this time, I’m going to precede the tables with the introduction of our second question of the day. We can
easily check whether the player was on last year’s list, but why stop there? If
he was on last year’s list, where did he appear, and how does that compare to
this year?
This is a trickier question than one might think. The
immediate instinct is to examine this by looking at change in ranking, but that
is easily dismissed as a viable option; going from #93 to #73 is a very
different outcome than going from #23 to #3, even if both are 20-position
improvements. I settled on an option that made me wish I’d been using different
ranking buckets in all the prior posts in this series: square roots. If the
square root of a player’s ranking decreases by at least 2 between seasons (say,
if the ranking goes from 100 to 64, or 25 to 9), we’ll call that a large jump;
if it changes by the same amount in the other direction, we’ll call it a large drop. Everything else,
we’ll consider reasonably similar to the prior year. Here are the player totals
in each category:
Rank |
Jump |
Similar |
Drop |
1-10 |
77 |
86 |
0 |
11-25 |
70 |
94 |
8 |
26-50 |
85 |
137 |
31 |
51-75 |
5 |
122 |
52 |
76-100 |
0 |
78 |
64 |
The 0’s are there for obvious reasons; it’s impossible to
have your square root improve by 2 and stay between 76-100, and nearly
impossible to do the opposite while staying in the top 10. The 8 and 5 totals
in the next categories toward the middle are small enough both to demonstrate
the unlikelihood of those criteria being met, and to be discarded as usable
samples. But the other groups are big enough to be functional.
(Just for the sake of completeness: The five jumpers in the
51-75 range include four unimpressive results, but the fifth player was Jim
Thome. Three of the eight droppers from 11-25 posted at least 10 WAR: Brad
Penny, Josh Hamilton, and Desmond Jennings. Hamilton is more or less the
essence of what you hope to eventually get out of a declining prospect, as he
eventually recovered to have a few seasons consistent with his initial hype.)
On to the percentiles! We’ll start with top-10 players,
broken down between new, similar, and jump:
Percentile |
Jump |
Similar |
New |
90 |
57.6 |
59.1 |
56.6 |
80 |
41.6 |
41.8 |
39.3 |
70 |
34.5 |
27.9 |
30.5 |
60 |
23.6 |
17.2 |
21.2 |
50 |
18.3 |
11.9 |
15.9 |
40 |
11.9 |
7.3 |
12.0 |
30 |
9.1 |
5.0 |
7.5 |
20 |
7.3 |
0.5 |
3.2 |
10 |
2.5 |
0.0 |
0.0 |
Not overpowering differences, especially at the top end. But
for anything between 20th and 70th percentile, you’d rather have someone new to
the top 10 (one way or another) than someone who’s hung out there for a while. There may be something to this, as the lion’s share of top-10 prospects will usually graduate to the majors by the end of the season if they perform well and stay healthy. However, our method of selection may also be contributing to this effect, as
jumpers are more heavily weighted toward the top of the rankings than players
who stayed in similar ranges (that is, a player ranked #2 is more likely to be
a jumper than a player ranked #8).
Next group, rankings 11-25:
Percentile |
Jump |
Similar |
New |
90 |
49.3 |
35.0 |
43.4 |
80 |
31.4 |
26.8 |
26.4 |
70 |
27.8 |
19.2 |
18.3 |
60 |
16.2 |
14.4 |
12.3 |
50 |
9.5 |
8.7 |
7.2 |
40 |
6.0 |
6.6 |
4.1 |
30 |
2.8 |
2.5 |
0.6 |
20 |
0.4 |
-0.1 |
0.0 |
10 |
-0.2 |
-1.6 |
-0.8 |
The jumpers stand out again, this time more on the high end
than toward the middle. Again, there’s likely some quality leakage going on
here; ranks 11-13 include 23 jumpers over our sample, while ranks 23-25 include
only eight. By comparison, I’m surprised how competitive the new player
category is, as this group skews lower in the rankings.
With that, let’s get to the middle group (26-50), the only
one in which we have both Jump and Drop categories in a reasonable sample size:
Percentile |
Jump |
Similar |
Drop |
New |
90 |
37.1 |
34.3 |
38.0 |
31.9 |
80 |
20.4 |
17.7 |
21.2 |
18.6 |
70 |
12.9 |
11.5 |
13.5 |
12.3 |
60 |
8.2 |
6.5 |
8.9 |
8.7 |
50 |
6.0 |
5.0 |
5.6 |
4.2 |
40 |
2.6 |
1.7 |
2.8 |
1.5 |
30 |
1.4 |
0.1 |
0.0 |
0.1 |
20 |
0.0 |
-0.4 |
-0.3 |
-0.2 |
10 |
-0.8 |
-1.0 |
-1.2 |
-0.6 |
Unfortunately, what might in theory have been the most
interesting group (since it covers the most categories) produces results that honestly
look like a lot of nothing. The Similar group is maybe a bit on the low end (it doesn’t
lead a single category), but not significantly so; meanwhile, Drop scores highest in the upper percentiles, but by small margins each time. Let’s move on to the 51-75
and 76-100 tables, in order.
51-75:
Percentile |
Similar |
Drop |
New |
90 |
27.5 |
20.3 |
28.2 |
80 |
19.6 |
13.2 |
15.2 |
70 |
14.3 |
8.7 |
8.5 |
60 |
8.4 |
3.5 |
4.6 |
50 |
2.0 |
1.6 |
1.4 |
40 |
0.3 |
0.7 |
0.1 |
30 |
0.0 |
0.0 |
0.0 |
20 |
-0.5 |
-0.1 |
-0.3 |
10 |
-1.3 |
-1.2 |
-0.9 |
Similar stands out nicely in the upper middle here, which is
a nice change of pace compared to its performance in every other ranking split so far.
76-100:
Percentile |
Similar |
Drop |
New |
90 |
15.3 |
15.1 |
22.4 |
80 |
6.5 |
11.9 |
12.9 |
70 |
2.1 |
5.8 |
8.4 |
60 |
0.4 |
1.6 |
4.7 |
50 |
0.1 |
0.8 |
1.4 |
40 |
0.0 |
0.0 |
0.1 |
30 |
-0.6 |
0.0 |
0.0 |
20 |
-0.9 |
-0.3 |
-0.3 |
10 |
-1.3 |
-1.1 |
-1.0 |
This one makes sense to me – your 76-100 prospects are guys
that you don’t have enormous expectations for, but you hope they’ll progress.
If a guy is in that range twice in a row, he failed to progress the prior year.
If he dropped into that range, then you actually had higher hopes for him
previously, and if he can put it together then the original expectations might
still be realized. And if he’s new, then the initial hope of progress hasn’t
yet been denied. (None of the medians are especially impressive, but that’s to
be expected from the bottom group.
Conclusions appear to be a little thin on the ground this
time, but I’ll throw a couple tentative ones out there. Even accounting for
quality leakage, I suspect you’d still slightly prefer a jumper in the top 25
over someone who’s stayed steady. And at the bottom of the rankings, it seems
pretty clear that you’d rather have a new guy over a repeater. Beyond that? It
looks like BA’s evaluators have again done a pretty solid job.
Up next, we’ll move on from year-to-year top 100 standing, and instead look at what we can glean from the most basic information about the teams’ own prospect evaluations: draft position.