The Wins Produced model begins with a simple idea. Wins are a function of offensive and defensive efficiency. This idea can be found in the work of John Hollinger and Dean Oliver. And with a bit of math, you can show that wins are indeed a function of these efficiency metrics.
Once we see that wins are a function of offensive and defensive efficiency, a simple regression (of wins on the efficiency metrics) allows us to determine the value, in terms of wins, of the following statistics tracked for a player: three point field goals made, two point field goals made, free throws made, field goals attempted, free throws attempted, missed field goals, missed free throws, offensive rebounds, defensive rebounds, turnovers, steals, and personal fouls. Now the efficiency metrics are not only comprised of these player statistics. From the wins regression we can also derive the value for the opponent’s three point field goals made, opponent’s two point field goals made, opponent’s turnovers (that are not steals), and team rebounds. For the most part, these team factors are associated with defense. Before I get to these team defense statistics, let me review two basic lessons from this wins model.
Wins are a function of offensive and defensive efficiency. This is not only what theory tells us, but also the empirical evidence. The efficiency metrics explain 95% of wins and hence provide an accurate depiction of the quality of a team.
Each element in the efficiency metrics is linked to actions taken by the players on the floor. It is the players who are responsible for made shots, rebounds, turnovers, team defense etc.. Thus, our wins model – because it accurately links these actions to wins – allows us to accurately value the contributions of an individual player. In sum, Wins Produced links all the elements of the wins model back to the players. Hence, Wins Produced also explains 95% of team wins.
For the player statistics, evaluating the player’s contribution in terms of wins is easy. We simply need to multiply a player’s statistics by the corresponding value. The result of this calculation is then compared to the average performance at the player’s position.The team variables, which are associated with team defense, are not tracked for individual players. So how can we incorporate these factors? Here is what we say in the book:
…we have constructed what we call the team’s statistical adjustment. We then follow a convention we have observed and personally employed in the economics literature. Specifically we allocate the impact of the team statistical adjustment according to the minutes each player was on the court.
The convention we followed was originated by Frank Scott, James Long, and Ken Sompii. In 1985 these authors published an article in the Atlantic Economic Journal examining the link between salary and performance in the NBA. For the team factors employed in their model these authors simply allocated each factor across the players according to minutes played.
What does that mean? We don’t have a measure of defense for each player. What we do have, though, is knowledge about how good the team played defense. We argue that if a player played 15% of the team’s minutes, then he is responsible for 15% of the team’s defense [or specifically, the opponent’s made field goals, opponent’s turnovers (that are not steals), and team rebounds accumulated].
And this makes some sense. Defense, especially in today’s NBA, is a team effort. If your team is good at this, you should receive credit. If not, you should be penalized. The Wins Produced metric, by allocating the opponent’s made field goals, opponent’s turnovers, and team rebounds across the players, thus takes into account the quality of a team’s defense.So as one can see, the infamous team statistical adjustment is simply a measure of team defense. Despite its simplicity, there has been a bit of confusion regarding what this is and how it impacts our results. Let me try and answer these questions.
1. Is this a “giant fudge factor”?
In simple terms, no. Wins are a function of offensive and defensive efficiency. The factors comprising what we called the team statistical adjustment are the opponent’s field goals made, opponent’s turnovers (that are not steals), and team rebounds. In essence, this is the quality of the team defense. The players are the individuals playing defense, so they should get credit if a team does well or poorly with respect to this aspect of the game.
This factor has been described as an addition to our wins model designed to increase its accuracy. That charge is false. The opponent’s statistics are part of defensive efficiency, not an addition to the model.
People have also said that you could take points scored and a “team adjustment” and predict wins just as accurately as we do. I do not know what the author of this statement means by a “team adjustment.” If you employ the team factors listed above, this statement is false. To demonstrate this point, I regressed wins on points scored, opponent’s three point field goals made, opponent’s two point field goals made, opponent’s turnovers (that are not steals), and team rebounds. These five factors explained 79% of team wins, not the 95% that is explained by offensive and defensive efficiency.
Again, the team statistical adjustment is not some random factor we yanked out of our ass. It is simply team defensive factors that must be accounted for in the measurement of defensive efficiency and player performance.
2. How does this impact our evaluation of players?
What I have mentioned, more than once, is that these team defensive variables do not have much impact on our evaluation of players. PAWSmin, which is position adjusted win score, does not consider these team variables. WP48, or wins produced per 48 minutes, does have these team variables. The correlation between these two is 0.994. So the team defense measures do not matter much in our evaluation of players.
3. Can you predict wins without the team defensive factors?
People have asked: Can you predict well without the team variables? And my answer is “I don’t think so, but then again, I haven’t looked.” Without the opponent’s field goals made, the opponent’s turnovers, and team rebounds, the wins model is mis-specified (in other words, you have left out relevant variables, which introduces a number of statistical issues for your model).
Nevertheless, since people have been clamoring to know, I thought I would look. Without the opponent’s field goals made, opponent’s turnovers, and team rebounds, we are explaining wins with the following variables: three point field goals made, two point field goals made, free throws made, field goals attempted, free throws attempted, missed field goals, missed free throws, offensive rebounds, defensive rebounds, turnovers, steals, and personal fouls. A regression of wins on these factors reveals that 84% of wins can be explained without opponent’s field goals made, opponent’s turnovers, and team rebounds.
Of course the model is mis-specified and the results cannot really be used to evaluate individuals. The easy fix to this problem is to specify the model in a fashion that is theoretically correct, which we did in The Wages of Wins.
Getting Rid of the Team Adjustment
Is it possible, though, to do “better”? Maybe we can change how we allocate opponent’s field goals made, opponent’s turnovers, and team rebounds. 82games.com has several measures of how well a team’s opponent does when each player is on the court. If we use one of these measures, we can scrap the team statistical adjustment (which allocates defense in terms of minutes played), and allocate the defensive variables across each player according to their ability to play defense.
I did this for the Washington Wizards. 82games. com reports how well the opponent scores with each player on the court. The team surrenders about 104 points per game. From 82games.com we see that when Jarvis Hayes is on the court the team gives up 106.2 points per game. So by this measure he is a relatively poor defender. In contrast, Brendan Haywood gives up 100.6 points, so he appears to be a relatively better defender. Previously we are allocating the opponent’s field goals made and opponent’s turnovers in terms of minutes played, so each player – on a per-minute basis – was evaluated the same on defense for each team. There were differences across teams, but on any one team it was the same.
Now we have a measure that differentiates players on a team. What happens if we allocate the opponent’s statistics with the 82games.com measure of defense? The results reveal a difference, but not very big.
Table One: Examining the Impact of Team Defense with the Washington Wizards
Brendan Haywood improves, posting a WP48 of 0.133 when we treat every player the same and a mark of 0.142 when he is credited for his defensive ability. Jarvis Hayes looks a bit worse. In general, though, the results are very similar.This should not surprise. There are apparently large differences in the defensive abilities of teams. And these differences across teams are incorporated in the evaluation of players with Wins Produced. Yet we found that with or without adjusting for these factors our evaluation of players was essentially the same. Now when we do this for an individual team, we tell the same story.
The research at 82games.com, if it accurately measures individual defense, allows us to get rid of the team adjustment. The model still explains 95% of wins. It is also still tells the same story about player performance. And what is that story? The primary factors in evaluating a player are shooting efficiency, rebounds, and turnovers.
A Comment on Baseball
The box score statistics in basketball are often discredited because these do not account for on-the-ball defense. We can see defensive ability at the team level, but without the work at 82games.com we cannot get at the individual’s contribution. Does the inability of the standard box score statistics to measure on-the-ball defense really matter in our evaluation of players?
I want people to step outside basketball for a moment and think about baseball. In baseball there is the issue of fielding. OPS and linear weights are measures of performance in baseball. They both ignore fielding entirely. Yet, very often (not always, but often) people discuss the merits of a player strictly in terms of his offensive output.
Of course there is also the issue of evaluating pitchers. A pitcher cannot strike everyone out. Pitchers depend on fielders. Yet in evaluating pitchers, I often see people reference statistics (ERA, Wins, etc…) that do not separate the pitcher from his fielders.
Imagine if we regressed wins in baseball on a team’s ability to hit, pitch, and field. We would be able to explain about 90% of wins, which is what happens when you regress wins on runs scored – your ability to hit — and runs surrendered – your ability to pitch and field.
Now let’s say you isolated a pitcher’s contribution from his fielders. Then you regressed wins on runs scored (what the hitters did) and just the pitcher’s contribution. How well would explain wins now? I do not know the exact answer, but without any measure of fielding your explanatory power would have to be a good deal less.
Despite this issue, traditional measures of hitters and pitchers do not consider fielding. Is this a problem? Should we throw out all the traditional box score statistics in baseball because they ignore fielding, a factor that must have some ability to explain wins?
The answer is of course not. Yes, fielding matters. But our evaluation of players will generally not change very much if we ignore this factor. Barry Bonds and Manny Ramirez will still be great players. Greg Maddux will still be a great pitcher.
I think the same thing can be said for defense in basketball. Yes some players are better at on-the-ball defense than others. But if it truly mattered a great deal we would see vast differences in our evaluation of players when defense is included and when it is not. We would also see that rebounding totals fluctuate for players dramatically depending on who their teammates were. When we look at the data, though, we see none of this.
Perfect Models?
I want to return to a statement I made a couple of weeks ago. Models are not supposed to be “perfect” (whatever that means). When I and my colleagues construct models, we are trying to construct a simplified version of reality that allows us to focus on what is important (and answer the various questions we pose in our research).That is what I think Wins Produced does. It is a simple and accurate measure of performance, based on the theoretically sound idea that wins are determined by a team’s offensive and defensive efficiency. This model ultimately tells us that wins are primarily determined by shooting efficiency, rebounds, and turnovers. Yes, other issues matter. But players who do not score efficiently, who fail to rebound (given their position), and/or turn the ball over excessively, will not help you win games.
Let me repeat what I said a few days ago:
Now it’s not the case that factors like blocked shots, assists, and personal fouls don’t matter. But none of these factors are as important as shooting efficiency, rebounds, turnovers, and steals. And once we see this, we can understand the outcomes we observe.
For example, the Rockets lost Yao Ming to a devastating injury, yet managed to maintain their winning percentage. Once we see the importance of rebounding, though, we can see how having an extraordinary rebounder like Dikembe Mutombo come off the bench mitigated the loss of Yao.
We also see why the 76ers improved after Iverson departed. Iverson has problems hitting shots and avoiding turnovers, so despite his scoring totals, our model tells us he does not produce as many wins as his star power suggests. In other words, we should not have expected a team that replaced Iverson with Andre Miler to get worse (as many people who focused on scoring predicted).
And on Collaboration
I have written a paper entitled “A Simple Measure of Worker Productivity in the National Basketball Association.” An earlier working version of this paper – which details the math behind the Wins Produced metrics — was referenced in The Wages of Wins. Later this year this paper should finally be published. Although I am the solo author on this paper, one should not think that Wins Produced and the related models are only the results of my efforts. As noted in The Wages of Wins, many conversations with Dean Oliver clearly impacted my thinking.
In addition, a collection of co-authors also should be credited (or blamed, depending on your perspective). At the top of this list are Martin Schmidt and Stacey Brook (co-authors on The Wages of Wins). The Wins Produced metric has also been employed in papers I have written with Anthony Krautmann, Aju Fenn, Bernd Frick, Erick Eschker, Young Hoon Lee, Rod Fort, Michael Leeds, and Michael Mondello.
This is how I described my tendency to collaborate at my website in April of 2006: Beyond my publication rate, we can also see that virtually all of my work is co-authored. Over the course of my career I have published research with eleven different writers, with Martin Schmidt being my most frequent collaborator. In all, Marty and I have published twelve papers together. One can look at this as I work well with others, or lack the skills to complete projects by myself. I like the “works well with others” story, although my many co-authors might play up the “lacking skills” angle.
Almost all of my work is a collaboration with someone else. All of these c0-authors are accomplished researchers who have the corresponding list of academic publications as evidence of their abilities. But even with their help, academic research still requires the help of more people.
My list of co-authors leaves out the various editors and referees who review our academic papers. Research in academia is never a solo effort. Other people comment and critique your work constantly, and these efforts improve the final product. Hence when people argue that I do not “collaborate” I am left very puzzled. Academic research – be it in journals or at academic conferences – is always a collaboration. And anyone with an understanding of academic research would see this point.
The Future
We are on the quarter system at California State University – Bakersfield. This is a good deal in the fall, since we don’t begin until late September and finish around Thanksgiving. In the spring, though, it’s a bit of drag. Today we begin the Spring Quarter, which does not end until June. This quarter I am teaching three classes. I also have a paper to complete for the Western Economic Association and several other research projects to finish.
All of this means that over the next 10 weeks, postings in this forum will be less frequent. I will simply not have time to post as often as I have across the last year. Hopefully this summer I will have time to resume posting on a daily basis. Thanks to everyone who makes this a part of their day. Unfortunately, for a few weeks at least, this will no longer be a part of each day for me.
– DJ
Mark T
March 26, 2007
Good post. Maybe I am missing something, but I never thought of WP48 as a predictor of future team performance. I always thought of it as a way to help value the individual players across statistical categories. Which is different from team-centered predictions. I do think there is much to do on the defensive side of this kind of analysis and hopefully data compilation will evolve to enable that sooner rather than later. But take a look for example at the March 21 blog entry for Mavsmoneyball.com which breaks down 7 minutes of the Mavs defense against the NYK and you’re reminded how hard it is. Were I retired from my day job, and my kids out of the house, I would gladly contribute some time to compilation of the data but, for now, I just kibitz. Enjoy your break!
Mark from APBR
March 26, 2007
This was a welcome post and I will think about it and re-read it before commenting or asking questions here now though I have made some recent posts at apbr on related points. I take note of and try to make use of the analysis and discussion in this blog. While I may disagree with portions that is true for other methods, other writers. And my thinking isnt set in stone and necessarily right on all questions yet but evolves as I am exposed to alternative viewpoints, data and methods.
anon
March 26, 2007
A post idea:
not sure if you’ve covered this, but it might be interesting to see what Wages of Wins has to say about the myth of players who make their teammates better. I know you’ve tackled this a bit with respect to Steve Nash. But it would be interesting to explore whether there is any evidence that someone like Jordan or Bird made their teammates better.
My intuition is there isn’t and that the myth is just around to provide fuel for critiques of great players surrounded by a lousy supporting cast (I’m specifically thinking about Kobe).
Mark from APBR
March 26, 2007
Ok I’ve re-read and thought about a few things and will post while they are fresh.
Using team defense while a player is on the court does seem better than minutes based share of overall defense but I have thought of “defense” as being somewhere between a 50-70% local product of the counterpart and 30-50% team as a whole. If defense scoring were blended with offense on that basis WP or any metric would have the potential to change more significantly and I’d support such a change as better reflecting reality and player value.
Second on the comparison to baseball, isn’t the spread between good defense and bad defense in basketball much bigger than the spread between the two in baseball?
Mark from APBR
March 26, 2007
With respect to baseball defense, I was referring only to fielding above but since I am not a knowledgeable baseball fan able to separate pitching from fielding impacts I probably shouldnt have commented on that comparison.
Julian
March 26, 2007
Hi Dave
Great post.
I follow your work from Spain where I’m trying to do some things with stats in spanish league.
I believe in plus/minus stats used to complete the WP48 but it’ll be better if this plus/minus use the possessions of each player instead of minutes. Only this you will have a great idea of the impact of the player in the team defense considering that plus/minus is a stat useful when you have a great amount of sample.
Okapi
March 27, 2007
Nice post.
A couple annotations ..
(1) There is a popular Sabremetrics stat, DIPS (defense independent pitching stats)– http://www.baseballprospectus.com/article.php?articleid=878
(2) Manny Ramirez is bad at defense so any evaluation of his worth should incorporate that element. There was a column in the NY Times a week and a half ago about exactly this subject– http://www.nytimes.com/2007/03/18/sports/baseball/18score.html?ex=1331870400&en=d59fbc0208da1586&ei=5088&partner=rssnyt&emc=rss
Huey
March 27, 2007
What would it take to get your colleagues to post more often? :)
Jake
March 28, 2007
I’d like to offer a possible explanation about why player evaluations do not change much when you change from a playing time based analysis of defense to one like you explain here.
It could be that NBA coaches are good at recognizing which players are better on defense, and play them more minutes. For instance, Andre Iguodala is considered a top defender, he also is in the top 5 in the league in minutes per game. meanwhile, other players do not play as many minutes, in part because they are not as good at defense.
therefore, assigning the team adjustment for defense based on playing time has more of an effect on the rating of the players whom the coach leaves on the court the most.
Mark
March 29, 2007
The last comment seems like an important path to me. But the current team adjustment based on team defense overall is actually more the defense from the players on the court a lot than not. So if it changed to team defense when on court these players might not change as much as in the cases of true bench players whose play is a minority of time. Nonetheless there are good and bad defenders who play a lot. Unless individual one on one defense gets a fairly large weight there will be players who unfairly get team defense scores that only have a small, watered down element that is their impact and make that score too high or too low.
dberri
March 29, 2007
Hi everyone,
Let me try and respond to a few comments.
anon… in The Wages of Wins we talk about The Law of Diminishing Returns. What this means for the NBA is that the more productive your teammates, the less productive you will be. In other words, Michael Jordan’s immense productivity levels probably reduced the productivity of his teammates.
Mark from APBR… not sure what you mean by the following: “Using team defense while a player is on the court does seem better than minutes based share of overall defense but I have thought of “defense” as being somewhere between a 50-70% local product of the counterpart and 30-50% team as a whole.”
Okapi,
I agree that DIPS makes an effort to separate the pitcher from his fielders. I also agree that Manny Ramirez is bad at defense. Not sure most GM’s wouldn’t still like to add him to the line-up. I could be wrong, but I think his offensive abilities still overwhelm his defensive lapses.
Huey,
Stacey teaches more than I do. So he can post occasionally, but not that often. Marty is really not interested in posting anymore. I will try and keep posting once or twice a week, as my schedule allows.
Thanks to everyone for the comments. I do read these. Responding to these frequently, though, is going to be a challenge.
Jason
March 29, 2007
It is the rare Major Leaguer whose defense is so bad as to use it as a reason to keep a productive bat out of the lineup.
Mark from apbr
March 29, 2007
Dave, sorry I was unclear.
I just meant on average I assign 50-70% of the responsibility to stopping a player from scoring or not to the man assigned to guard him. The other 30-50% I count as team responsibility to cover switches, zone coverages, etc.
The 82 games shot defense data is imperfect but it can help in many cases to identify above average, near average and below average 1 on 1 defenders on a team. Defense seems like it is better modelled as the sum of 2 weighted scores- local and team as whole so everyone doesnt get the same or close to the same team defensive score. There will still be interactions from having a Duncan behind you or a Duhon in front of you that may affect your individual defensive score. Perhaps analysis of defensive results adjusted by who is on the floor would be the next step.
David Arnott
April 1, 2007
Yes, fielding matters. But our evaluation of players will generally not change very much if we ignore this factor. Barry Bonds and Manny Ramirez will still be great players. Greg Maddux will still be a great pitcher.
I think the same thing can be said for defense in basketball.
What? Are you kidding? This is from the co-author of a book that caused an uproar by saying Allen Iverson hasn’t been as great a baller as he’s generally credited because people tend to improperly weight the different aspects of a basketball player’s production. The EXACT SAME THING happens in baseball. Ignoring defense is improperly weighting the available data. To wit, who is the better third baseman: Aramis Ramirez or Adrian Beltre? Ramirez just got a $75 million contract, and most baseball fans would say Beltre isn’t worth nearly that much. However, Dave Cameron of USSMariner explains why Ramirez and Beltre are essentially equal. Armando Benitez in 2004 was a pretty good, not historically awesome, pitcher, the surface stats be damned, and I say that based on data that separates his contribution and his team’s contribution to the production credited to him. I see no reason why, once the proper data is collected, defense can’t be similarly incorporated into NBA player evaluation. Really, it’s only a matter of time until someone designs a computer program to examine footage of games and evaluate how much space each defender controls.
Mark
June 26, 2007
Maybe at some point adjusted team +/- for defensive side of play only will be available for current season data and included in player ratings. Defensive evaluations short of this all seem to have flaws. Til then most player ratings handle offensive better and most overrate it and undercapture or miscapture individual defensive quality.
Mark
June 26, 2007
Actually, while adjusted team +/- for defense would have value/ relative strength, it seems to assume consistency of defensive effort & affect and certainly there is variation in player behavior and variation in effeciveness in certain combinations and facing certain opponents. Ultimately intelligent and partial credit /demerit scoring of defense on videotape (thru use of Synergy database) would be the right answer and teams will invest in that understanding to the extent they wish. Unlikely to ever be complete and in public. So it goes. Folks will get by with lesser approaches or tacked on subjective understanding of defensive value.
Bill Schauer
September 3, 2007
I don’t get it. I thought the whole point of the of these metrics was to allocate the actual wins for a team to the effort of the individual players. If you just allocate defensive effort equally across all players of course you will not alter the total wins predicted, what you will do is mis-allocate credit for the defensive effort.
You seem to defend the process by saying that it accurately predicts games won. This is certainly a requirement if the numbers are accurate but it is not enough. What has been done to verify that the wins are actually distributed accurately between the players on a team?
I am not sure why you decided to test the 82games.com data with the Wizards, I would like to see you do it with the Spurs since there is serious anomaly there. According to the wins produced metric Bruce Bowen was essentially a worthless player since in the regular season he produced negative wins (which I guess translates to losses). This is at such variance with what the coaches think that I think some investigation is in order. I could believe that if the Spurs were an average or below average team that the coach could be mis-judging his talent. But I do not see how coach Greg Popovich can win four titles and be so terribly wrong about evaluating one of what is, in his judgment, a key player. Not only is Bruce rated suspiciously low but Brett Barry is rated suspiciously high although I can swallow that one a little easier.
Pete
December 1, 2007
Thank you everyone for the informative replies. I especially like the idea of combining +/- with win score to better attribute defense at the individual level. (In this good post that Owen pointed out to me– https://dberri.wordpress.com/2007/03/26/incorporating-defense)