Tuesday, July 31, 2012

Introducing the Contest of the Week: What's Your Lineup?

What's the best lineup, order included, that can be made from the 2002 San Francisco Giants?  Think you can come up with a better one than other readers?

---------------------------------------------------

Today I'm introducing a new feature of the blog: the contest of the week.  Each week, I'll announce a contest to readers.  Each week, the winner of the contest gets three Shadow-points, second place gets two Shadow-points, and third place gets one Shadow-point.  Every two months, the person with the most Shadow-points in the period gets their name put on a Hall of Fame widget to the right of the blog (I'll create it once it's needed), a wallet-busting $2 prize, and the opportunity, if they want, to write a guest article for the blog.

So, on to the first contest:

What's Your Lineup?


I wrote Basim, a python script that simulates baseball games based on the stats of the people in the lineups.  The details of how it works are in the link above; you can also see Basim simulating games live on the right  of the blog.  I've recently been looking into evaluating players with it, but this contest has to do with the original use of Basim: evaluating batting orders.

So, the first contest is to construct the best batting order from the 2002 San Francisco Giants roster.  To enter the contest, submit a lineup from their players; I'll run Basim on all submissions, and the three highest average runs per game are the winners.

Rules:

1)  The players you can draw from are listed here.

2)  You can only use players who had at least 100 plate appearances with the Giants that year.

3) Your lineup must be defensively valid.  That is to say you must have a first baseman, second baseman, etc.  Shortstops and second basemen are considered interchangeable, and all outfielders and first basemen are also interchangeable.  Third basemen and catchers can play first base, but not vice versa.  The position that a person can play, up to interchangeability, is the one listed here. (Technical note: Dunston can play 2B, SS, OF, and 1B, in case you want to play him for some reason).

4) Your pitcher (in your lineup) must be Jason Schmidt.

5)  I will run 1,000,000 simulations on each submitted lineup to find its average runs scored per game.  The highest value will win and get three Shadow-points; second will get two, and third will get one.

So, for example, a submission might look like:

1.  Benito Santiago (C)
2.  J. T. Snow (LF)
3.  Jeff Kent (SS)
4.  Kenny Lofton (1B)
5.  Barry Bonds (CF)
6.  Jason Schmidt (P)
7.  David Bell (3B)
8.  Ramon Martinez (2B)
9.  Marvin Bernard (RF)


Some (quite obvious) things to think about: where do you put Bonds?  What do you think of stolen bases?  What do you want in a leadoff hitter?  Where do you put the pitcher?

Submissions are due by Monday, August 6th; I'll announce winners the next day.  To submit a lineup, either email it to me (sambf at mit dot edu), or post it as a comment on this post.  (Make sure to include your name.)

Good luck!


Sunday, July 29, 2012

Basim Live: Like baseball, but without the commercial breaks

Earlier I posted about Basim, a baseball simulator I designed that simulates baseball games based on the stats of people in the lineups.  My long term project has been to design a stat eWAR, empirical wins above replacement, that attempts to determine how good offensively a baseball player is by seeing how many runs lineups including them produce.

Recently, I put up a widget on the top right of the site that runs Basim simulations live every time you load the  page.  It randomly chooses two 2011 mlb teams from the same league and plays a game between them, the same way I simulate games to calculate eWAR.

I'll continue to work on it--eventually I'd like to set up a simulated baseball season, shown live on the blog--but for now I'm interested in feedback on how to improve it.

Known things that could be improved:

1) Groundouts and flyouts could be treated differently, with batters replacing runners on first on groundouts that don't turn into double plays.

2) The batting orders, roughly speaking, are the most frequently used lineups from the team from 2011, but this can have some weird consequences, like not using someone who is generally in the lineup because they switch spots frequently enough that no single lineup involving them has been used very much.  So, some of the orders that I'm using aren't very good.  To that extent, if you see a lineup that is not very representative of a team, tell me and I can change it.  (There are also potential oddities where the lineup lists I'm reading from only use last names, so, for example, for a while the Pirates were using Donald McCutchen in their lineup instead of Andrew McCutchen.)

3) Right now people are always thrown out trying to take an extra base on 3% of balls put in play; if someone could find better player-by-player data for this, that'd be awesome.

Anyway, feel free to shoot me any comments you have.  And in the mean time, I hope you spend as much time staring at text-based baseball games as I have.


Addendum: if you know a good way to synthesize pitcher stats with hitter stats to predict an at bat, I'm all ears; right now the approximation I'm thinking of using is to add up the deviations from average of the two statistics, though of course  in the end what I really want to do is do a regression....

Friday, July 27, 2012

The Playoffs and the Trade Deadline

Long story short: you should make aging all star for prospect trades at the start of the season, instead of at the trade deadline, if you're at least 70% sure, before the season starts, that the team getting the aging all star will be in playoff contention and that the other team will not.

_______________________


In Major League Baseball, teams are not allowed (or, in some cases, just significantly hindered) in making trades after the July 31st trade deadline.  This left me wondering: in what cases does it make sense to wait until the trade deadline to make trades, and when does it make sense to do it at the start of the year?

Put another way, it's been about ten years since the Seattle Mariners were in  playoff contention, and about twenty years since the Yankees weren't--the fact that this is also true this year shouldn't surprise anyone.  So why didn't the Mariners trade Ichiro to the Yankees at the beginning of the year?

Most trade deadline trades, like the one involving Ichiro Suzuki, involve a team in playoff contention trading some future prospects and/or money to a team not in contention in return for an (often aging) borderline all-star level player.  The advantage of making this trade at the beginning of the year, instead of waiting until halfway through for the trade deadline, is that the team in contention gets the good player for longer and thus the player  has more utility for them; the disadvantage is that you might accidentally trade away a good player and then find yourself in playoff contention, or vice versa.

The first thing I investigated was the following: how much does gaining an Ichiro-like player help in the regular season, and how much does it help in the post season?  Here were the assumptions I made.

1) All that matters is maximizing chances of winning the world series.
2) The player will add roughly 3 wins to the team (i.e. have a WAR of 3 more than the person they're replacing).
3) The team will be facing teams of equal strength not including the player traded for in the playoffs, and will win with a probability over .500 corresponding to the difference in their inherent winning percentage and their opponents'.
4) The team suspects that they will end up doing something between winning the division by 7 games, and losing by 10 games (roughly the average difference between division leader and second in division in 2011).


The probability of a team winning the world series p_WS = p_PL + p_PLW, the probability of making the playoffs + probability of winning the playoffs.

p_PL will be increased by ~3/20 by making the trade at the beginning of the season, meaning that p_PL will go from 50% to 65%, a proportional increase of .65/.5 = 1.3

To calculate the change in p_PLW, note that they have gained 3/162 ~ .0185 in winning percentage, meaning that they have a 51.85% chance of winning a playoff game (by assumption 3).  So, whereas before p_PLW = 12.5%, now p_PLW = 15.6%, a proportional increase of about 1.25.

I had not been expecting this result; I had assumed that the playoffs would be random enough that player quality wouldn't matter as much as it would  in the regular season (Because there are more games).  But it seams that, in fact, a good player added to a good team might make about as much of an impact in the playoffs as in the regular season.  (Note, however, that if you give yourself some partial victory credit for making the playoffs even if you don't win, that that would argue in the other direction.)

So, what's the upshot of this, as far as the trade deadline is concerned?  Without trading you have a 50%*12.5% = 6.25% chance of winning the world series.  If you trade at the beginning of the year, you have a .65*.156 = 10.14% chance of winning the world series.  And if you make a trade at the trade deadline you get about 1.5 extra regular season wins from the player, so (by the same logic) you have roughly a .575*.156 = 8.98% chance of winning the world series.

So, making a before-season trade increases your probability of winning the world series by about 3.89%, whereas making it right before the trade deadline increases it by about 2.74%.

This means that the advantage of making the trade pre-season is that it's about 3.89/2.74 = 1.42 times as effective.  Assuming that you'll be pretty sure by mid season whether you're in a playoff race, the question then becomes: before the season, are you at least 1/1.42 ~ 70% sure that your team will be in a competitive playoff race, and the other team won't be?  In another post, I'll look at this question.

Thursday, July 26, 2012

Utilitarianism, part 5: Who Counts?

This is the fifth post in a series about utilitarianism.  For an introduction, see the first post.  For a look at total vs. average utilitarianism, see here.  For a discussion of act vs. rule, hedonistic vs two level, and classical vs. negative utilitarianism, see here.  For a response to the utility monster and repugnant conclusion, see here.

_______________________


Another question that sometimes gets asked about utilitarianism is: who, exactly, is included in the calculation of world utility when evaluating possible scenarios?  People alive now?  People who would be alive then?  People who will live no matter what?  This is a question which is also often bundled into total vs. average utilitarianism, but I've already written on that here, and, I hope, demonstrated that average utilitarianism doesn't work.  I've also attempted to refute the "repugnant conclusion", the most commonly leveled criticism against total utilitarianism, with the "Sam is Great" rule here.  So I will assume that no matter what group of  people counts, the correct thing to do with them is to total up their utility in order to evaluate a scenario.

The first thing to note is that total utilitarianism naturally handles the somewhat unpleasant question  of whether dead people matter, and if so at exactly what point someone is "dead": dead people have zero utility and so won't affect the total utility of the world anyway.  You should include them in the calculation the whole time, but it will stop mattering once their dead enough to not feel any pain or pleasure.


Not Yet Living


The most common form of not counting some people in utilitarian calculations is generally called prior-existance utilitarianism, which states than when calculating the utility of possible universes you should only include people who are already alive, i.e. not people who will be born sometime between the current time and the time that you are analyzing.  There are a number of variants on this, but the central idea is the same: there are some people who shouldn't yet count for evaluating the utility of future scenarios.

This idea, however, has two damning flaws.  To understand the first flaw, look at the following scenarios.

The Future Pain Sink


There are currently ten people in the world W, each with utility 1 (where 0 is the utility of the dead).  You are one of the ten, and trying to decide whether to press a button which would have the following effect: a person would be born with -10 utility for the rest of their life, which would last for 100 years, and you would gain 1 happiness for a year.  (If you like, you can imagine a much more realistic version of "pressing the button" which results in short-term pleasure gain in return for an increase in unhappy population, but for the sake of reducing complications I will stick to the button model.)  Do you press the button?  Well, if you are a prior-existence utilitarianism, you would do it: you don't care about the not-yet-existing person's utility.  You would prefer the button-pressed universe to the non-button-pressed one, because the ten existing people would be net happier, and so you would press the button.  But then the new person comes into being, and you value its utility.  And now you prefer the universe where you hadn't pressed the button to the one where you had, even though you haven't yet gotten any of the positive effects of the button press.  You disagree with your past self, even though nothing unexpected has happened.  Everything went exactly as planned. In fact, your past self could have predicted that your future self would, rationally, disagree with your current self on whether to press the button.


The Once King


Now, say that you're the king of an empire, the only empire on earth.  You could manage your grain resources well, saving up enough for future years, encourage farming practices that preserve the soil, and reduce carbon emissions from anachronistic factories--in short, you could make some small short term sacrifices in order to let your kingdom flourish in fifty years.  But the year is 1000, and so no currently living people are going to be able to survive that long except yourself (being a king, you have access to the best doctors of the age).  And so, even though you're a utilitarian, you don't plan for the long term because it'll help people not yet born.  But then fifty years pass and your kingdom is falling into ruin--and all of your subjects are suffering.  And so you curse your past self for having been so insensitive to your current universe.


The problem with both of these scenarios, and in general with prior-existence utilitarianism, is that your utility function is essentially changing over time: its domain is the set of living people, a set which is constantly changing.  And so your utility function will disagree with its past and future selves; it will not be consistent over time.  This will give some really weird results, like someone repeatedly pressing a button, then un-pressing it, then pressing it again, etc., as your utility function goes back and forth between including a person and not including them.  Any morality had better be constant in time, or it's going to behave really weirdly.



The second flaw is much simpler: why don't you care about future generations' happiness?  Why did you think this would be a good philosophy in the first place?  Why would you be totally insensitive to some people's utilities because of when they're born?  Their happiness and their pain will be just as real as the current peoples', and to ignore it would be incredibly short-sighted, and a little bit bizarre, like a bad approximation to attempting to weigh your friends' utilities more than strangers'.


Friends and Family (and Self)


Speaking of which, the other common form of valuing people differently comes in valuing people close to you more (call it self-preferential utilitarianism).  So, for instance, you might weight your own happiness with an extra factor of 1,000, your immediate family's with a factor of 100, and your friends with a factor of 10, or something like that.

Let me first say that there are of course good practical reasons to generally worry more about your own utility, and that of close friends and family, than that of strangers: it's often a lot easier to influence your own utility than someone you've never heard of before; you can easily have significant control on your own life and the lives of those close to you; and maintaining close relationships (and living a reasonably well off life, at least by global standards) can help to prevent burnout.  But this is all already built into normal utilitarianism; to the extent that the utilitarian thing to do is to make people happy by talking to them and hanging out with them it's naturally going to be the case that this is best done with friends because you already have an established connection with them, you know that you'll get along with them, and it'd be difficult to find a random person and have an interesting conversation with them.  Once again, it's important to make sure not to double count intuitions by explicitly adding a term for something that is already implicitly taken care of.

Even if you are undeterred by this and want to weigh friends, family and self more than strangers, as a philosophy this is going to run into more problems.  First, to the extent that your friend base changes over time you could run into the same problems as prior-existence utilitarianism has with non-constant utility functions.  Second, it has the weird property that two people will disagree about what the right thing to do is even if they have the same information and see the same options for how the universe should be.  Third, as a consequence of this everyone being an optimal friend-preference utilitarian would, in many circumstances, be dominated by everyone being normal utilitarians.  The easiest example of that is the prisoner's dilemma.  Say there are two people in the world, A and B; each have a button they could press that would make them 1 util happier but cost the other 2 utils.  If both are self-preferential utilitarians they would both push the button, leaving both worse off than if they had both acted as normal utilitarians--even by self-preferential utilitarian metrics.  That is to say, everyone following self-preferential utilitarianism does not always lead to the optimal self-preferential outcome for everyone, and can in fact be dominated by other strategies that everyone could follow.  Now that's not a problem with a way to describe people's motivations behind their actions, but it seems a bit weird to endorse a philosophy that produces sub-optimal results by its own measure.  Fourth, there comes the sticky question of exactly what weights each person gets, and how that's decided.


All of these problems, in the end, come from the fact that self-preferential utilitarianism took utilitarianism and added an arbitrary, hard to define, non-universal, non-constant in time wrench into it.  This is the downfall of many flavors of utilitarianism; I've made similar points about average utilitarianism, negative utilitarianism, and high/low pleasure utilitarianism.  Like with average utilitarianism, in the end in some sense the problem with self-preferential utilitarianism is that it's not normal utilitarianism.


The astute observer will note that there is another divide in utilitarianism that would fit well under this title.  But it deserves quite a bit more space than this, and so it will have to wait until a later day.


Friday, July 20, 2012

The Utilitarian Boogeymen

This is my fourth in a series of posts on utilitarianism.  The first is an introduction.  The second is a post on average vs. total utilitarianism.  The third is a post on act vs. rule, classical vs. negative, and hedonistic vs. high/low pleasure utilitarianism.

____________________________________



This post is a response to various objections people often claim to utilitarianism.


The Repugnant Conclusion


The Repugnant Conclusion (or mere addition pradox) is a thought experiment designed by Derek Parfit, mean as a critique of total utilitarianism.  The thought experiment goes roughly as such:  Suppose that there are 1,000 people on Earth, each with happiness 2.  Well, a total utilitarian would prefer that there be 10,000 people each with happiness 1, or even better 100,000 people each with happiness 0.5, etc., leading eventually to lots and lots and lots of people each with almost no utility: in fact, for any arbitrarily small (but positive) utility U, there is a number of people N such that one would prefer N people at U utility to a given situation.  This, says Parfit, is repugnant.

I would argue, however, that this potential earth--with e.g 1,000,000,000 people, each with 0.015 utils of happiness, is far from a dystopia.  First of all, it is important to realize that this conclusion only holds as long as U--the happiness per person--remains positive; when U becomes negative, adding more people just decreases total utility.  So, when imagining this potential planet it's important not to think of trillions of people being tortured; instead it's trillions of people living marginally good lives--lives worth living.

Still many people have the intuition that 1,000,000,000,000, people at happiness 2 (Option A) is better than 1,000,000,000,000,000 people with  happiness 1 (Option B).  But I posit that this comes not from a flaw in total utilitarianism, but instead from a flaw in human intuition.  You remember those really big numbers with lots of zeros that I listed earlier in this paragraph?  What were they?  Many of you probably didn't even bother counting the zeros, instead just registering them as "a really big number" and "another really big number, which I guess kind of has to be bigger than the first really big number for Sam's point to make sense, so it probably is."  In fact, English doesn't even really have a good word for the second number ("quadrillion" sounds like the kind of think a ten year old would invent to impress a friend).  The point--a point that has been supported by research--is that humans don't fundamentally understand numbers above about four.  If I show you two dots you know there are two; you know there are exactly two, and that that's twice one.  If I show you thirteen dots, you have to count them.

And so when presented with Options A and B from above, people really read them as (A): some big number of people with happiness 2, and (B): another really big number of people with  happiness 1.  We don't really know how to handle the big numbers--a quadrillion is just another big number, kind of like ten thousand, or eighteen.  And so we mentally skip over them.  But 2, and 1: those we understand, and we understand that 2 is twice as big as one, and that if you're offered the choice between 2 and 1, 2 is better.  And so we're inclined to prefer Option A: because we fundamentally don't understand the fact that in option B, one thousand times as many people are given the chance to live.  Those are entire families, societies, countries, that only get the chance to exist if you pick option B; and by construction of the thought experiment, they want to exist, and will have meaningful existences, even if they're not as meaningful on a per-capita basis as in option A.

Fundamentally, even though the human mind is really bad at understanding it, 1,000,000,000,000,000 is a lot bigger than 1,000,000,000,000; in fact the difference dwarfs the difference between things we do understand, like the winning percentages of the Yankees versus that of the Royals, or the numbers 2 and 1.  And who are we to deny existence to those 900,000,000,000,000 people because we're too lazy to count the zeros?

I have one more quibble with Parfit's presentation of the thought experiment: the name.  Naming a thought experiment "The Repugnant Conclusion" is kind of like naming a bill "The Patriot Act" so that you can call anyone who votes against it unpatriotic.  I'm all in favor of being candid about your thoughts, but do so in analysis and discussion, not in naming, because a name is something that everyone is obliged to agree with you on.

By the way, I'm naming the above thought experiment the "if-you-disagree-with-Sam-Bankman-Fried-on-this-then-you-probably-have-a-crush-on-Lord-Voldemort conclusion", or "Sam Is Great" for short, and would appreciate it if it were referred to as such.


The Utility Monster



The Utility Monster is the other of the two famous anti-utilitarianism thought experiments.  There are a few different version of it running around.  All of the versions revolve around a hypothetical creature known as the Utility Monster.  In some versions it gains immense amounts of pleasure from torturing people--more pleasure than the pain they feel--and in others it simply gains more pleasure than others from consuming resources, e.g. food, energy, etc., and so the utilitarian solution would be to allow it all the resources, while the rest of humanity either withers off or continues a greatly diminished existence.

There's something that seems intuitively wrong about giving all societal resources to one utility monster, but what's going on here is really the same thing as the intuition behind negative utilitarianism:  because it's so much easier to make someone feel really bad (e.g. torture them) than really good, no normal person would come close to being able to gain from consuming the world's resources anything close to the losses associated with seven billion tortured people.  In fact if you took a random person and gave them unlimited resources, it's unlikely they'd be able to make up for a single tortured person (especially given the lack of basically any happiness gained from income above about $75,000).  In order to make up an additional factor of seven billion, the utility monster in question would have to be a creature that behaves fundamentally differently from any creature we've ever encountered.  In other words, in order for utilitarianism to disagree with out intuitions and endorse a utility monster, the situation would have to be way outside the set of situations we have ever encountered.

And, in fact, it would be a little bit weird if the optimal decisions didn't disagree with our moral intuitions in weird-ass situations, because our intuitions are not meant to deal with them.  This is a phenomenon frequently seen in physics: when you get to extreme situations outside of the size/speed ranges humans generally interact with, our intuitions are wrong.  It seems really weird that if you travel the speed of light you don't age relative to the rest of the universe, but that's because our intuitions were developed for 10 mile per hour situations, not for the speed of light.  It seems really weird that objects do not have well defined positions or velocities but are instead complex-valued probability distributions floating through space, but that's because our intuitions weren't developed to deal with electrons.  It seems really weird that snapping a piece of plutonium into smaller pieces can destroy a city, but it can.

Long story short, utility monsters are generally really bad bargains even for a utilitarian, and trying to work around it is just double-counting this effect (like negative utilitarianism double counts our intuition that it's harder to be really happy than really sad).  And when a utility monster really is the utilitarian option, you're in a really bizarre situation that we shouldn't expect out intuitions to work in any way.



I Don Care About Morality



One of the more common responses I get to utilitarianism is some combination of "well screw morality", "you can't define happiness", and "the universe has no preferred morality".  And all of these are, in a sense, true: in the end it's probably not possible to truly define happiness, and the universe does not have a preferred morality.  And all of these things are fine things to use to reject utilitarianism, as long as you wouldn't have any objections to any possible universes, including ones in which you and/or other people are being tortured en mass for no good reason.  But as soon as you say that it's "wrong" to steal or torture or murder, you've accepted morality; as soon as you say that it's bad to be racist or sexist or you have political positions you've accepted morality.  You can't have it both ways.



Please, sir, can I have some more?



Perhaps the most common response I get to utilitarianism, however, is a combination of two statements.  The first, roughly speaking, is "then why aren't you just high all the time?"; the second, roughly, is "then why aren't you in Africa helping out poor kids?"  Yes, these two statements are contradictory; and yes, I often hear them together, at the same time, from the same people.  I'm going to ignore the first statement because it's just wrong, and instead  focus on the second.

Imagine that you lived in a universe where the marginal utility of your 5th hour of time after work was greater volunteering at a soup kitchen than hanging out with friends.  (It's probably not too big a stretch of your imagination.)  Well, it's probably also true of the fourth hour after work.  And the third...  At some point you might start to loose sanity from working/volunteering to much and/or your productivity might significantly decline, but until that point it seems that utilitarianism says that if you've decided that some of your time--or money--can be better spent on others than on yourself, well, then, why not more of it?  Why not all of it?

I'm going to write much more on this later, but for now I have two points.  The first is that this, truly, is why people aren't utilitarians: in the end what scares people most about utilitarianism is that it encourages selflessness.  And the second point is that it would be really weird if a philosophy held that selfishness was good for the wold.  Yes, of course dedicating some of your life to making the world a better place is good, and of course donating more is better.  Defecting on prisoner's dilemmas is bad, and cooperating is good.

I don't, of course, believe that everyone does or ever really will act totally selflessly and totally in the interest of the world, but to argue that they shouldn't is sacrificing an almost tautologically true statement in an effort to reclaim the possibility that you're acting "well enough".

Next up: taking issue with decision theories.




Tuesday, July 17, 2012

Utilitarianism, part 3: Classical, Act, One Level

This is part three in a series of posts on utilitarianism.  For an introduction, see part one.  For a discussion of total vs. average utilitarianism, see part two.

___________________________________



In this post I'm going to address a number of divisions in utilitarianism that don't require quite so much space as total vs. average utilitarianism.



High/low pleasure vs. Hedonistic utilitarianism


High/low pleasure or two-level utilitarianism is the thought that some forms of pleasure are intrinsically better than others, independent of their function for society.  Hedonistic utilitarianism rejects this claim.

But why invent this distinction in the first place?  In general adding new rules to a philosophy should be looked upon critically: they often just serve as ways for people to save themselves from the consequences of their beliefs, and this is a perfect example of it.  I cannot help but note that in high/low pleasure utilitarianism, the "higher" form of pleasure--the more important one--is intellectual pleasure, and that it is a philosophy invented by intellectuals.  It would be like if I created "Starcraft utilitarianism", which is like normal utilitarianism but pleasure gained playing the Starcraft II computer game is weighted more heavily--a relatively transparent attempt to justify my current activities as worthwhile.  This is not to say, of course, that intellectuals are bad for the world--much of society's advancement is due to them--but they should have to justify their work and lifestyle on their own merits, not by inventing philosophies that arbitrarily favor what they do with their life.

This is all, of course, putting aside the issue of defining what exactly an intellectual pursuit is.

I am a hedonistic utilitarian.


Act vs. Rule utilitarianism


Act utilitarianism is what you usually think of as utilitarianism: the philosophy that you should try to maximize utility.  Rule utilitarianism, roughly, states that when evaluating a rule--for instance, a possible law, or maybe a standard like "bike to work instead of driving" (depending on the particular interpretation of rule utilitarianism), you should evaluate the rule on utilitarian grounds..  It suffers from two quite obvious flaws.  First, what exactly is a rule?  If it's defined as a possible algorithm to follow, then it just reduces to act utilitarianism (e.g. take the set of "rules" designed to produce actions corresponding to the set of possible universes).  Otherwise it's going to be impossible to define rigorously, or even semi-rigorously.  Second, why would you think that utilitarianism would be good for evaluating laws, but not everyday decisions?

This is not to say, of course, that whenever you have to make a decision you should get out a pen and paper and start adding up the utility of everyone in the world.  In many everyday situations the act utilitarian thing to do is to create a (not fully well defined) rule and follow it in unimportant situations to save your brain power for more important things.  When I try to decide how to get to campus I don't spend time calculating the marginal effect that my driving would have on traffic for other drivers; I assume it'd probably make it worse, and that that'd probably decrease their utility.  I also don't calculate the effect on global warming, etc: instead I understand those are all things that incentivize biking over driving, and so as a general matter just bike to campus and spend my time and brain power thinking about much more pressing issues that the world needs to confront, like Tim Lincecum's ERA.  Similarly I would generally be in favor of laws that incentivize biking over driving so as to discourage defections on prisoner's dilemmas, so long as the laws were well designed.  But this is not an argument for rule utilitarianism as a philosophy, just that sometimes it would argue for similar things as act utilitarianism.

For all of  those reasons, I am an act utilitarian.

Classical utilitarianism vs. Negative utilitarianism


The terminology is a bit confused here, so to clarify by this I mean: what is the relative weighting of pain and pleasure?  Classical utilitarianism does not tend to make much of a distinction between the two (or, if you wish, weights them equally), whereas negative utilitarianism weights pain more heavily: sometimes just by an extra factor (i.e. h(p) = Happiness(p) - k*Pain(p) for some k > 1), and sometimes infinitely more, i.e. reducing suffering is always more important than increasing happiness.  Still others try to split the difference between these two types of negative utilitarianism and weight all pain above some threshold infinitely, while pain below it equally the pleasure.

The strictest form of negative utilitarianism is clearly silly; it implies that if anyone anywhere is suffering, than no pleasure for anyone anywhere matters; the threshold version also has this problem to the extent that the threshold is crossed.  In addition neither of these are likely to be linear.

So, I'll just look at the weakest version: that h(p), the happiness of a person, is Happiness(p) - k*Pain(p) for some constant k > 1.  There are a number of problems with this:

1) Are happiness and pain calculated separately for each person, or first added together?  I.e.: say p, at a given point, is feeling 10 utils of happiness and 20 utlils of pain, and say k=2.  Should you be calculating h(p = 10-20*2=-30, or h(p) = h(10-20) = k*(-10)=-20?  The problem with the first version is that it is not going to be linear in how you view an individual's emotions: if you consider all emotions about an upcoming interview as one set of emotions, and everything else as another set, you'll get a different answer than if you consider everything about their hunger as one set  of emotions and everything else as another.  So, let's say that you're going with the second version: then what you're really saying is that you care more about net unhappy people than net happy ones.

2) k is totally arbitrary; why is pain twice as bad as pleasure is good?  Why not 10 times?  Why not 1.001 times?  For that matter, why not 0.7 times?

3) Why do you care about pain more than pleasure in the first place?  The reason, I think, is that you're grabbing on to the following fact about people: it's a lot easier to be really really unhappy than it is to be really really happy; there is just about no experience in the world as good as being tortured is bad.  But that fact will already be reflected in classical utilitarianism; naturally someone who is being tortured will have much more negative utility than someone who won the lottery will have positive utility because that's how humans work, and so the correct actions will care about them more; introducing k is just double counting for that.

For all of these reasons, I am a classical utilitarian.

Monday, July 16, 2012

Some calculations about Tim Lincecum

Tim Lincecum is a starting pitcher for the San Francisco Giants.  For the first five years  of his career he was one of the best pitchers in the majors, with a cumulative ERA of 2.98.  This year, however, he has been atrocious, with an ERA currently at 5.93.  I decided to take a look at pitch-by-pitch data to see if I could make anything of it.

I noticed that he had an unusually large difference between home and away ERA--3.43 at home, but 9.00 on the  road.  Coupled with the source of home field advantage, I decided to investigate something: could his difference in play this year come from umpires restricting his strike zone?

As it turns out, no.  In both 2011 and 2012 about 10% of his pitches that batters didn't swing at were balls miss-classified as strikes by the umpires, and about 2.5% were strikes that were called balls; umpires were no harsher this year than last.

I then took a look at placement.  In particular, all else equal better pitches are generally around the edge of the strikezone, and worse pitches are generally either right down the middle or way outside the strikezone.  So, I decided to look at the average distance from his pitches to the vertical and horizontal edges of the strikezone; this time there was a difference.  In 2011 the sum of the vertical and horizontal misses from the edges of the strikezone averaged .923 feet; in 2012 it averaged .961 feet.  It doesn't seem like a huge difference, but it is statistically significant, with just a 2% chance of occurring randomly (having a t-test value of -3.23).  So, it does seem like his control is down from last year.

I also looked at the velocity of his pitches.  There's been a lot of talk about his velocity decline; the decline, as it turns out, is real but not that precipitous: his fastballs and sliders have slowed down by about a mile per hour on average, though his changeup and curveball are still at roughly the same speeds there were in 2011.

So Lincecum's pitches are slower and less controlled than last year; in the end it just looks like Lincecum is pitching worse this season.

Is there anything else I should look at?

Contributors