Install Theme

nostalgebraist.tumblr.com →

slatestarscratchpad:

“One of them is the same old issue about the validity of correlation-based IQ research.  In your example, we can do controlled experiments.  With people, we often can’t, and are relegated to doing statistics techniques on uncontrolled demographic data.

There are better and worse ways to do this.  That Glymour article I linked to earlier today discusses this issue, the upshot being that the ways Murray and Herrnstein did it in The Bell Curve was very wrong, and an object lesson in what is wrong with statistical practice in social science today.  Of course, Glymour is speaking against the psychometric mainstream here, and so he is certainly outnumbered by people with impressive credentials, but his argument makes sense to me.”



Darnit, we stopped agreeing about everything substantial!

[snip]

It seems like we’re now getting to the point where we’re really disagreeing on some pretty fundamental issues – not even about IQ, but about science and how to live one’s life.

If I could phrase the central gist of your reply in my own words, it’d be something like

"What we have are imperfect but interesting bits of scientific data, and if we choose to ignore these because they’re imperfect, we’re effectively deciding to go with our existing prejudices, or with the status quo.  This may have socially regressive consequences, as [you say] it does in college admissions.”

One attitude I could possibly have toward this idea is: “well, yeah, going with the status quo is bad and should be avoided if we have information that is any good at all, but our IQ information is actually not even ‘any good.’  [insert thousands of words poking holes in IQ research in an attempt to convince you of this]”

That isn’t what I really think, though!  My actual attitude is “the status quo isn’t so bad!  Existing, non-scientific ideas have their own virtues, such as being stable over time in a way that makes life easier to live, being based in some and perhaps many cases on a pretty good analysis of the data, and often being bounded in how much harm they can cause (they haven’t yet gotten us into any mess worse than the one we’re in, after all).”

(In some senses of the term, I guess I’m a conservative.)

For clarity, I’ll first describe this attitude in the now-familiar world of pseudo-speedometers.  As I stipulated way back when, the pseudo-speedometers only look at rate of footfalls.  This works pretty well for most humans, and works abysmally for creatures with stride lengths very different from the average human: for instance, they say that horses go way more slowly than they clearly do.

Now, one possible response to this is “wow, look at this new, exciting, counter-intuitive result from the new science of pseudo-speedometry!  Sure, it upends what seems obvious about horses, but even imperfect scientific information is better than mere unscientific intuition!  I’m going to sell my horse and go around talking about how we as a society use horses more than The Latest Research tells us we should.”

Another possible response is “but just look at a horse.  It moves really fast!  I’m going to keep riding my horse; my intuition tells me it’ll get me places fast, The Latest Research be damned.”

Then, a year later, someone makes a new version of the pseudo-speedometer with some ad hoc correction for horses, and notes that this new version can be used to make better predictions, such as the correct prediction that a horse really will get you places faster.  New papers are published and reported on with great excitement.

So now all of the advocates of The Latest Research read this stuff, change their tune entirely, start riding horses again.  Meanwhile, the other character in this story was riding their horse the whole time.  They had a stabler, simpler life – they weren’t being jerked around by the mercurial whims of The Latest Research.  This would have been a nice perk of their strategy even if they had been wrong, but as it turned out they weren’t.

Note that status quo-using character here could concede that pseudo-speedometry has some uses, even if they never let it override their intuitions.  In this character’s view, pseudo-speedometry is just a standardized (but imperfect) version of the intuitive practice of having a person look at something and asking them if it’s going fast.  In being standardized and numerical, pseudo-speedometry is nice; it might be useful for comparing different horses, for instance.  (One would of course make sure it’s doing what one wants by watching the horses oneself and comparing.)  On the other hand, if it tells you a horse goes no faster than a human, you don’t have to believe it.

The situation described above strikes me as pretty similar to the situation in various areas of human science – say, the field of nutrition science.  It seems to me that The Latest Research is constantly going back and forth on which sorts of food are horrible for you and which aren’t, and that even at any given time, there’s less a consensus than a heated controversy between several different factions.

One could let oneself be yanked around by all this, reasoning that some suitable average of The Latest Research at any given time is simply the best information we have, imperfect as it is.  But the alternative, the status quo, is not informationless!  It consists of things like “eating what you like best” and “going with your body’s signals” and “eating what’s worked for your ancestors (whose genes you share).”  These signals are always here, stably and reliably emitting their nontrivial share of information, while The Latest Research jerks back and forth on the basis of esoteric shifts in statistical methods and the like.

Is it worth your while to spend hours brushing up on statistics and methodology and reading abstruse academic disputes so you can determine whether the latest Science Diet is a good thing or not?  The alternative isn’t some reasonless, random, arbitrarily risky void; the alternative is “just eating a normal diet” (whatever that happens to mean to you), which is not so bad.

Now that brings us to another point, which is that it matters who the “we” is in questions like “how much should we care about The Latest Research?”  You give the example of college admissions.  If I were someone who actually worked in (a relevant aspect of) college admissions, I would consider it my responsibility to read up on all the relevant science and try to come to some conclusion about it.

This is the equivalent of having some sort of transportation policy statistics job in the pseudo-speedometry world.  You wouldn’t want to just say “horses aren’t slower than people, that’s absurd!” because any old fool can say that; that isn’t what they pay you for.  You would want to dive into the arguments for and against the “horses are actually slow” contention and come out with an informed view.  If the counter-intuitive result is actually right, you’d want to know, because a lot is riding on it.

But if you’re just a regular person, you don't have to do this.  That’s a lot of hard work that intuition suggests will give sparse returns.  (Cf. gwern’s comments on whether it’s really worth it for a non-expert to read up on the IQ and race debate.)  Sure, you might be required to make political decisions relevant to transportation policy, such as voting.  But given an appropriate government structure, you can just elect someone who strikes you as smart (the person in the previous paragraph) to do the thinking for you.

In the particular case of college admissions, for instance, my impression is that the predictive value of the SAT is a hot topic of debate, not a settled issue.  (This is relative to high school grades, mostly – not interviews and extracurriculars.)  My impression of the topic at this point is formed of stuff like “I read a New Yorker article ages ago which said the SAT was less predictive than high school grades but the SAT II was more predictive among those who take it” and “Scott seems to think the SAT is very predictive” and “my friend Isaac who works in SAT prep says the SAT isn’t predictive at all” and “when I do a Google Scholar search for this I get a bunch of conflicting opinions, some of which are produced by the College Board itself, which is another issue I’d have to disentangle.”

I’m sure glad it’s not my job to work this stuff out!  As it is, extracting anything reliable from the debate seems difficult enough that I’m pretty happy going with my intuitions, as a layman.

There’s a related point here, which is that I think exposing yourself to many low-quality arguments can be dangerous in itself, because it can transform your relatively benign initial position – agnosticism or status quo – into something else simply by barraging your mind with the sense that you should be nudging your view a little bit towards something, until eventually (because the human mind doesn’t have very many gradations of belief) you hit the wall and become a believer.  I talk about this here.  This is one of my longstanding issues with Less Wrong rationalism in practice – it seems to value updating in the direction of a very noisy signal rather than staying with your pre-existing biases, and this seems to lead people astray.  If the smallest update I can make is still a pretty big jump in belief, and the Latest Research sucks, I’d rather not update at all in the direction of the Latest Research.  Sorry for being irrational!

I think I’ve made the points I wanted to make (probably in more space than they needed).  Now I can imagine that you may be incredulous about all this.  You may be thinking you’ve forced me – by pressing me to justify my anti-IQ stance – into a conservative, traditionalist position that I would never espouse in general.  I mean, doesn’t this same view endorse people like racists who look at the famous evidence about race not correlating with genetic categories and say, “well, whatever, I’ll still go with my gut”?

Yes, it does, and I’m OK with that.  Because I don’t think The Latest Research is a good post on which to pin your anti-racist opinions (or much else).  This is an argument I’ve heard a lot (I think I first saw it in Pinker’s The Blank Slate): if you use, say, the Latest Research on “gay genes” to make a “scientific case” against homophobia, what if that research gets overturned?  Were the homophobes then right all along?  Surely not; presumably “homosexuality is genetic” wasn’t your real reason for opposing homophobia anyway.  (And it shouldn’t be.)  More generally, I think many of my opinions are based on a wealth of experience and non-scientific knowledge stronger than the Latest Research.  My advice for the hypothetical racist here is not “don’t go with your gut,” but “get a better gut.”  Make some friends who aren’t white and see how your intuition evolves in response; this is a much more reliable and time/energy-efficient source of interesting and relevant information than delving into the extremely noisy and confusing signal that is “the science of race.”  I’m not making a blanket endorsement of tradition, but tradition is only one source of intuition, and I am saying that intuition is better than The Latest Research, a lot of the time.

(via slatestarscratchpad)

jonomancer:

binghsien:

Ah, here’s the thing.

You don’t have to be smart to be right. You don’t have to be smart to be good.

Being smart doesn’t make you right. Being smart doesn’t make you good.

Intelligence is like beauty: often admirable, but basically an arbitrary social construction that we nonetheless equate with being a good person.

If you read something, if you hear something, why should you care if the person who said it is smart? Care if the thing that they said was right, or was good. Or, hopefully, both.

Agreed!

Sometimes, all that “smart” means is that you come up with more complicated rationalizations for your wrong, harmful ideas.

Silicon Valley is rife with what I call “dumb ideas for smart people”. Like Bitcoin, and Singularity, and (flashback to 1999) the idea that the internet had so changed the rules of economics that your company didn’t need to make an actual product anymore.

The east coast has its own set of dumb-ideas-for-smart-people, including the idea that we’re always to the right of the hump on the Laffer curve, and the theory that we could sell subprime mortgage derivatives forever because housing prices would never stop going up.

These are all pretty counter-intuitive ideas. You have to be a certain level of smartness before you can even understand the argument for the idea. So you only see people above a certain smartness level believing the idea. Joining this group makes you feel smart, because your contrarianism separates you from the not-smart people.

But none of this proves the idea is correct, it just proves it’s complicated. I see a lot of otherwise smart people suddenly forget to be skeptical when an idea is sufficiently complicated.

I was a kid who skipped a lot of grades. I was constantly praised for my smartness by grown-ups. Because of this, I learned to value smartness above all else, and I also learned that I should just be smart all the time without having to put in any effort at things that aren’t naturally easy for me to understand. These are both TERRIBLE lessons, and it took me decades to unlearn them.

My IQ was measured at 160, but that doesn’t change the fact that I’m a dude who regularly, regularly spends 20 minutes looking for his keys only to discover that he left them hanging in the outside of his front door the night before.

“Stupid is as as stupid does.” This is why I don’t believe IQ tests measure anything real, and also why I no longer aspire to be smart: I aspire to be wise.

(via accordion-druid)

Okay, just edging a little bit into the territory of “things that make me angry and people will probably find obnoxious”:

It seems like I see a lot of people with this almost fetishistic distaste for scientists (or people in science-related jobs) who talk widely about things outside of science they know nothing about.

But this is a human trait that goes far beyond science.  For every Richard Dawkins, Steven Pinker or Eliezer Yudkowksy, there are tons of journalists, humanities professors, think tank employees, non-scientific bloggers, etc. who have the exact same attitude of “I think everyone needs my opinion about everything at all times, even if I know nothing about it, because I have the magical Smartness abilities you lack.”

These people are not as clumsy when they do this, so it’s less easy to make fun of them.  They speak fluently and appealingly and do not use argumentative tactics that seem strange outside of a science department.  But this just makes them more insidious.

In my experience, most scientists simply view their work as a kind of specialized technical knowledge, not a mandate to opine on everything under the sun and get in everyone’s face, in all possible faces.  It’s a strange leap to convince yourself that being good at science lets you do that, and so only a certain sort of person makes that leap.  By contrast, college education in many non-scientific fields at this point practically tells you you have that mandate.  Your degree doesn’t give you technical knowledge, it “teaches you to think,” it lets you “theorize” about anything and everything.  Go out into the world and explain things, explain people you don’t know or understand, to themselves – after all, you know how to think, right?

People who talk this way are everywhere and they are insidious because they are not laughable, unlike scientists who talk out of their field and out of their rhetorical depth.  The people who know how to make you nod sagely as though they’re conveying Insights, because they know “how to think” better than you do, are the ones that really worry me.

(I’m not saying non-science degrees are bad, or will not teach you things, or will turn you into an asshole if you weren’t already one; I’m saying that from what I experienced in college it seemed like they were very good at enabling assholes who wanted mandates to opine.)

Um, sorry? <3

It’s not your fault!  I mean, I’m sure you know that, but I’m saying it anyway

What is frustrating is more this feeling that my social capital is mostly built on saying things that no one is likely to disagree with?

Like, making fun of Eliezer Yudkowsky is a great gimmick because as far as I can tell no one feels threatened by him, his “enemies” just find him amusing and entertainingly ridiculous at worst, and his fans are happy to take my mocking in good humor and then have good-natured arguments with me on the basis of it (or, at worst, savior my Big Yud tag and then be friends with me)

But when I see someone doing the same things that squick me out about EY, but it’s not as much of a comfortable position to take, I don’t really think I can do it

I guess I feel like a coward who mocks the people it is socially acceptable to mock in my social context and doesn’t ever take socially unacceptable stands

blurds replied to your post “For the last few years I’ve felt more socially competent than I did in…”

There is nothing in this process besides learning and internalizing new rules. Put me in a situation I don’t have a workaround for and I’m still a silent mess

Maybe, but that doesn’t leave me understanding why there is so much animosity for people who aren’t as good at following the rules

E.g. a lot of my relative “lack of arrogance” is knowing now that I have to surround a lot of my statements with a bunch of disclaimers, and I don’t actually like doing this (if these disclaimers get used this often they should be considered always implicit as part of some Principle of Charity + Gricean Maxims thing?), but people who don’t perform the ritual get punished, and I’m left feeling like I’m dodging bullets again and again, like I’ve infiltrated something

People seem to put a lot of moral weight on rules that I’m only following because people seem to put so much moral weight on them.  (Maybe everyone is doing this simultaneously and it’s just a bad equilibrium)

Remember the post I made a few days ago about how it’s annoying that you can’t talk about certain things sometimes without bringing up the specter of “feminism” and derailing the conversation?  In some conversations I’ve learned to make my points in a careful way so they just sound like “common sense” instead of “feminism”

But here’s the thing – I have also learned the exact same thing with talking about heuristics and biases and the kind of stuff Less Wrong talks about when they’re being sensible; I’ve learned that I have to talk about this kind of stuff carefully and talk around the dread specter of “rationality”

Neither of these makes any internal sense to me, I have just learned to deal with not bringing up specters that derail the conversation.  There are various such specters and sometimes I understand why they exist for a certain person but often it’s just baffling.  Like I have this sense of “what will not derail this conversation” that is becoming better and better over time, but that makes me feel manipulative, because it involves steering around the signals (“feminism,” “rational,” etc.) that people would normally be using as red flags, and meanwhile I’m here steering around these whirlpools while not even understanding why they’re there

Which I guess might just be what “social skills” are?  But that creeps me out, then

hpgross:

nostalgebraist:

hpgross:

nostalgebraist:

I’m really tired and I just remembered that weird anecdote about the guy who claimed his goal in life was to experience an amount of pleasure describable by a Dirac Delta function (an infinite amount of pleasure for an infinitely short time, adding up to a “normal” amount of pleasure), because that would mean he had “won” (??????)

what the fuck

But that doesn’t even make any sense. Dirac delta functions only work as limits of functionals right? That would require there to be no upper bound on pleasure. Though the fundamental notion that there is some metric of pleasure that would work like this is also crazy.

Like, there’s so many things wrong I don’t know where to begin.

I don’t know either, man.  (The anecdote is from this post, and doesn’t give any more information than I gave — either it’s being badly reported, or it makes no sense, or there were a lot of justifying arguments that made it make sense that are lost to time, unless someone else runs into this guy at some point.)

Fun Theory, then, is the field of knowledge that would deal in questions like:

  • “How much fun is there in the universe?”
  • “Will we ever run out of fun?”
  • “Are we having fun yet?”
  • “Could we be having more fun?”

This sounds like Douglas Adams satire more than anything. How is this not just dressed up utilitarianism?

From what I remember, it’s mostly about how to imagine a utopia that doesn’t sound like it would get boring after a short time, which is surprisingly hard.  (A lot of the utopias people have proposed in the past have that problem.)

(via matchazed)

hpgross:

nostalgebraist:

I’m really tired and I just remembered that weird anecdote about the guy who claimed his goal in life was to experience an amount of pleasure describable by a Dirac Delta function (an infinite amount of pleasure for an infinitely short time, adding up to a “normal” amount of pleasure), because that would mean he had “won” (??????)

what the fuck

But that doesn’t even make any sense. Dirac delta functions only work as limits of functionals right? That would require there to be no upper bound on pleasure. Though the fundamental notion that there is some metric of pleasure that would work like this is also crazy.

Like, there’s so many things wrong I don’t know where to begin.

I don’t know either, man.  (The anecdote is from this post, and doesn’t give any more information than I gave – either it’s being badly reported, or it makes no sense, or there were a lot of justifying arguments that made it make sense that are lost to time, unless someone else runs into this guy at some point.)

(via matchazed)

Quick response to hot-gay-rationalist’s most recent post – I have to leave to get on a plane soon so I don’t have as much time as I’d like to think and write about this, but I want to note down my initial, possibly stupid responses to your points so any easily resolvable confusions can be resolved.

First, about the principle of indifference: I agree about the relabeling and the idea that the 4 outcomes shouldn’t be treated differently.  However, this is (sort of) why I don’t think it’s appropriate to represent my state of uncertainty as a probability distribution.  Assigning probability 0.25 to each outcome implies that the behavior of the lights is independent, but that’s something I feel like I, in a definite/positive sense, do not know.

I would not be “more surprised” by non-independent behavior than independent behavior, it’s just that all of the possible non-independent behaviors “cancel out,” as it were (because of the invariances), so that the probability distribution has independent behavior.  But that doesn’t mean that independence is part of my knowledge.

Perhaps using the correct P_a distributions would reflect this fact?  What I want to capture is that I don’t believe any more strongly that the behavior of the lights is independent than that it isn’t.  I really do know nothing about the machine (let’s bracket the issue of whether total uncertainty is realistic for now – not enough time on my end).

Second, about MaxEnt.  I don’t think I’m doing the Mind Projection Fallacy.  I’m not looking for “what the real PDF is.”  I’m supposing that what I know about a phenomenon is limited to a mean and variance, and I’m trying to come up with a mathematical object that describes what I know about it.  For a Bayesian, this is “prior construction.”  Even if I’m not Bayesian, I might want some object that represents what I know.

What I’m saying is that using a Gaussian prior doesn’t feel like “what I should do based on what I know” in that situation.  If the variable is x and I care about E[(x-mean(x))^4], and want to make predictions about it, the fact that the variable might (in the real world as opposed to my head) be distributed like a Student’s t is a thing that would actually enter my head (like, in a real situation, if I were thinking about a thing and all I knew was its mean and variance).  That this fact seems to be “missing” from the Gaussian prior suggests that the Gaussian prior is somehow not a good representation of my actual understanding of the situation.

Note that my complaint is not that the Gaussian might not match the real world (which would be Mind Projection), but that my internal state of knowledge involves information that the Gaussian doesn’t seem to capture (all stuff in my own mind).  I don’t really have “the fourth moment of my uncertainty” because “my uncertainty” ranges over possibilities like “maybe the fourth moment isn’t even finite” and I can’t see how the Gaussian includes that.

hot-gay-rationalist:

nostalgebraist:

hot-gay-rationalist:

nostalgebraist:

hot-gay-rationalist:

nostalgebraist:

raginrayguns:

nostalgebraist:

“Where do Bayesians get their numbers from anyway,” installment (n+1)

[cut cut]

[snup]

[snop]

[snip]

[snrp]

[snep]

I think it’s because the uniform Ap distribution does not represent a state of “full uncertainty.” It in fact represents the statement “I think any probability is just as plausible as any other one,” and for most possible hypotheses, that is not a good description of most agents’ state of uncertainty about them! And once again we run into the problem of conflicting intuitions because I think that states of really total uncertainty are the ones that need to obey the rules of probability the most because they’re the easiest to fuck up.

And this problem is just isomorphic to the “finding priors” problem, which is the greatest weakness of the method. The Solomonoff solution, priors proportional to the negative exponential of the hypothesis’ complexity, is one that also appeals to me intuitively and mathematically - the basic argument that adding one bit makes twice as many hypotheses available and therefore should cut probability in half looks very good to me.

But there is no such thing as total uncertainty. Even in the ice cream case, you know something about the human population in the present and how it’s likely to change in the next 25 years, you know something about the prevalence of ice cream, etc - and even if you were completely ignorant, I hardly think you’d believe that “it’s as likely as not that that proposition is true,” or that literally any probability assignment is as justifiable as any other. The uniform Ap distribution is not special nor does it represent full uncertainty, because there is no such thing as full uncertainty, and because different amounts of information are represented by different Ap distributions (another example Jaynes uses is that of the existence of life in Mars at some point in the past, whose Ap distribution for himself he claims he’d describe as something like the Haldane prior).

—EDIT to add:

But also, as you and raginrayguns pointed out, it’s in practice impossible to really use the same heuristic rule about assigning more “reasonable” plausibilities to things when the thing is murky enough, as humans. But… all of my argumentation isn’t supposed to apply to humans :P I’m not concerned with what we can actually accomplish in the physical universe, I’m concerned with what’s the optimal way of reasoning, what’s the golden standard, hyper ideal, the Platonic Form of reasoning. Whether we can implement that and what we can do when we can’t is a completely different topic. But if I can determine that Bayes is indeed the ideal perfect unachievable golden standard of reasoning, then I can move on to see what approximations I can make and when and how I should deviate from the strict uncomputable solution.

OK, first of all a response to the last point: this conversation started with me objecting to some statements by a human, which are still floating up there above all the [snip]s and [snop]s.

As I’ve been interpreting it, this conversation has been about human approximations and not ideal reasoning, and to the extent that you’re only talking about ideal reasoning, you aren’t addressing the original question, which was “is the behavior of Robin Hanson (who is a human, not a JaynesBot) justified here?  Why or why not?”

Second: ultimately, I think the conflict here comes down to this question:

“Is it right to ‘add’ information we don’t feel like we know in order to make our representation of our uncertainty obey the probability axioms?”

It seems like our intuitions are diametrically opposed here.  I actively think this extra information shouldn’t be added, while you think that rationality demands that we add such information, and refer to not doing so as “fucking up.”

I’m going to give two examples of what I mean by “adding information.”  Note that I’m using “information” in an informal sense, not in the Shannon  sense or anything.  And also that, as always, I’m an amateur in these things and I wouldn’t ever want to imply that I’m the first to have ever thought of these objections – merely that I don’t know what the expert responses to them look like (though they presumably exist).

A thing I was getting at with my “conjunction” points yesterday is that if you know a probability distribution, you know information about the dependence structure of the events in it.  However, sometimes I don’t feel like I know the dependence structure of a set of “murky” events.  Representing my uncertainty as a distribution requires me to choose one, and this feels wrong.

For instance, let’s look at your example of the machine with two lights.  In that example there was no problem because the dependence structure was given (we know P(blue|red) = 0 and vice versa).

But now suppose I have the same machine, just as mysterious, except now I know any combination of the lights would be possible.  The possibilities are now {neither, red only, blue only, both}.  We could split this into various events like “red” which would be the set {red only, both}.  (I’ll give the event {both} the name “red&blue.”)

Knowing nothing about the machine, I don’t know what the dependence structure is.  Maybe the two lights are independent like two flipped coins, and P(red&blue) = P(red)*P(blue).  Or maybe they have some kind of dependence: maybe only “red only” and “blue only” are possible, or maybe only “neither” and “both” are possible.

What should my A_p distributions be here?  They can’t be uniform for each outcome because there are 4 outcomes and the means have to sum to 1, not 2.  (They still should have support over the whole interval [0,1] because maybe the machine just does blue every time or w/e.)  There’s probably a MaxEnt answer here?

In any case, whatever answer I choose, it will imply a dependence structure.  If I have a probability distribution over the space {neither, red only, blue only, both} then I can compute things like P(red|blue).

But actually I feel totally uncertain about those things, which I was not informed of at the outset.  They are “extra information,” and the idea that an object involving this information is the “right” representation of my state of uncertainty seems strange to me.

Is there any one dependence structure here that “correctly” represents my total lack of knowledge about the dependence structure of the machine’s behavior?  It feels counter-intuitive that there would be, though maybe there is.  Maybe I’m missing something here?

Here’s a second example of “extra information.”

Famously, if all you’re given are a mean and a variance, the MaxEnt prob. dist. on R is a Gaussian.

In going from “mean mu, variance sigma^2” to “N(mu,sigma^2)” I become able to compute many new things.  For instance, I can now compute any moment of this distribution.  I could tell you its fourth moment, say (and it would be finite).

However, the information provided is also consistent with other distributions, such as the Student’s t with nu > 2 (technically, a non-standardized Student’s t).  However, the Student’s t does not have defined nth moment for n >= nu.  So, the information provided is consistent with a Student’s t with nu = 3, but the fourth moment of that distribution is undefined (in the sense of being “infinite,” i.e. the integral diverges to +infinity).

So, suppose you come along and say, “Rob, there’s a distribution with mean mu and variance sigma^2.”  And I think, okay, there's some nonzero chance it’s a Student’s t.  After all, that’s a distribution that comes up in real life.

Now you ask “okay, Rob, what’s its fourth moment?”  And if I were a MaxEnt machine I’d happily spit out the fourth moment of N(mu,sigma^2).  But I, Rob, know that Student’s t is out there in the world, and that its fourth moment is infinite!  How do I incorporate this knowledge into a probability distribution?  If I have, say, some distribution over distributions in which the relevant Student’s t has a nonzero probability epsilon of being the right one, even if epsilon is tiny, the “expected value” of the fourth moment will still be infinite (infinity * epsilon = infinity), and I’ll spit out “infinity,” not “fourth moment of N(mu,sigma^2).”

So it seems like doing MaxEnt makes me forget that Student’s t is a possibility, in that it leads me to draw conclusions that seem unreasonable unless I think a Student’s t is actually impossible.  The information “mean mu, variance sigma^2” doesn’t just fail to tell me the fourth moment; it fails to tell me that it’s even finite.  The state of uncertainty I’m in, having received that information, seems very poorly represented by a Gaussian.

Again, I don’t doubt that people have thought about these issues and come up with answers to them; I just don’t know what the answers are.  And in sum, the disagreement here seems to come down to the question of whether states of uncertainty should or shouldn’t be “altered” to give me a prior that acts like a probability distribution.

(via hot-queer-rationalist-deactivat)

hot-gay-rationalist:

nostalgebraist:

hot-gay-rationalist:

nostalgebraist:

raginrayguns:

nostalgebraist:

“Where do Bayesians get their numbers from anyway,” installment (n+1)

[cut cut]

[snup]

[snop]

[snip]

tl;dr my subjective feelings about very inferentially distant propositions don’t feel like subjective plausibilities, and I get the sense that this is true for most people.  The way Bayesians quote numbers seems strange to many people, not just because it is unfamiliar, but because it seems to conflate the “sure uncertainty” one has about a fair coin with the “unsure uncertainty” one has about inferentially distant events.

About the difference in feeling, that’s indeed true, and in fact Jaynes has a whole chapter about this subjective difference and I wrote a post about it.

But I’m fairly certain that I wouldn’t say 50% to a question like that, because… well, I don’t know, full uncertainty doesn’t feel like “it could go either way” to me? And that’s part of the thing in the link, probabilities aren’t “brute numbers,” they have distributions. I’m also uncertain about my uncertainty, and if someone asked me that, I’d probably say something like “40% with a very very very wide tail.”

(By the way, I did read the whole thing, I just wanted to save space.)

It seems like you’re one of “nature’s plausibilists” — someone whose mind just naturally assigns a subjective plausibility to every proposition.  Which is pretty cool, don’t get me wrong, but my hunch is that this is not a common trait.  And the ultimate justifications for plausibilism are intuition-based, as ultimate justifications in philosophy tend to be.

Hmmm… yeah maybe. Though I don’t feel like those justifications are really intuition based, they’re more like “what I’d want to reason like”? I don’t know, maybe I’m projecting, or I’m just happy to have found my qualia represented in Bayesianism, but the Cox Axioms look like what I would want to reason like, even if I didn’t in fact reason like that - and of course I don’t really reason like that, I’m a biased human, and my reasoning deviates predictably from perfect Bayes, but it still looks like I’d want to be Bayesian and that whenever I reason in a way that’s inconsistent with that I’d feel bad about it. I do feel bad about it.

(Last point: I think there’s an even more fundamental state of uncertainty one can have, which is being uncertain about whether a proposition even describes a state of the territory at all.  For instance, if you asked me whether I thought Max Tegmark’s “Mathematical Universe Hypothesis” was true, I would feel a fundamental uncertainty caused in part by the fact that I’m not even sure yet what it would mean for it to be “true,” or whether that’s a meaningful question.  That is, uncertainty about whether or not a proposition is vacuous is a second kind of uncertainty that I don’t think can be captured well with plausibilities.  I have no idea if the Mathematical Universe Hypothesis is “plausible”; I don’t even know if it can be true or false, and will have to do more thinking to resolve that question.)

Well, here I think we sort of shrug and just go with logic? In logic, a meaningless sentence is always false (has no model, describes no possible world), so I think you can give that some form of probability. Maybe.

Though that’s in fact an Open Problem in FAI, in the same order of “how to reason about hypercomputers?” So this is a part where I confess complete epistemic confusion and say that this is an unsolved problem and that it may well be the point where Bayesianism unravels completely and we find out that we need some othermore universal form of reasoning to be our unattainable golden standard.

But I don’t know that Bayesianism can’t solve that, and I don’t know that simply giving potentially-meaningless propositions a probability will give me headaches. I haven’t yet read this paper but maybe there’s some potential stuff there.

I understand that such an A_p distribution can be constructed, and I guess then what I’d say is “for most futurological and other highly speculative propositions, my A_p distribution is nearly uniform.”

You can then take the first moment of all these uniform distributions and get 0.5 out and play games with the 0.5s, but this will at best just be a way of reflecting that I have no clue about the answers to any of these questions.  Saying “I have a subjective plausibility and it corresponds to probability 0.5” seems misleading; I don’t have a subjective plausibility.

Anyway, if we try to represent this state with uniform A_p distributions, doesn’t that run into the conjunction problem I mentioned?  "P and Q" should be (in the general case) less likely than P or Q alone, but supposing any of these are sufficiently far from my experience, I simply feel in a state of complete uncertainty about them.  So if you first asked me about “P and Q” I would give you “0.5” or “A_p is uniform” or whatever, but I would also have said that if you had presented me with P or Q alone.  (Taken literally, this set of judgments would seem to imply that all the events I’m totally uncertain about are really the same event, i.e. imply each other with probability 1, which is something I certainly don’t believe!)

For the above reason the A_p formalism seems like an awkward way to express the state of really total uncertainty I feel about many things; it still assumes my states of total uncertainty can play by the rules of probability when they can’t.  I don’t feel like they should, either – that would imply more knowledge than I really have, viz. some sense of which events are relatively big in the probability space and which are relatively small (to resolve the conjunction issue).

I guess what I’m looking for here is the mental state “more research is needed” – like the mental state I was in when I first learned about the “P=NP?” question but before I had learned that most people thought P != NP, and knew that if I really wanted to have a subjective sense of how plausible P=NP was, I should look up what experts thought about it.  I don’t think this can be captured by A_p because of the conjunction problem I mentioned, though maybe I’ve gotten that wrong.

(via hot-queer-rationalist-deactivat)