To be complete, we mention here
that possibility measures can be also interpreted as limits of probability measures in the large deviation convergence
sense (See [8], [11]).
Recall that the purpose of the study
of large deviations in probability theory
is to find the asymptotic rate of convergence of sequences of probability measures of rare events. For example, in the
simplest setting, while the law of large
numbers asserts that the sequence of
sample means (of i.i.d. Xj; j ≥ 1, with
EX = µ; V ar(X) = σ2 < 1) Xn converges surely to µ, it is also of interest to
find the rate at which P (jXn − µj > a)
(probability of large deviations from the
mean) goes to zero, as n ! 1, for some
a.
Now, for Zn = pn(Xn − µ)=σ, we
have, by the Central Limit Theorem,
By finding the asymptotics of this
integral for x ! 1, we arrive at the
well-known Cramer result of the form :
The sequence of sample means satisfies
the large deviation principle in the sense
that, for " > 0, we have for n sufficiently large, where the exponential rate of convergence being
e−nh(a).
Một vài quy trình này sẽ được đề cập trong
phần sau.
12 trang |
Chia sẻ: hachi492 | Ngày: 15/01/2022 | Lượt xem: 190 | Lượt tải: 0
Bạn đang xem nội dung tài liệu On the calculus of subjective probability in behavioral economics, để tải tài liệu về máy bạn click vào nút DOWNLOAD ở trên
49Asian Journal of Economics and Banking (2020), 4(1), 49–60
Asian Journal of Economics and Banking
ISSN 2615-9821
On the Calculus of Subjective Probability in Behavioral
Economics
Hung T. Nguyen
New Mexico State University (USA) & Chiang Mai University (Thailand)
Article Info
Received: 06/02/2020
Accepted: 16/3/2020
Available online: In Press
Keywords Bayesian Probability,
Belief Functions, Choquet Ca-
pacity, Coarse Data, Frequen-
tist Probability, Fuzzy Sets,
Granularity, Idempotent Analy-
sis, Linguistic Variables, Possi-
bility Distributions, Possibility
Measures, Random Sets, Sub-
jective Probability.
JEL classification
C10, C11
MSC2020 classification
60A05, 62C10
Abstract
In elaborationg upon the recent thought-provoking
paper“Subjective probability in behavioral economics
and finance: A radical reformulation” by H. Joel Jef-
frey and Anthony O. Putman [5], we proceed to spec-
ify the calculus of their “probability (uncertainty) ap-
praisals” as possiblity measures, i.e., the “radical” re-
formulation of the usual calculus of subjective proba-
bilities is that of idempotent uncertainty. With pos-
sibility measures as quantitative uncertainty for sub-
jective probabilities, we discuss the necessary mariage
of possibility measures and Kolmogorov probability
measures in a new Bayesian analysis for economic ap-
plications.
Corresponding author: hunguyen@nmsu.edu
50 Asian Journal of Economics and Banking (2020), 4(1), 49–60
1 INTRODUCTION
As stated in the abstract, this paper
is about the recent thought-provoking
paper [5]. It is thought-provoking be-
cause it proposed a “radically new” way
to really understand the old notion of
subjective probability, from which ap-
plications, e.g., in decision-making in
social sciences, will be “radically differ-
ent”, say, more “compatible” with what
psychological experiments revealed (in
behavioral econometrics and finance).
As such, it is important to expose
the essentials of [5] to a large audi-
ence of econometricians. In fact, by do-
ing so, we accomplish several important
tasks, namely specifying the calculus of
subjective probabilities as (idempotent)
possibility as previously suggested [2],
and discussing the necessary mariage of
possibility calculus with standard Kol-
mogorov probability calculus, which is
needed in a “Bayesian” framework of
statistical inference.
Since the “beginning”, we are taught
that, in constrast to frequentist prob-
ability (Kolmogorov), subjective (or
Bayesian) probabilities are in our mind,
which are used as a mode of judge-
ment. When using subjective proba-
bilities in applications, such as in auc-
tions (as Bayesian games) or Bayesian
statistics, not only we need to manip-
ulate quantitative subjective probabil-
ity, but also combine subjective prob-
ability with frequentist probability (to
form Bayesian statistics).
While, the mathematical (language)
foundation for quantitative (frequen-
tist) probability is Kolmogorov’s axioms
(by analogy with measure theory), in-
cluding additivity, what are the axioms,
i.e., the calculus of subjective probabil-
ity, that the Bayesians use to establish
their Bayesian statistical theory?
Can you guess? “Imagine” if the cal-
culus of subjective probabilities is dif-
ferent than Kolmogorov probability cal-
culus! Then how can Bayesian anal-
ysis be carried out? Specifically, how
to incorporate prior information into a
frequentist framework? Well, we know
that, as stated again in [5], the stan-
dard “view” is that with respect to cal-
culus (of uncertainties), frequentist and
subjective probabilities are two sides of
the same coin, i.e., while their meanings
are different, their calculi are the same.
This allows Bayesian statistics to exist
to “beat” frequentist statistics in several
fronts, e.g., in hypothesis testing.
I can’t resist to point out an op-
posite situation: while the meaning
of the (intrinsic) uncertainty in quan-
tum mechanics is the same as that of
ordinary randomness, their calculi are
different, namely, quantum probabil-
ity is non-commutative whereas Kol-
mogorov probability is not (but, in
fact, quantum probability is simply a
non-commutative generalization of Kol-
mogorov probability).
Should we ask “why it is so?”. More
specifically, “Why subjective probabil-
ities are additive, or even σ−additive,
just like Kolmogorov probabilities?”.
Well, texts on subjective probability ex-
plain that, by using betting schemes,
they are so by existence of Dutch book
arguments for these properties.
Note that, although Bertrand Rus-
sel once wrote in 1929 “Probability is
the most important concept in mod-
Hung T. Nguyen/On the Calculus of Subjective Probability in Behavioral Economics 51
ern science, especially as nobody has
the slighest notion what it means”, peo-
ple equate uncertainty with probabil-
ity whose calculus is Kolmogorov. Thus,
“traditionally”, while subjective and ob-
jective probabilities are different man-
ifestations of uncertainty, people ma-
nipulate them according to the same
calculus. After Dennis Lindley attended
a seminar at UC Berkeley in 1981
in which both Lotfi Zadeh and Glenn
Shaffer talked about their non-additive
uncertainty measures (possibility mea-
sures, and belief functions, respec-
tively), he wrote [7] claiming that all
non-additive uncertainty measures are
inadmissible, i.e., the message is “we
cannot avoid probability” (where, of
course, again, by“probability”, we mean
additive set functions). It turns out that
Lindley’s message is not exactly what he
claimed! What he did show is that “Us-
ing scoring rule approach, an admissible
uncertainty measure must be a function
of a probability measure”. But then,
for example, Shaffer’s belief functions
which are non-additive, are functions of
probability measures, and hence admis-
sible in Lindley’s sense! A complete re-
sponse to Lindley’s paper was [4].
One more thing about the coexis-
tence of objective uncertainty (say, in
von Neumann-Morgenstern utility the-
ory) and subjective uncertainty (say,
in Savage’s subjective/qualitative prob-
ability theory): this is possible in appli-
cations since these two different types of
uncertainty are forced to obey the same
calculus. For a rigorous treatment, see
[6], where it reminded the reader that
“One warning: when mathematicians
use the term probability , they almost
always mean a σ−additive probability
measure defined on a σ−algebra”! It
is clear in [6] that uncertainty theo-
ries are proposed to be used in mod-
eling behavior of individuals in their
decision-making, and as such, e.g., in
physics, models of choice must be con-
firmed by experimental evidence: they
were not (See the last Chapter of
[6] on “The Experimental Evidence”).
Thus, the door was open ever since for
non-additive uncertainty measures. It
should be emphasized for statisticians
that Kolmogorov probability is just one
quantitative modeling of one type of un-
certainty. There are other types of un-
certainty whose modelings might not
be “probability”, i.e., not additive. Do
not equate uncertainty with probabil-
ity. How to find out reasonable quan-
titative theories of uncertainty? Well,
just follow physics (I mean quantum
mechanics)! Quantitative modeling of a
type of uncertainty is used to model the
behavior of something, and as such, a
proposed model must be tested by ex-
periments. The non-commutativity and
non-additivity of probability in quan-
tum mechanics, as observed in experi-
ments, led to the establishment of a firm
theory of quantum probability. In so-
cial sciences, including economics and fi-
nance, people make decisions under un-
certainty. Thus, it is so clear that any
quantitative theory of uncertainty must
be validated by experimental evidence.
The thought-provoking paper under re-
view is precisely in this scientific spirit.
Somewhat clearly (!) that statisti-
cians and econometricians are still not
aware of non-additive set functions
which were proposed to model various
52 Asian Journal of Economics and Banking (2020), 4(1), 49–60
different types of uncertainty we face in
everyday decision-making, say in Arti-
ficial Intelligence, (not just one type of
uncertainty, traditionally attributed to
randomness or epistemic uncertainty, as
two sides of the same coin, i.e., obey-
ing the calculus of Kolmogorov proba-
bility theory), let alone fuzzy set the-
ory of Zadeh (1965, see [10] for a com-
plete update of the theory) from which
are founded concepts such as approxi-
mate reasoning, soft computing, gran-
ular computing, possibility theory (see
e.g., [1,3,13–16]). As we will see shortly,
the“radical reformulation”of the notion
of subjective probability (uncertainty)
in [5] is “probability appraisals” which
are nothing else than Zadeh’s “linguis-
tic variables” [17] in the context of fuzzy
set theory. In this context, it looks like
we are heading back into the territory
of fuzzy set theory which, since 1965,
only attracted computing and engineer-
ing fileds. Perhaps it will be so since, af-
ter all, there are “real” contributions of
fuzzy theory to social sciences, includ-
ing economics and finance.
The purpose of this present paper
is multifold: First, we elaborate of the
“radical reformulation” of [5] which we
believe that it was in a right direction
for improving methods in behavioral
economics and finance. Next, we place
this reformulation completely in the set-
ting of linguistic variables and granu-
lar information. Then, we complete [5]
by specifying a reasonable calculus of
subject probabilities, namely possibil-
ity measures (as opposed to probabil-
ity measures). Finally, we discuss a
“Bayesian” analysis in which possibil-
ity uncertainty coexists with probability
uncertainty.
2 UNCERTAINTY APPRAISALS
In a sense, as opposed to objective
uncertainty (quantitatively modeled as
frequentist probabilities), by subjective
uncertainty we mean the type of un-
certainties (e.g., epistemic uncertainty)
which cannot be modeled quantitatively
by a frequentist approach (e.g., for
events which cannot be repeated), it is
said that, for such uncertainty, we judge
it by using our mind. But how exactly
our mind perceives it, let alone manipu-
lates it (i.e., what is the calculus of sub-
jective probabilities?). Well, you might
say : it’s an old story and it has been
resolved long time ago! Note that, if
we adapt the current Bayesian calcu-
lus of subjective probabilities, then fre-
quentist and Bayesian statistics are just
two sides of the same coin (in the sense
that they use the same calculus of prob-
abilities).
While it has been known, also for
a long time, from experimental evi-
dence, that humans do not necessarily
manipulate their subjective “probabili-
ties” (uncertainties) according to Kol-
mogorov calculus, the paper [5] seems
to spell it out specifically as “uncer-
tainty appraisals” which could lead to
a “radical” calculus of subjective uncer-
tainty. We elaborate next the main mes-
sage in [5], namely“Subjective probabil-
ities are uncertainty appraisals” in so-
cial decision-making context, with em-
phasis on Bayesian statistics framework
in econometrics.
When facing, say, an epistemic un-
certainty (e.g., on an unknown parame-
Hung T. Nguyen/On the Calculus of Subjective Probability in Behavioral Economics 53
ter of a population), we take a “closer
look” at it, then use any information
we have about it to “appraise” it, i.e.,
how to describe the uncertainty a lit-
tle more precise for, say, actions in a
decision-making. Of course, as in gen-
eral thinking processes, humans tend to
use natural languages before numerical
languages. As such, an appraisal of an
uncertain situation should be a linguis-
tic variable in Zadeh’s sense [17], i.e., a
variable whose possible values are words
in a natural language. If the variable X
is an appraisal of uncertainties, then its
possible set of values could be “likely,
very likely, unlikely, ....” which can be
modeled as fuzzy subsets of the unit in-
terval [0, 1], i.e., fuzzy probabilities. In
this quantitative modeling process, the
subjective aspect of the appraisals is
reflected in the shapes of membership
functions of fuzzy probabilities.
When trying to model, say, the epis-
temic uncertainty concerning an un-
known parameter θ of a population (as
in standard practice of Bayesian statis-
tics), we view θ as a linguistic vari-
able instead. The parameter space Θ
(containing the true parameter) can be
coarsen into a fuzzy partition to provide
granular information about θ, i.e., a set
of possible linguistic values for θ, from
which we could consider granular infor-
mation of the form “θ is A is λ”, where
A is a fuzzy subset of Θ, and λ is a fuzzy
probability (e.g., “θ is “small” is likely”).
In order to “figure out” how to ma-
nipulate (in other words, how to derive
a calculus of) uncertainty appraisals,
it is necessary to dig into the model-
ing of membership functions of fuzzy
“probabilities”. Roughly speaking, say-
ing that something is, e.g., likely, is say-
ing that it is “possible” it is so. We
used to hear statements such as “some-
thing is improbable but possible”, re-
vealing that possibility is a weaker no-
tion than probability. Moreover, it
seems that a quantitative theory of pos-
sibility was first systematically formu-
lated by Zadeh [14]. See an update
in [1]. In the next section, we will pro-
ceed to advocate that the qualitative
approach to uncertainty appraisal in [5]
could be quantitatively formulated by
Zadeh’s possibility theory.
3 POSSIBILITY MEASURES
Probability theory (Kolmogorov) is
a quantitative theory of objective uncer-
tainty. The meaning of this uncertainty
is based on the notion of frequency, and
its calculus (i.e., the way probabilities
are combined) is based upon an anal-
ogy with measure theory in mathemat-
ics (with its main axiom of additiv-
ity). There are various different types
of uncertainty, e.g., subjective probabil-
ity, fuzziness. When trying to establish-
ing a quantitative theory for a kind of
uncertainty, we first examine carefully
the meaning of the uncertainty under
consideration, then from which, proceed
to postulate its calculus (i.e., axioms).
This is exemplified by Kolmogorov’s for-
mulation of (objective) probability the-
ory.
Now, the very reason that we are
concerned with uncertainty is decision-
making: how we make decisions in the
face of a type of uncertainty? The an-
swer to this question dictates an appro-
priate calculus for that type of uncer-
54 Asian Journal of Economics and Banking (2020), 4(1), 49–60
tainty. It is “interesting” to ask whether
we should have an axiomatic theory of
uncertainty in advance or derive it from
experimental evidence?
Suppose in natural sciences, uncer-
tainty is objective so that we could
use the frequentist viewpoint to model
uncertainty as (Kolmogorov) probabil-
ity. Does that mean that its calculus
must always obey Kolomogorov’s calcu-
lus? Well, as we all know, one century
ago, that was not true in quantum me-
chanics: while the meaning of probabil-
ity is the same, the calculus of quan-
tum probability is different than that
of Kolmogorov (e.g., non additive and
non commutative). Note however that
quantum probability calculus is a gener-
alization of Kolmogorov probability cal-
culus.
Perhaps following physics, social sci-
entists have looked at various types
of uncertainties encountered in social
problems since quite some time, reveal-
ing various different calculi of uncertain-
ties (See [6] for a survey).
Now, it is well known that there is
a type of uncertainty, called epistemic
uncertainty in Bayesian statistics. The
quantitative modeling of this uncer-
tainty is called “subjective probability”
because of the following reasons. First
of all, it is subjective! The “probability”
of an uncertain phenomenon is varies
from person to person, and is used to
quantify the uncertainty to make deci-
sions. Secondly, it is only in your mind.
And, finally, it is called “probability”
since its calculus (axioms) is taken to be
exactly Kolmogorov’s calculus, despite
their difference in semantics (subjective
probabilities are not frequentist-based).
We have pointed out earlier why we are
led to a situation like this!
Is it the time to take a “closer look”
at the calculus of subjective probabil-
ities? say, in the “spirit” of quantum
physics! Here, we examine [5] and try to
figure out, in view of their “radical re-
formulation”of subjective probability in
behavioral economics and finance, what
should be a new and reasonable calculus
for subjective probabilities?
Since possibility measures, as rea-
sonable quantitative candidates for
“subjective probabilities” in social sci-
ences, seem new to social scientists, we
devote this entire section to a tutorial
on how the intuitive (and well-known)
notion of possibility arises in scientific
studies. Upfront, as stated earlier, we
are going to justify Zadeh’s possibility
theory. Possibility theory is a mathe-
matical theory of a weaker form of un-
certainty (possibility). While the in-
tuitive concept of possibility is famil-
iar in everyday language, it has not
been systematically used as a“scientific”
tool (like probability) in, say, decision-
making. Now, as mentioned above, the
“acceptable” notion of “subjective prob-
ability” (as Bayesians call it) seems to
resemble to possibility rather than prob-
ability (in [5], subjective probability is
not probability!), it is time to make it
clear: What is the calculus of subjec-
tive probability? i.e., how possibility
measures are axiomatized? (just like
objective probability measures are ax-
iomatized by Kolmogorov, from which,
their calculus follows). In the language
of Bayesian statistics, we are basically
talking about the “prior information” of
the true but unknown “parameter” θo in
Hung T. Nguyen/On the Calculus of Subjective Probability in Behavioral Economics 55
the parameter space Φ. We are going
to call it a “possibility distribution” of
θo, i.e., a function pi(.) : Θ→ [0, 1] with
supθ∈Θ pi(θ) = 1, where pi(θ) represents
the possibility (a value in [0, 1]) that θ
is the true parameter. From pi(.), the
notion of possibility that the true pa-
rameter is in a subset A ⊆ Θ is “postu-
lated” to be pi(A) = supθ∈A pi(θ). This
is Zadeh’s theory of possibility measures
[16].
We are going to justify the Zadeh’s
calculus of possibilities. Just like “sub-
jective probabilities”, possibilities are
somewhat subjective and are used to
specify“appraisals”for decision-making.
They can come from your mind (in fact,
based on your available information).
Now, in everyday language, probability
is often used to strengthen possibility,
so that there is some “intuitive” link be-
tween these two notions of uncertainty.
While, once the meaning of possibilities
is spelled out, their calculus could be de-
rived from the so-called “experimental
evidence”, we offer here a justification
based on the intuitive link of probabil-
ity and possibility.
Probability and possibility are some-
what related? We often say “that
some phenomenon is improbable but
possible”, meaning that possibility is a
weaker degree of belief. Typically, such
statement appears in situations where
we face coarse data (imprecise or low
quality data). Let (Ω,A, P ) be a prob-
ability space. Let Θ be a set. Sup-
pose when we perform an experiment,
we observe only the outcome in a sub-
set of Θ (rather than a precise value in
it), i.e., our random experiment is de-
scribed by a random set S(.) : Ω→ 2Θ,
so that our random variable of inter-
est X(.) : Ω → Θ is an a.s. selec-
tor of S. An event A ⊆ Θ is realized
(occured) if X(ω) ∈ A, but if we only
know that X(ω) ∈ S(ω), what can we
say? Well, if S(ω) ⊆ A, then A occurs,
but if S(ω) ∩ A 6= ∅, then it is possi-
ble that A occurs. Thus, we could take
pi(A) = P (ω : S(ω) ∩ A 6= ∅) to quan-
tify the possibility that the event A oc-
curs, i.e., the possibility measure of the
event A. Since X is an almost sure se-
lector of S, we have {ω : X(ω) ∈ A} ⊆
{ω : S(ω) ∩ A 6= ∅}, and hence P (ω :
X(ω) ∈ A) ≤ P (ω : S(ω) ∩ A 6= ∅),
i.e., “probability is smaller than possi-
bility”, as expected. Let A = {θ} be
a singleton set, then the possibility of
θ is pi(θ) = P (ω : S(ω) 3 θ), which
is the coverage function of the random
set S. Thus, formally, a possibility dis-
tribution is the coverage function of a
random set.
The question now is this. What is
the relation between pi(θ) and pi(A)? Of
course, you are thinking about a possi-
ble analogy with probability theory in
which pi(θ) plays the role a a probability
denstity function, whereas pi(A) plays
the role of a probability measure.
Now, let pi(.) : Θ → [0, 1] such that
supθ∈Θ pi(θ) = 1. Then there is a ran-
dom set S(.) on Θ such that P (ω :
S(ω) ∩ A 6= ∅) = supθ∈A pi(θ), for each
A ⊆ Θ. In other words, there is a
canonical random set S whose coverage
function is precisely the possibility dis-
tribution function pi(.), and the possi-
bility measure of any event A ⊆ Θ is
given as supθ∈A pi(θ). Such a result jus-
tifies Zadeh’s postulates pi(.) : Θ ∈ [0, 1]
with supθ∈Θ pi(θ) = 1, and pi(A) =
56 Asian Journal of Economics and Banking (2020), 4(1), 49–60
supθ∈A pi(θ), noting that pi(∅) = 0.
Here is the proof. Let α(.) :
(Ω,A, P ) → [0, 1], uniformly dis-
tributed. Consider the random set S(.) :
(Ω,A, P ) → 2Θ which is the random-
ized level set of the function pi(.) : Θ→
[0, 1], i.e.,
S(ω) = {θ ∈ Θ : pi(θ) ≥ α(ω)},
Then, for A ⊆ Θ, we have
pi(A) = P (ω : S(ω) ∩ A 6= ∅)
= P{ω : α(ω) ≤ sup
θ∈A
pi(θ)}
= sup
θ∈A
pi(θ),
Remark. From pi(A) = supθ∈A pi(θ),
we see that, in particular, the possibility
measure pi(.) on 2Θ, is maxitive (rather
than additive), i.e., for any A,B, we
have pi(A∪B) = max{pi(A), pi(B)}.It is
interesting to note that maxitive opera-
tor can arise from additive operator in
some limiting process. Let Pn, n ≥ 1 be
a sequence of probability measures on a
measurable space (Ω,A), and εn > 0,
with εn → 0 as n→∞. Then it is pos-
sible that the sequence of submeasures
P εnn , n ≥ 1 converges to a (σ−) maxitive
set-function. Here is an example. Let
(Ω,A, P ) be a probability space, and
f(.) : Ω→ R+, f ∈ L∞(Ω,A, P ). Since
P is a finite measure, it follows that
f ∈ Lp(Ω,A, P ), for all p > 0. Con-
sider
Pn(A) =
A
|f(ω)|ndP
Ω
|f(ω)|ndP ,
then, as n→∞,
[Pn(A)]
1
n =
||f1A||n
||f ||n →
||f1A||∞
||f ||∞ ,
where the set function pi(A) =
||f1A||∞
||f ||∞
is (σ−) maxitive.
Of course, if we“consider”a function
pi(.) : Θ → [0, 1] with supθ∈Θ pi(θ) = 1,
and call it our possibility distribution
(just like “probability density”, where
“possibility” reflects our “subjective” as-
signments/ appraisals), then our possi-
bility measure of A ⊆ Θ is taken to
be pi(A) = supθ∈A pi(θ), although we
do not need to relate it to the notion
of probability. In other words, uncer-
tainties might not need to be a function
of probability. For example, a function
pi(.) as above plays the role of a mem-
bership function of a fuzzy subset of Θ,
generalizing ordinary subsets of Θ (see
[15] for more discussions on how to ob-
tain possibility distributions in applica-
tions). What we have done above is to
give a meaning to the concept of possi-
bility distribution (from which possibil-
ity measures are built), namely“a possi-
bility distribution is the coverage func-
tion of a random set”. If you refer to the
concept of confidence intervals in statis-
tics, then you realize the flavor of “sub-
jective” assessments.
It is interesting to note that, for-
mally, the concept of possibility dis-
tribution is somewhat “hidden” within
probability theory! In fact, the notion of
possibility distribution“appears”also in
probability theory as limits of probabil-
ities. When X is an absolutely continu-
ous random variable, e.g., taking values,
say, in R = (−∞,∞), with distribution
function F (x) = P{ω : X(ω) ≤ x},
then its density function value
Hung T. Nguyen/On the Calculus of Subjective Probability in Behavioral Economics 57
f(x) = lim
h→0
F (x+ h)− F (x− h)
2h
= lim
h→0
P{ω : x− h ≤ X(ω) ≤ x+ h})
2h
can be interpreted as a possibility value
(and not a probability value). For ex-
ample, in the context of Bayesian statis-
tics, when the parameter space is Θ =
[0, 1], the uniform prior density f(θ) =
1, for any θ ∈ [0, 1] is a possibility dis-
tribution, resulting in supθ∈A f(θ) = 1
for any A ⊆ [0, 1]. And this can be
applied to unbounded parameter spaces
like R = (−∞,∞) with a possibility
distribution f(θ) = 1, for any θ ∈ R,
without “calling” it an improper prior
probability density function !
To be complete, we mention here
that possibility measures can be also in-
terpreted as limits of probability mea-
sures in the large deviation convergence
sense (See [8], [11]).
Recall that the purpose of the study
of large deviations in probability theory
is to find the asymptotic rate of conver-
gence of sequences of probability mea-
sures of rare events. For example, in the
simplest setting, while the law of large
numbers asserts that the sequence of
sample means (of i.i.d. Xj, j ≥ 1, with
EX = µ, V ar(X) = σ2 < ∞) Xn con-
verges surely to µ, it is also of interest to
find the rate at which P (|Xn − µ| > a)
(probability of large deviations from the
mean) goes to zero, as n→∞, for some
a.
Now, for Zn =
√
n(Xn − µ)/σ, we
have, by the Central Limit Theorem,
P (Zn > x) ≈
∞
x
1√
2pi
e−
t2
2 dt
By finding the asymptotics of this
integral for x → ∞, we arrive at the
well-known Cramer result of the form :
The sequence of sample means satisfies
the large deviation principle in the sense
that, for ε > 0, we have
P (|Xn − µ| > a) ≤ e−nh(a)+ε
for n sufficiently large, where the ex-
ponential rate of convergence being
e−nh(a).
The general setting of large devi-
ation principle is this. Let Θ be a
complete, separable metric space, and
B its borel σ−field. A sequence of
probability measures Pn on (Θ,B) is
said to obey the large deviation prin-
ciple (LDP) if there exists a lower semi-
continuous function I(.) : Θ → [0,∞]
(the rate function) such that :
(i) for each closed set F of Θ,
lim sup
n→∞
[
1
n
logPn(F )] ≤ − inf
θ∈F
I(θ)
(ii) for each open set G of Θ,
lim sup
n→∞
[
1
n
logPn(G)] ≥ − inf
θ∈G
I(θ)
If we let ϕ(A) = supθ∈A e
−I(θ), then
the LDP is formulated as
lim sup
n→∞
[Pn(F )]
1
n ≤ ϕ(F )
and
lim sup
n→∞
[Pn(G)]
1
n ≤ ϕ(G)
58 Asian Journal of Economics and Banking (2020), 4(1), 49–60
If we refer the above to the Port-
mamteau theorem of weak convergence
of probability measures, we can view the
set function ϕ(.) as a limit of Pn in the
sense of LDP.
Remark. Since the possibility dis-
tribution ϕ(θ) = e−I(θ) is upper semi-
continuous (with values in [0, 1]), the
associated possibility measure ϕ(.) char-
acterizes the probability distribution of
a random closed sets on Θ (See [9] for
details).
In summary, a possibility distribu-
tion (for an appraisal) is a function
pi(.) on an arbitrary set Θ, taking val-
ues in the unit interval [0, 1], such that
supθ∈Θ pi(θ) = 1. Its associated possi-
bility measure, denoted also as pi(.), is
a map from the power set 2Θ of Θ to
[0, 1], defined by pi(A) = supθ∈A pi(θ),
for A ⊆ Θ. Equivalently, a possibility
measure is a map pi(.) : 2Θ → [0, 1]
such that pi(∅) = 0, pi(Θ) = 1, and
pi(∪i∈IAi) = supi∈I pi(Ai), for any index
set I.
Remark. For a general theory of
idempotent probability, see [11].
4 BAYESIAN ANALYSIS WITH
POSSIBILITY MEASURES
First of all, in the context of
Bayesian statistics, the additional in-
formation required is a prior proba-
bility measure (or often a probabil-
ity density function) on the parame-
ter space, reflecting the subjective in-
formation about the possible location
of the true (but unknown) parameter
of the population. Such an informa-
tion is quite compatible with a possibil-
ity distribution where the value of the
possibility distribution of an element in
the parameter space acts as the degree
of membership of that element in the
statistician’s perception.
Secondly, while the literature did
contain research works on the possi-
bility to consider “Bayesian statistics”
where both prior and objective proba-
bility measures are replaced by possi-
bility measures, it seems that the more
practical situation where the prior sub-
jective (epistemic) uncertainty is pos-
sibilistic, but the objective uncertainty
(in the statistical models) remains Kol-
mogorov is not completely worked out.
This is the situation where two different
types of uncertainty calculi need to be
combined!
The paper [5] is, in a sense, a con-
firmation of [2], from an original idea
in [16]. And now, with all necessary
justifications spelled ont in the previous
section, it is about time to see how “tra-
ditional”Bayesian statistics is affected?!
This will happen clearly when we simply
replace the calculus of subjective prob-
abilities by the calculus of possibilities
in the formulation of the “Bayes the-
orem”, combining objective probability
with subjective probability.
A situation closely related to what
we wish to discuss here is robust
Bayesian statistics, as exemplified by
[12]. In a sense, robust Bayesian in-
ference refers to the situation where we
consider a set of possible priors (say, as
prior probability measures on the pa-
rameter space) rather just one given
one. Specifically, let P be a set of prob-
ability measures on (Θ,B(Θ)). With-
out knowing a specific prior in P , we
must work with the lower and upper en-
Hung T. Nguyen/On the Calculus of Subjective Probability in Behavioral Economics 59
velops of P , i.e., the non-additive set-
functions L(.) = infP∈P P (.), U(.) =
supP∈P P (.), respectively. However, for
each P ∈ P , we can use Bayes theorem
to get the corresponding posterior prob-
ability measure. Bounds on these pos-
terior probability measures can be ob-
tained from L(.) and U(.). Specifically,
as in [12], these bounds are expressed
as (Choquet) integrals of monotone in-
creasing set functions (called Choquet
capacities) L(.) and U(.), as a general-
ization of ordinary integral calculus.
Now, in one hand, possibility mea-
sures are Choquet capacities, and on
the other hand, if we view a possibil-
ity measure pi(.) as the envelop of a
set of prior probability measures P , i.e.,
say, P = {P : P (.) ≤ pi(.)}, then
we can proceed as in [12], noting that
a possibility measure pi(.) is , in par-
ticular, maxitive, i.e., for A,B in 2Θ,
pi(A∪B) = max{pi(A), pi(B)}. But, any
such set-function is “alternating of infi-
nite order” (and hence, in particular, 2-
alternating), i.e., for any A1, A2, ..., An ,
we have
pi(∩ni=1Ai) ≤
∑
∅ 6=I⊆{1,2,...,n}
(−1)|I|+1pi(∪i∈IAi)
where |I| denotes the cardinality of the
set I. For a proof, see [9].
In general, the problem seems open.
References
[1] Dubois, D., & Prade, H. (2011). Possibility theory and its applications: Where
do we stand? Mathware and soft computing, 18(1), 18–31.
[2] Dubois, D. (2006). Possibility theory and statistical reasoning. Computational
Statistics and Data Analysis, 51, 47-69. DOI:10.1016/j.csda.2006.04.015.
[3] Goodman, I. R., & Nguyen, H. T. (1985). Uncertainty Models for Knowledge-
Based Systems, North-Holland
[4] Goodman, I. R. , Nguyen, H. T., & Rogers, G. S. (1991). On the scoring
approach to admissibility of uncertainty measures in expert systems, J. Math.
Anal. and Appl., 159, 550–594.
[5] Jeffrey, H. J., & Putman, A. O. (2015). Subjective probability on behavioral
economics and finance: A radical reformulation, J. Behavioral Finance, 16,
231–249.
[6] Kreps, D. V. (1988). Notes on The Theory of Choice. Westview Press.
[7] Lindley, D. V. (1982). Scoring rules and the inevitability of probability. Intern.
Statist. Review, 50(1), 1–11.
[8] Nguyen, H. T., & Bouchon-Meunier, B. (2003). Random sets and large devi-
ations principle as a foundation for possibility measures. Soft Computing, 8,
61–70.
60 Asian Journal of Economics and Banking (2020), 4(1), 49–60
[9] Nguyen, H. T. (2006). An Introduction to Random Sets. Chapman and Hall
/ CRC Press.
[10] Nguyen, H. T., Walker, L. C., & Walker, E. A. (2019). A First Course in
Fuzzy Logic (Fourth Edition). Chapman and Hall / CRC Press.
[11] Puhalskii, A. (2001) Large Deviations and Idempotent Probability. Chapman
and Hall / CRC Press.
[12] Wasserman, L. A., & Kadane, J. B. (1990). Bayes’ theorem for Choquet ca-
pacities. Ann. Statist., 18(3), 1328–1339.
[13] Yao, J. T., Vasilakos, A. V., & Pedrycz, W. (2013). Granular computing:
Perspectives and challenges. IEEE Trans Cybern., 34(6), 1977–1989.
[14] Yao, M. X. (2019). Granularity measures and complexity of partition-based
granular structures. Knowledge-Based Systems, 163, 885–897.
[15] Zadeh, L. A. (1996). Fuzzy sets and information granularity. In G.J. Klir and
B. Yuan (Eds), Fuzzy Sets, Fuzzy Logic and Fuzzy Systems (Selected papers
of Lotfi A. Zadeh) (433-448). World Scientific.
[16] Zadeh, L. A. (1978). Fuzzy sets as a basis for a theory of possibility. J. Fuzzy
Sets and Systems, 1, 3–28.
[17] Zadeh, L.A. (1975). The concept of a linguistic variable and its application to
approximate reasoning. I (Information Sci., 8, 199-249), II (Infomation Sci.
8, 301-357), III (Information Sci. 9, 43-80)
Các file đính kèm theo tài liệu này:
- on_the_calculus_of_subjective_probability_in_behavioral_econ.pdf