The purpose of this article is to apply Pascal’s Triangular plus Monte Carlo Analysis instead of the method used in above articles where the template is the same and referred to fuzzy set theory. Then I will compare the final results in which we can say that both of them (methods) are compatible with together while the method of Pascal’s Triangular plus Monte Carlo Analysis is very easier and more reasonable than the method applied in previous articles.
I had illustrated the method of Pascal’s Triangular plus Monte Carlo Analysis on below links:
Therefore, I directly start to analyze previous cases again by using the method of Pascal’s Triangular plus Monte Carlo Analysis as follows:
Case Study: Asset Pricing Model
Example (1)
To find probability distribution inferred from Pascal’s triangular, I chose:
n = 200 for X 1 = 5000 and X2 = 10000
Then, we have below probability distribution:
Cut Offs
|
X
|
0
|
5025
|
7.72E-45
|
7100
|
0.01
|
7275
|
0.1
|
7450
|
0.4
|
7725
|
0.9
|
8550
|
1
|
10000
|
I assigned the formulas of Rand and Vlookup for a1 and a2 as follows:
x
|
0
|
|
|
|
Left Side
|
|
|
Right Side
|
|
Random
|
0.626742
|
|
Random
|
0.58275
|
a1
|
7725
|
|
a1
|
7725
|
a2
|
10000
|
|
a2
|
10000
|
x
|
0
|
|
x
|
0
|
(Formula)1
|
0
|
|
(Formula)1
|
1
|
(Formula)2
|
-3.3956
|
|
(Formula)2
|
4.395604
|
(Formula)3
|
1
|
|
(Formula)3
|
0
|
Alpha -cut
|
0
|
|
Alpha -cut
|
1
|
Then, I got two ways data table for (x) between 5000 and 10000 with 400 iterative calculations for left side and right side as follows:
Finally, I calculated the average for α –Cut (Left) and α –Cut (Right) as follows:
x
|
8400
|
8600
|
8800
|
9000
|
9200
|
α –Cut (Left)
|
0.301045
|
0.366571
|
0.471798
|
0.557612
|
0.642259
|
α –Cut (Right)
|
0.696507
|
0.622444
|
0.535629
|
0.441788
|
0.349999
|
|
|
|
|
|
|
|
|
|
α - Cut
|
0.503714
|
|
As we can see, α –Cut (Left) and α –Cut (Right) are approximately close around x = 8800 and α –Cut = 0.5
Of course, I chose ∆x = 200. If we consider ∆x = 100, the final results will be more accurate.
In the previous article, the best price of the first try to advertise was equal to $8883.33 with confidence level of 0.52 (α = 0.52). We can see that the final results are compatible.
Example (2)
To find probability distribution inferred from Pascal’s triangular, I chose:
n = 200 for X 1 = 4000 and X2 = 6000 (Right side)
n = 200 for Y1 = 6000 and Y2 = 8000 (Left side)
Then, we have below probability distribution:
Cut Offs
|
X
|
|
4010
|
7.72E-45
|
4840
|
0.01
|
4910
|
0.1
|
4980
|
0.4
|
5090
|
0.9
|
5420
|
1
|
6000
|
Cut Offs
|
Y
|
|
6010
|
7.72E-45
|
6840
|
0.01
|
6910
|
0.1
|
6980
|
0.4
|
7090
|
0.9
|
7420
|
1
|
8000
|
I assigned the formulas of Rand and Vlookup for a1 and a2 as follows:
x
|
4000
|
|
y
|
6000
|
Right Side
|
|
|
Left Side
|
Random
|
0.858574
|
|
Random
|
0.701062
|
a1
|
5090
|
|
a1
|
7090
|
a2
|
10000
|
|
a2
|
10000
|
x
|
4000
|
|
y
|
6000
|
(Formula)1
|
0
|
|
(Formula)1
|
1
|
(Formula)2
|
-0.222
|
|
(Formula)2
|
1.37457
|
(Formula)3
|
1
|
|
(Formula)3
|
0
|
Alpha -cut
|
0
|
|
Alpha -cut
|
1
|
Then, I got two ways data table for (x) between 4000 and 10000 and for (y) between 6000 and 10000 with 400 iterative calculations for left side and right side as follows:
(Of course, the best range for both of them is (x) and (y) between 6000 and 10000)
Finally, I calculated the average for α –Cut (Left) and α –Cut (Right) as follows:
For ∆x = 100, we have:
(x, y)
|
7900
|
8000
|
8100
|
8200
|
8300
|
8400
|
α –Cut (Right)
|
0.574654
|
0.593947
|
0.614103
|
0.63463
|
0.65446
|
0.675476
|
α –Cut (Left)
|
0.71642
|
0.681869
|
0.649753
|
0.61348
|
0.58049
|
0.547297
|
|
|
|
|
|
|
|
|
|
|
α –Cut
|
0.62405
|
|
|
For ∆x = 200, we have:
(x, y)
|
7600
|
7800
|
8000
|
8200
|
8400
|
8600
|
α –Cut (Right)
|
0.512874
|
0.553806
|
0.594213
|
0.634706
|
0.67472
|
0.715272
|
α –Cut (Left)
|
0.820029
|
0.753847
|
0.681916
|
0.614954
|
0.548512
|
0.479935
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
α –Cut
|
0.62483
|
|
|
As we can see, α –Cut (Left) and α –Cut (Right) are approximately close around x = 8200 and α –Cut = 0.62
In the previous article the best price of the first try to advertise was equal to $8183.33 with confidence level of 0.4 (α = 0.4). We can see that the final results are compatible.
Case Study: Predictions in temperature transferring
Warming to Cooling
|
|
Cooling to Warming
|
Ta
|
5
|
|
Ta
|
100
|
T0
|
100
|
|
T0
|
5
|
t
|
34
|
|
t
|
1
|
k
|
0.055
|
|
k
|
0.0106
|
T (t)
|
19.641748
|
|
T (t)
|
6.001681708
|
µ
|
0.1964175
|
|
µ
|
0.060016817
|
α-cut
|
0.1964175
|
|
α-cut
|
0.060016817
|
Firstly, I chose the same random range for “t” and also “k”. For instance, t1 = 0 to t2 = 100 and k1 = 0 to k2 = 100.
After that, I found probability distribution inferred from Pascal’s triangular for:
n = 200 for k1 = 0 and k2 = 100
Where we have below probability distribution:
Cut Offs
|
k
|
0
|
0.5
|
7.716E-45
|
42
|
0.01
|
45.5
|
0.1
|
49
|
0.4
|
54.5
|
0.9
|
71
|
0.9999999
|
100
|
Then, I used the formulas of Rand and Vlookup for “k” as follows:
t
|
0
|
|
|
|
Warming to Cooling
|
Cooling to Warming
|
Ta
|
5
|
|
Ta
|
100
|
T0
|
100
|
|
T0
|
5
|
t
|
0
|
|
t
|
0
|
Rand
|
0.4782881
|
|
Rand
|
0.0027308
|
k
|
54.5
|
|
k
|
42
|
T (t)
|
100
|
|
T (t)
|
5
|
µ
|
1
|
|
µ
|
0.05
|
α-cut
|
1
|
|
α-cut
|
0
|
I got two ways data table for “t” and α –Cut as follows:
As we can see, all results are same and equal to zero. Therefore, I decreased the range of “k” to [0, 10] and I repeated above steps where I found the same data table as follows:
Finally, I decreased the range of “k” to [0, 0.1] and I repeated again above steps where I found below data table:
Cut Offs
|
k
|
|
0.0005
|
7.72E-45
|
0.042
|
0.01
|
0.0455
|
0.1
|
0.049
|
0.4
|
0.0545
|
0.9
|
0.071
|
0.9999999
|
0.1
|
Thus we can consider the range of “k” between 0 and 0.1.
In the next step, I did a sensitivity analysis for “t” in the range of [0, 100] and α –Cut (Left) and α –Cut (Right) as follows:
t
|
α-cut
|
α-cut
|
mean
|
STDV
|
CV
|
0
|
0.949611
|
0.115111
|
0.532361
|
0.59008
|
1.108421
|
1
|
0.949611
|
0.100389
|
0.525
|
0.60049
|
1.143791
|
2
|
0.917367
|
0.148106
|
0.532736
|
0.543949
|
1.021048
|
3
|
0.856708
|
0.171214
|
0.513961
|
0.484718
|
0.943102
|
4
|
0.813919
|
0.219088
|
0.516504
|
0.420609
|
0.814338
|
5
|
0.806696
|
0.2766
|
0.541648
|
0.374835
|
0.692026
|
6
|
0.758013
|
0.276962
|
0.517487
|
0.340154
|
0.657319
|
7
|
0.740877
|
0.351305
|
0.546091
|
0.275469
|
0.504437
|
8
|
0.664287
|
0.461675
|
0.562981
|
0.143268
|
0.254481
|
9
|
0.631704
|
0.418296
|
0.525
|
0.150902
|
0.287433
|
10
|
0.631995
|
0.44915
|
0.540573
|
0.129291
|
0.239174
|
11
|
0.571632
|
0.424084
|
0.497858
|
0.104332
|
0.209562
|
12
|
0.543964
|
0.594767
|
0.569365
|
0.035923
|
0.063094
|
13
|
0.552433
|
0.497567
|
0.525
|
0.038796
|
0.073897
|
14
|
0.528407
|
0.521593
|
0.525
|
0.004818
|
0.009178
|
15
|
0.50553
|
0.580543
|
0.543036
|
0.053042
|
0.097676
|
16
|
0.447209
|
0.602791
|
0.525
|
0.110013
|
0.209549
|
17
|
0.42614
|
0.715861
|
0.571001
|
0.204864
|
0.35878
|
18
|
0.443256
|
0.735336
|
0.589296
|
0.206531
|
0.350471
|
19
|
0.424451
|
0.625549
|
0.525
|
0.142197
|
0.270852
|
As we can see, “t” between 12 min and 15 min has the least CV and STDEV. Therefore, I used Pascal’s Triangular plus Monte Carlo Analysis for “t” as follows:
n = 200, t1 = 12, t2 = 15 where the probability distribution is below cited:
Cut Offs
|
x
|
0
|
12.015
|
7.72E-45
|
13.26
|
0.01
|
13.365
|
0.1
|
13.47
|
0.4
|
13.635
|
0.9
|
14.13
|
0.9999999
|
15
|
I applied the Rand and Vlookup in excel in which the algorithm was changed as follows:
k
|
0
|
|
|
|
Warming to Cooling
|
Cooling to Warming
|
Ta
|
5
|
|
Ta
|
100
|
T0
|
100
|
|
T0
|
5
|
Rand
|
0.85862
|
|
Rand
|
0.978536
|
t
|
13.635
|
|
t
|
14.13
|
k
|
0
|
|
k
|
0
|
T (t)
|
100
|
|
T (t)
|
5
|
µ
|
1
|
|
µ
|
0.05
|
α-cut
|
1
|
|
α-cut
|
0
|
I obtained the sensitivity analysis for “k” and α –Cut (Left and Right). The final results are as follows:
k
|
α-cut
|
α-cut
|
mean
|
STDEV
|
CV
|
0.039
|
0.608189
|
0.441811
|
0.525
|
0.117647
|
0.224089
|
0.04
|
0.600629
|
0.460166
|
0.530398
|
0.099323
|
0.187261
|
0.041
|
0.593173
|
0.450781
|
0.521977
|
0.100686
|
0.192893
|
0.042
|
0.589543
|
0.464183
|
0.526863
|
0.088642
|
0.168246
|
0.043
|
0.57856
|
0.47144
|
0.525
|
0.075746
|
0.144278
|
0.044
|
0.571402
|
0.478598
|
0.525
|
0.065623
|
0.124996
|
0.045
|
0.568174
|
0.496989
|
0.532582
|
0.050336
|
0.094513
|
0.046
|
0.557376
|
0.488759
|
0.523067
|
0.04852
|
0.09276
|
0.047
|
0.550505
|
0.499495
|
0.525
|
0.036069
|
0.068703
|
0.048
|
0.543727
|
0.506273
|
0.525
|
0.026483
|
0.050445
|
0.049
|
0.53704
|
0.509006
|
0.523023
|
0.019823
|
0.037901
|
0.05
|
0.530445
|
0.515575
|
0.52301
|
0.010514
|
0.020103
|
0.051
|
0.527943
|
0.526062
|
0.527003
|
0.00133
|
0.002524
|
0.052
|
0.52413
|
0.53248
|
0.528305
|
0.005904
|
0.011176
|
0.053
|
0.515239
|
0.532164
|
0.523702
|
0.011968
|
0.022852
|
0.054
|
0.509015
|
0.545057
|
0.527036
|
0.025486
|
0.048357
|
0.055
|
0.498782
|
0.547127
|
0.522954
|
0.034185
|
0.065369
|
0.056
|
0.492704
|
0.557296
|
0.525
|
0.045673
|
0.086996
|
0.057
|
0.486709
|
0.559164
|
0.522937
|
0.051234
|
0.097973
|
0.058
|
0.484937
|
0.565063
|
0.525
|
0.056657
|
0.107918
|
0.059
|
0.474961
|
0.575039
|
0.525
|
0.070766
|
0.134793
|
0.06
|
0.473377
|
0.580794
|
0.527085
|
0.075956
|
0.144105
|
0.061
|
0.467712
|
0.586471
|
0.527092
|
0.083976
|
0.159319
|
Above table shows us, for k = 0.051 and α – cut = 0.53 we have the least CV and STDEV.
Note: “All spreadsheets and calculation notes are available. The people, who are interested in having my spreadsheets of this method as a template for further practice, do not hesitate to ask me by sending an email to: soleimani_gh@hotmail.com or call me on my cellphone: +989109250225. Please be informed these spreadsheets are not free of charge.”
In the previous article, we had the answer for “k” approximately equal to 0.055 per min and
α – cut = 0.54
Therefore, we can see both methods are compatible with together.
But why am I following the constant points or values, intersections and distance among fuzzy numbers? Because, if we are willing to utilize Fuzzy Logic Control (FLC) as the methodology to analyze and solve some complex cases, the most important step is to evaluate the rules in FLC. In classic or theory physics, if we have to solve the complex cases, the evaluation of the rules will be easier than other cases because we usually encounter the universal laws or constant points (referred to article of “The Constant Issues, Universal Laws and Boundaries Conditions in Physics Theory” posted on link: http://emfps.blogspot.com/2011/10/constant-issues-universal-laws-and.html). In the other complex cases such as strategic management or financial management, the tracking the intersections and consequently the distance among fuzzy numbers to evaluate the rules in FLC methodology is very important. On the other hand, if we are using FLC as the methodology to analyze the complex cases in the field of strategic management or designing a strategic plan, we should know that the rules are changing the moment to moment and we have to replace new rules instead of old rules rapidly. How? I think that the article of “EMFPS: How Can We Get the Power Set of a Set by Using of Excel?” posted on Link: http://emfps.blogspot.com/2012/08/emfps-how-can-we-get-power-set-of-set.html) let us increase our speed to replace new rules.