15

I have some problem while trying to calculate time complexity for this code:

function foo (int a):
    if a < 1: 
        return 1
    else:
        for i = 1 to 4:
            foo(a - 3)
        for i = 1 to 4:
            foo(a / 2)
end function

As far as I can go:

T(n) = 1 if n<1

T(n) = 4T(n-3) + 4T(n/2)     if n>=1

     = 4(4T(n-6) + 4T((n-3)/2))  +  4(4T(n/2 - 3) + 4T(n/4))

     ~ 4^2 (T(n-6) + T((n-3)/2) + T(n/2-3) + T(n/4))

Now, it is very complicated, since number of the next T increase by 2^n and also the child is quite complicated.

Is there any other ways to solve this problem?

Claudiu
  • 224,032
  • 165
  • 485
  • 680
Frzzy
  • 299
  • 1
  • 2
  • 16
  • Foo will at best just return 1 with the code posted – Ed Heal Sep 17 '15 at 16:36
  • 5
    There's the empirical way. Run it with n=10, 100, 1000, 10.000, ... and plot the curve. It should at least give you an intuiton, if that's enough – vmg Sep 17 '15 at 16:37
  • 2
    This doesn't look like C or C++ code to me. Is it pseudocode? What is the meaning of the indentation of the second for loop, is it part of the "else" block or not? In C / C++ it wouldn't be. – Chris Beck Sep 17 '15 at 16:40
  • 1
    @vmg: That's not really true, the low end data points are going to be highly affected by caching, pipelining, optimization, inlining and all sorts of other things the chip and compiler will actually do. If you want to get into the regime where the big O complexity of the algorithm is sure to dominate you have to run with very large numbers. I think it's not usually very practical to try to 'deduce' the big O complexity this way. – Chris Beck Sep 17 '15 at 16:42
  • @EdHeal: although the result is a constant function, the question deals with the time complexity. – Willem Van Onsem Sep 17 '15 at 16:43
  • 4
    @ChrisBeck: timing is indeed out of question, but adding a counter of the number of calls to `foo` is trivial. –  Sep 17 '15 at 16:44
  • True. Chris Beck is right. – vmg Sep 17 '15 at 16:44
  • @Frzzy: are you interested in the time complexity, or the exact number of steps. – Willem Van Onsem Sep 17 '15 at 16:45
  • @CommuSoft "I have some problem while trying to calculate time complexity for this code:" – nbro Sep 17 '15 at 16:45
  • @YvesDaoust: that's a pretty good point – Chris Beck Sep 17 '15 at 16:46
  • @Axl: yeah, but experience shows that most people use both concepts interchangeably. – Willem Van Onsem Sep 17 '15 at 16:46
  • I'm interseted in time complexity more @commusoft. I'm trying to calculate this big(o) – Frzzy Sep 17 '15 at 16:47
  • I would say that for `a > 3` (or something similar), the second loop should take less time, since division by 2 should reach a smaller number faster than subtracting `-3`, but this is just a intuition. So, you should just concentrate your brain on trying to understand the complexity of the first loop and all the recursion calls involved... – nbro Sep 17 '15 at 16:47
  • @Axl: that is my idea as well. I think it is safe to ignore the second `for` loop since addition in complexity theory is the same as maximum so to speak. – Willem Van Onsem Sep 17 '15 at 16:50
  • I'm trying to the problem by avoid unnecessary constant, and it may lead too T(N) = 4^n (T(n) + 2^n T(n/2) + 2^(n-1) T(n/4) + .... + 2^(n-k) T(n/k)) and then i'm stuck again – Frzzy Sep 17 '15 at 16:52
  • Check my answer, it gives precise results. –  Sep 18 '15 at 07:51

3 Answers3

14

Let's expand the recursive cost function:

T(n) = 4   [T(n-3) + T(n/2)]
T(n) = 4^2 [T(n-6) + T((n-3)/2) + T((n-6)/2) + T(n/4)]
T(n) = 4^n [T(n-9) + 2*T((n-6)/2) + T((n-9)/2) + T((n-12)/4) + T((n-3)/4) + T((n-6)/4) + T(n/8)]

From the moment the x in T(x) drops below 1, you should replace T(x) with 1. And from that moment on, the T(x) doesn't generate any "children" anymore so to speak.

what does this means? It means that after the k-'th expansion of T(n), the function will look like:

T(n) = 4^k [number of paths with length `k`]

and keep increasing k until all paths have "died". This is definitely the case after n/3 iterations, because that's the longest possible path.

We thus have some kind of graph, for instance for n=9:

9 + 6 + 3 + 0
  |   |   ` 1
  |    `3 + 0
  |       ` 1
   `4 + 1
      ` 2 + -1
         `  1

so 6 paths. Now the problem is how to count the number of paths. In order to do this we will first represent the main path: n, n-3, n-6, etc. as a horizontal line of nodes, this is definitely the longest path:

n    n-3  n-6  n-9  ...  1

Now out of all these nodes, do originate nodes of i -> i/2 (except for one)

n    n-3      n-6      n-9     ...   4   1
|     |        |         |
n/2  (n-3)/2  (n-6)/2  (n-9)/2 ...   2

(the second row shows all nodes created by divisions by 2). Now these nodes, generate again offsprong n -> n-3, which is, since it is divided by two n/2 -> (n-6)/2, in other words, there are edges that make jumps of two:

n    n-3      n-6      n-9     ...   4   1
|     |  /-----+-------(n-9)/2       |
n/2  (n-3)/2  (n-6)/2  (n-9)/2 ...   2
  \---------->(n-6)/2 \------->...

on other words, except for the first two elements, all other nodes in the second row count for two. If we would represent it as some kind of graph with the nodes labeled by their weight, it would look like:

   1 -- 1 -- 1 -- 1 -- 1 -- .. -- .. -- 1
   |    |    |    |    |    |     |
   1 -- 1 -- 2 -- 2 -- 2 -- .. -- 2

Or if we keep doing this for this process:

   1 -- 1 -- 1 -- 1 -- 1 -- .. -- .. -- .. -- .. -- ..-- 1
   |    |    |    |    |    |     |     |     |     |
   1 -- 1 -- 2 -- 2 -- 2 -- .. -- .. -- .. -- .. -- 2
   |    |    |    |    |    |     |     |
   1 -- 1 -- 2 -- 2 -- 3 -- .. -- .. -- 4

(the third row generates children 4 items further)

Now we need to calculate the sum of the last row. This is at most O(log n).

Which thus results in an upper bound of O(4^(n/3)*log n) maximum. It is definitely possible that the bound is tighter, or 4^(n/3+epsilon), the log doesn't really matter when it comes down to the exponent.

Experiments

One can turn the program into a program that calculates the cost (used Python):

def memodict(f):
    """ Memoization decorator for a function taking a single argument """
    class memodict(dict):
        def __missing__(self, key):
            ret = self[key] = f(key)
            return ret
    return memodict().__getitem__

@memodict
def foo (a):
  if a < 1:
    return 1
  else:
    return 1+4*(foo(a-3)+foo(a//2))

for i in range(1000) :
    print '{0} {1}'.format(i,foo(i))

mind the 1+ (this is due to the fact that calling a method not at the leaves requires computational cost as well).

It shows the following graph (with the y axis in log space):

complexity graph

If one looks very closely it looks as if log n is a better estimate. Although I don't know if it is safe to say this.

This results in a table (below, calculated it further up to 2'000).

1 9
2 41
3 41
4 201
5 329
6 329
7 969
8 2121
9 2121
10 5193
11 9801
12 9801
13 22089
14 43081
15 43081
16 96841
17 180809
18 180809
19 395849
20 744009
21 744009
22 1622601
23 3015241
24 3015241
25 6529609
26 12149321
27 12149321
28 26290761
29 48769609
30 48769609
31 105335369
32 195465801
33 195465801
34 422064713
35 782586441
36 782586441
37 1688982089
38 3131929161
39 3131929161
40 6758904393
41 12530692681
42 12530692681
43 27038593609
44 50129261129
45 50129261129
46 108166435401
47 200529105481
48 200529105481
49 432677802569
50 802142540361
51 802142540361
52 1730759807561
53 3208618758729
54 3208618758729
55 6923087827529
56 12834580197961
57 12834580197961
58 27692546388553
59 51338515870281
60 51338515870281
61 110770380632649
62 205354484822601
63 205354484822601
64 443082304393801
65 821418721153609
66 821418721153609
67 1772329999438409
68 3285676572873289
69 3285676572873289
70 7089323128099401
71 13142709421838921
72 13142709421838921
73 28357295642743369
74 52570844443284041
75 52570844443284041
76 113429195098690121
77 210283390300852809
78 210283390300852809
79 453716792922477129
80 841133588239028809
81 841133588239028809
82 1814867221812679241
83 3364534403078885961
84 3364534403078885961
85 7259468937373487689
86 13458137720469918281
87 13458137720469918281
88 29037875950010995273
89 53832551082396717641
90 53832551082396717641
91 116151504000561025609
92 215330204762252612169
93 215330204762252612169
94 464606016804360524361
95 861320819851126870601
96 861320819851126870601
97 1858424068019558519369
98 3445283281135218692681
99 3445283281135218692681
100 7433696275286804238921
101 13781133127749444932169
102 13781133127749444932169
103 29734785104355787117129
104 55124532517920818958921
105 55124532517920818958921
106 118939140430257623503433
107 220498130084517750870601
108 220498130084517750870601
109 475756561733864969048649
110 881992520365763354792521
111 881992520365763354792521
112 1903026246986798196986441
113 3527970081514391739961929
114 3527970081514391739961929
115 7612104987998531108737609
116 14111880326168337145401929
117 14111880326168337145401929
118 30448419952199478498431561
119 56447521304878702645088841
120 56447521304878702645088841
121 121793679809003268057207369
122 225790085219957892102885961
123 225790085219957892102885961
124 487174719236834490168119881
125 903160340880652986350834249
126 903160340880652986350834249
127 1948698876948159378611769929
128 3612641363524384274620912201
129 3612641363524384274620912201
130 7794795507795923189331694153
131 14450565454100822773368263241
132 14450565454100822773368263241
133 31179182031186978432211391049
134 57802261816410380413470806601
135 57802261816410380413470806601
136 124716728124761056435137057353
137 231209047265654664360174719561
138 231209047265654664360174719561
139 498866912499057368446839722569
140 924836189062647014733211275849
141 924836189062647014733211275849
142 1995467649996282044625046245961
143 3699344756250640629770532459081
144 3699344756250640629770532459081
145 7981870599985180749337872339529
146 14797379025002675948264700809801
147 14797379025002675948264700809801
148 31927482399940933280729262494281
149 59189516100010914076436576375369
150 59189516100010914076436576375369
151 127709929599763943406294823113289
152 236758064400044110022526700261961
153 236758064400044110022526700261961
154 510839718399056614758740495864393
155 947032257600177281223668004459081
156 947032257600177281223668004459081
157 2043358873596227300168523186868809
158 3788129030400710939761843707744841
159 3788129030400710939761843707744841
160 8173435494384912565208445703590473
161 15152516121602847123581727787094601
162 15152516121602847123581727787094601
163 32693741977539653625368135770477129
164 60610064486411395753795798399095369
165 60610064486411395753795798399095369
166 130774967910158627959610155397452361
167 242440257945645596473320805911925321
168 242440257945645596473320805911925321
169 523099871640634525296578233905353289
170 969761031782582414931158973141652041
171 969761031782582414931158973141652041
172 2092399486562538155018863817501086281
173 3879044127130329713557186774446281289
174 3879044127130329713557186774446281289
175 8369597946250152673908006151884018249
176 15516176508521318970380250897829106249
177 15516176508521318970380250897829106249
178 33478391785000610910962228937122943561
179 62064706034085276096851207920903295561
180 62064706034085276096851207920903295561
181 133913567140002443859179120078078644809
182 248258824136341104852010847685857284681
183 248258824136341104852010847685857284681
184 535654268560009776298037299361325027913
185 993035296545364420269364209792439587401
186 993035296545364420269364209792439587401
187 2142617074240039106053470016494310560329
188 3972141186181457682935880906387200447049
189 3972141186181457682935880906387200447049
190 8570468296960156427659163345381749723721
191 15888564744725830735188806904953309270601
192 15888564744725830735188806904953309270601
193 34281873187840625714081936660931506377289
194 63554258978903322948188923891891471159881
195 63554258978903322948188923891891471159881
196 137127492751362502870108879768266900279881
197 254217035915613291806536828692106759410249
198 254217035915613291806536828692106759410249
199 548509971005450011494216652197608475890249
200 1016868143662453167255882099869574254596681
Willem Van Onsem
  • 443,496
  • 30
  • 428
  • 555
  • yeah so usually the issue with these kinds of questions is not seeing intuitively what terms are negligible, but arguing it rigorously in a straightforward way. I agree with what you say, but it's not a very rigorous answer – Chris Beck Sep 17 '15 at 17:03
  • @ChrisBeck: True, I'm trying to come up with a better way to formulate it. – Willem Van Onsem Sep 17 '15 at 17:03
  • @ChrisBeck: I at least proved an upper bound of O(4^(n log n)), but I think the weight in the lower rows will be significantly less, do you have some inspiration to reduce the factor further? – Willem Van Onsem Sep 17 '15 at 17:44
  • I have a question, the loop of foo(a/2) contain inside foo(a/2-3) too. Can we avoid that too? – Frzzy Sep 17 '15 at 17:59
  • @Frzzy: that's what I was wondering about. Given we take this into account (see updated answer), the upper bound is at most `O(4^(n/3)*n*log n)` – Willem Van Onsem Sep 17 '15 at 18:01
  • The factor `n.Log(n)` is extraneous. –  Sep 17 '15 at 18:27
  • 1
    @CommuSoft: I rewrote my answer, I think I found a decent way to do it – Chris Beck Sep 17 '15 at 19:27
  • @ChrisBeck: I think this is indeed a good explanation. One must be careful because O(r^n) and O(r^(2*n)) mean something different. – Willem Van Onsem Sep 17 '15 at 19:48
7

(Rewrote to give a better answer.)

Here is a simple and rigorous analysis that shows why T(n) ~ 4^{n/3} is a tight estimate.

We have the recurrence

T(n) = 4T(n-3) + 4T(n/2)

To get the tight result, we want to see that T(n/2) is negligible compared to T(n-3). We can do this as follows.

First, T is nonnegative for all n, so in particular T(n/2) >= 0, so for all n we have an inequality,

T(n) >= 4T(n-3)

Now, we want to use that inequality to compare T(n-3) and T(n/2). By applying that inqeuality n/6 - 1 times, we get that

T(n-3) >= 4^{n/6 - 1} * T(n/2)

(Because, (n/6 - 1) * 3 = n/2 - 3, and n/2 - 3 + n/2 = n - 3).

It implies that T(n/2) is small compared to T(n-3):

T(n/2) <= 4^{-n/6 + 1} * T(n-3)

Now, for any epsilon > 0, there is an n_0 such that for n > n_0, 4^{-n/6 + 1} < epsilon. (Because, the limit of 4^{-n/6 + 1} is zero as n gets large.)

This implies that for any epsilon > 0, there is large enough n so that

4T(n-3) <= T(n) <= (4 + epsilon) T(n-3)

This yields the tight bound T(n) = 4^(n/3 + o(n)).


Getting a sharper estimate

There's some question in the comments about getting rid of the o(n) above, to get an even sharper estimate.

I fear this is basically just going to get pedantic -- usually no one cares about the low order terms, and nailing them down exactly is just some calculus work. But we can do a little more today anyways.

What's the difference

First of all, what is the difference between O(4^{n/3}) and 4^{n/3 + o(n)}? (Alternatively, we could write the latter as (4+o(1))^{n/3}.)

The difference is in how tightly they control the low order terms. O(4^{n/3}) controls them very tightly -- it says you don't exceed the (concrete) value 4^{n/3}) by more than a constant factor.

4^{n/3 + o(n)}, allows that you may exceed 4^{n/3} by more than a constant factor. But that factor is subexponential in n, it's negligible compared to 4^{n/3}.

For example, consider the function f(n) = n * 4^{n/3}. This function is not O(4^{n/3}). Indeed, it exceeds it by a factor n, more than a constant factor.

However, f(n) is in the class 4^{n/3 + o(n)}. Why? Because n = O(4^{epsilon n}) for every epsilon > 0.

When you have an inequality like,

4T(n-3) <= T(n) <= (4 + epsilon) T(n-3)

for every epsilon > 0, you can only deduce from this T(n) = (4 + o(1))^{n/3}.

To get a sharper bound, we need to treat epsilon as a function of n and not as a constant (like I did in the lazier version.)

Proof

Let epsilon(n) = 4^{-n/6 + 1} in what follows. Then we already showed

T(n) <= (4 + epsilon(n)) T(n-3)

and we want to see T = O(4^{n/3}).

This is can be expanded as an iterated product:

T(n) = PI_{i=1}^{n/3} (4 + epsilon(3i))

We can factor each term and pull out a factor of 4 to get

T(n) = 4^{n/3} * PI_{i=1}^{n/3} (1 + epsilon(3i)/4 )

The goal is now to show that

PI_{i=1}^{n/3} (1 + epsilon(3i)/4 ) = O(1)

and then we will be finished.

To do this we take the log, and show that that is O(1).

SUM_{i=1}^{n/3} log(1 + epsilon(3i/4))

We bound that using log(1+x) <= x for x >= 0.

SUM_{i=1}^{n/3} epsilon(3i/4)

Now we use the definition of epsilon. In fact we only need to know epsilon(n) <= C^{-n} for some C > 1. The above becomes

SUM_{i=1}^{n/3} C'^{-i}

for some constant C' > 1. But this is a geometric series, so it is bounded above by the infinite geometric series as

1 / (1 - 1/C') = O(1)

Thus T(n) = O(4^{n/3}). Since we already had T(n) = Omega(4^{n/3}) we now have it tight up to constants, T(n) = Θ(4^{n/3})

You can decide for yourself if this extra work made things any more clear :p Personally I prefer to leave the o(n)'s in there usually.

Chris Beck
  • 15,614
  • 4
  • 51
  • 87
  • +/2-1/. I think this is definitely the most conservative (and thus definitely correct approach). – Willem Van Onsem Sep 17 '15 at 19:46
  • (that was plus one, but that's apparently not allowed in a comment anymore). – Willem Van Onsem Sep 17 '15 at 19:57
  • `o(n)` is unnecessary. –  Sep 18 '15 at 07:53
  • @YvesDaoust: Yeah but it requires more calculus to see that... usually no one cares about the `o` terms. Just want a simple argument that gives a good estimate. Could try to eliminate the `o` terms I guess just to help a student see how that can be done, but that's the only reason I think. – Chris Beck Sep 18 '15 at 15:51
  • By the way, `T(n)` isnt' `4^(n/3 + o(n))`, it is "`O(4^(n/3 + o(n))`" (if that notation makes any sense), and this is strictly `O(4^(n/3))` as `n` absorbs `o(n)`. –  Sep 18 '15 at 16:22
  • Yves Daoust: No, that's playing a little fast and loose. Consider for instance the function `f(n) = n * 4^{n/3}`. It's true that `4^{n/3} <= f(n) <= (4 + epsilon)^{n/3}`, for any `epsilon > 0`. But it's not true that `f(n) = O(4^{n/3})`. It is true though that `f(n) = 4^{n/3 + o(n)}`. On the other hand, it's always true that `O(4^{n/3 + o(n)}) = 4^{n/3 + o(n)}`, because the constant outside can always be absorbed into the exponent, and `O(1) = o(n)`. – Chris Beck Sep 18 '15 at 16:30
3

IMO, the time complexity is Θ(r^n), where r=³√4.

Indeed, plugging this expression in the recurrence relation,

r^n = 1 + 4 r^n / r³ + 4 r^(n/2) = 1 + r^n + 4 √(r^n),

where the second term dominates asymptotically.

Here is a plot of the exact number of total calls to foo, divided by r^n for easy reading. We assumed the floor [n/2] in f(n/2).

The ratios tend to the repeating sequence 46.6922952502, 63.4656065932 74.1193985991. This seems to confirm Θ(r^n).

enter image description here

Update:

By induction we can show that for n >= 21,

T(n) < B(n) = 75.(s^(2n) - 4.s^n),

with s=³√2.

Indeed, by the recurrence equation and the induction hypothesis,

T(n+3) = 1 + 4.T(n) + 4.T([(n+3)/2])
       < 1 + 4.75.(s^(2n) - 4.s^n) + 4.75.(s^(2[(n+3)/2])) - 4.s^[(n+3)/2])

We compare this to the bound B(n+3) to establish

1 + 4.75.(s^(2n) - 4.s^n) + 4.75.(s^(2[(n+3)/2])) - 4.s^[(n+3)/2])
     < 75.(s^(2n+6) - 4.s^[(n+3)/2]

We can simplify the terms 4.75.s^(2n) and divide by 300.s^n:

s^(-n)/300 - 4 + s^(-(n+3)%2) - 4.s^([(n+3)/2]-n) < - s^([(n+3)/2]-n)

or

s^(-n)/300 + s^(-(n+3)%2) < 4 + 5.s^([(n+3)/2]-n).

This inequality is true for any n, so that T(n) < B(n) => T(n+3) < B(n+3).

Now for the base case, we use the table of T(n) given by @CommuSoft (and checked independently) and verify numerically

T(21) = 744009 < 75.(s^42 - 4.s^21) = 1190400
T(22) = 1622601 < 75.(s^44 - 4.s^22) = 1902217.444...
T(23) = 3015241 < 75.(s^46 - 4.s^23) = 3035425.772...
...
T(41) = 12530692681 < 75.(s^82 - 4.s^41) = 12678879361

This shows that the induction step can be applied from n=39 onwards ([(39+3)/2]=21).

Then

T(n) = O(75.(s^(2n) - 4.s^n)) = O(r^n).

(Actually, for all n >= 23, 46.r^n < T(n) < 75.r^n and this is very tight; T(n) = Θ(r^n).)

  • I don't know whether the fact that the first term dominates asymptotically is *enough*. I can imagine scenario's like the sum of the harmonic sequence, where the tail will grow enough that it will create a significant impact. – Willem Van Onsem Sep 17 '15 at 18:05
  • Well, you can say, you create `log n` of such processes stacked upon each other, since for each half, you still generate an approximately equivalent sequence calls. – Willem Van Onsem Sep 17 '15 at 18:10
  • No, you need to think of this as a matrix. The first sequence of processes is n,n-3, n-6,... 1, thus n/3 processes. Each of these processes will create a "second" layer of n/2, n-3/2,...,1 processes, and thus in a 2d matrix (although there are parts cut out at the right side, because the sequence will be shorter). Now these will again create processes,... – Willem Van Onsem Sep 17 '15 at 18:16
  • looks to me the OP's formula forgot to take something into account: the constant cost of each call, so the cost should be `T(n)=1+T(n-3)+T(n/2)`. – Willem Van Onsem Sep 17 '15 at 18:26
  • if I add a log factor, I approximately get the same plot. If you plot it in logscale, you see a small divergence of `r^n` which means on the (very long term) there is devergence. I indeed added a factor `n` that was not allowed. The bound is probably `r^n*log(n)` – Willem Van Onsem Sep 17 '15 at 19:07
  • @YvesDaust: well I've provided the table in my answer, and plotted using a logscale, which is more safe when it comes down to rounding errors. The plot you show looks like the shape of a normal `log x` curve. Can you provide your algorithm to calculate the number of steps? – Willem Van Onsem Sep 17 '15 at 20:22
  • Well first of all, I think it is clear that the time complexity is `O(r^n*log n)` at least (`O` is an upperbound). Furthermore I agree `r^n` is a dominant factor, but the question is whether it is the only serious factor. As said before you can visualize it as a first layer n, n-3, n-6,... But there is a layer below it n/2-3,(n-3)/2,(n-6)/2. Although it doesn't contribute "much", it contributes something, and it contributes with a same complexity (but way lower offset value). I think that Chris Beck gives the sharpest bound with a mathematical proof. – Willem Van Onsem Sep 17 '15 at 20:41