(06/14/2022, 06:02 PM)tommy1729 Wrote:(06/14/2022, 01:22 AM)JmsNxn Wrote:(06/13/2022, 10:52 PM)tommy1729 Wrote: i advise to use this for the computation around the fixpoint
(a x + k x^2 + ...)^[t] = a^t x + k a^(t-1) ( a^t - 1 )/(a-1) x^2 + ...
This quadration approximation is then plugged in
f^[t](x) = lim f^[n] ( a^t y + k a^(t-1) ( a^t - 1 )/(a-1) y^2 )
where y = f^[-n](x) and a and b can be easily computed from the closed form for x_n and taylors theorem.
Call it the quadratic fixpoint formula or so![]()
regards
tommy1729
Hey, Tommy. Not to let you down or anything, but this is a known procedure. It can be done upto any polynomial, not just quadratic.
It essentially operators off the principle of instead of choosing a fixed point, we choose a fixed polynomial. So we know the first \(3\) coefficients of the Taylor polynomial near the fixed point. Iterate a slightly different Kneser formula, but do it for a fixed point which is polynomial. You'll get a far faster solution. Same Kneser algorithm, you just do it for polynomials.
Hahaha ofcourse I am aware this fixpoint method is nothing new.
Im aware of it for over 30 years , it has been talked about here for over 13 years.
Gottfried mentioned it here and on his site.
Carleman matrices , bell matrix and sometimes vandermonde matrices are related.
It has been used for over 2000 years.
It occurs in many many books , webpages , wiki and handwritten notes since the 1930's.
There are even some trends and theorems about them for the higher degrees.
I probably wrote about it and related things before.
Now I do admit that I am not expert at it though.
I do not know degree 5 by heart nor the pattern and I have questions and conjectures.
Im willing to talk about it and I would love references.
Im currently talking about it in another thread ; comparing the methods , their speed accelerating methods and imo most important : if and when they converge to the same function.
But the main point here was to mention that I advise a faster method than schröder/kneser/koenings.
( I consider koenigs the correct term : kneser used a modified version of koenigs , whereas schröder is the equation it satisfies but not the method itself.
one can argue about it ... Btw Im also aware that sometimes it is better to transform the schröder equation to the Julia equation or the böttcher equation or others. Im also aware of what leo wrote. I wrote some computer codes to experiment. And I experimented with matrix splitting to accelerate the methods. There are connections to number theory too. Im not a total beginner![]()
)
I conjecture that using taylor polynomials as approximation is optimal on average ( for random functions ) for such formulas.
I apprecciate your comments though.
It is hard to smell what people know or do not , especially when you cant smell them.
***
It is not entirely clear to me what fixpoint method makes the idea work and produces analytic tetration.
But I think this " quadratic fixpoint formula " produces a working one.
I was not really claiming it as my own formula by giving it a name.
It is not " tommy's quadratic fixpoint formula " or so.
It is just that I am unaware of its real name ...
If it even has a real name.
"truncated polynomial method of degree 2" or "polynomial iteration mod x^3" do not sound so nice imo.
regards
tommy1729
Honestly, I don't know what it's called either. It's the central topic of Kouznetsov's method though. At least, to a degree.
We guess the solution to:
\[
\Psi^{-1}(\lambda z) = f(\Psi^{-1}(z)) + O(z^k)\\
\]
So that:
\[
\Psi^{-1}(z) = \sum_{j=0}^k a_j z^k + O(z^k)\\
\]
For some fixed \(k\). You can choose the Kneser solution to this, or the Koenigs solution, or what ever (you're right, I should be clearer about this). Then you iterate:
\[
f^{\circ n}(\Psi^{-1}(\lambda^{s-n}))
\]
Which will converge to a super function \(F(s)\) such that \(f(F(s)) = F(s+1)\). You can create Kneser, or you can create Koenigs (Kouznetsov and Bo call this regular iteration--I think that should be the standard), or you can create whatever (you can do this procedure for \(\beta\) too). The larger you let \(k\) be in the original solution, determines how fast the algorithm will converge. With much of this being justified by choosing \(k\) very large (at least as Kouznetsov talking about it).
You're doing something similar, but a little different. As you aren't looking for the same kind of thing per se. There are theories of complex dynamics which talk about this, I believe Milnor has a brief moment where he talks about it in his graduate text book; I believe it's only briefly though as a "speed up the calculations" observation. I'll double check. This is mostly a Kouznetsov thing, but there are dynamics results (Milnor uses it for Abel functions if I'm not mistaken...).
What you've done is a little different, you've written the first part of the Fourier series:
\[
\Psi^{-1}(\lambda^s\Psi(z)) = a_0 + a_1\lambda^s + a_2 \lambda^{2s} + O(\lambda^{3s})\\
\]
And gone from there, in your super function formula. You can do this arbitrarily though. I'll look for the part in Milnor where he describes speed ups in computation time (it might be the appendix), or I'm outta luck and I don't remember where I read this, lol. Milnor's so much my bible I forget other references, lol.

