• 3 Vote(s) - 4.33 Average
• 1
• 2
• 3
• 4
• 5
 Searching for an asymptotic to exp[0.5] tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 07/27/2014, 05:01 PM About that perpendicular thing... I think I was able to show that the first occurance - between the first 2 zero's starting from 0 going to -oo - is true if all zero's are on the negative real line. Basicly it uses hadamard and says that adding a linear term (by product) does not change the property. I started with a polynomial of degree 2 and then used induction. Probably this can be extended to all occurances. regards tommy1729 tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 07/28/2014, 12:17 PM Let f(z) be the entire fake half-iterate of exp(z). Now we do not have a sexp and hence also no theta(z) function. However at least at the positive real line there must other entire functions that satisfy the functional equation very well. Let u(z),v(z),w(z) give these other solutions f_i(z) : f_1(z) = f(z+u(z)) = f(z)(1+v(z)) = f(z) + w(z). (Here I assumed u,v,w are Always related , that is another subject) Now it seems logical to assume u,v,w also grow much slower then exp , likely like exp^[1/2]. This gives then a similar hadamard product for these functions. Let m be an integer > -1. Now U := u,v,w must satisfy U(f^[m](0)) = 0. Which btw is consistant with carlson and hadamard. So U is completely determined by the other zero's apart from U(f^[m](0)) = 0 (and a constant C of course). And that determines how well the functional equation is satisfied off the real line !! It would be amazing if one of u,v,w is somehow optimal WITHOUT other zero's then U(f^[m](0)) = 0. Of particular intrest is the cases where u,v,w have no zero's in the region where our Original f(z) satisfies the functional equation well. Because if it has a zero there (point "Q") then it must also have a zero at point f(Q) which complicates matters. regards tommy1729 tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 07/28/2014, 09:52 PM (This post was last modified: 07/28/2014, 09:54 PM by tommy1729.) (06/28/2014, 11:16 PM)tommy1729 Wrote: I wanted to remark that the "fake function" idea is generalizable. We can find any entire asymptotic with nonnegative derivatives by simply using a rising integer function T(n) : etc. This simple idea / equation is very powerfull. The fundamental theorem of fake function theory. In combination with the post about the inverse gamma function we could for instance find the asymptotic : with the method above. To stay in the spirit of the exp. regards tommy1729 To combine 2 ideas of me we can use this to find an alternative asymptotic half-iterate F(z) of exp(z) such that F(z) = F(-z). Or even the additional F^(z) = F^(-z) if we take F(0) = 0. We do that by defining F(z) = a0 + a1 x + a3 x^3 + a5 x^5 + ... In other words we take T(n) = 2n + 1 and then solve like described above. The main difference with 2sinh^[0.5](z) is that we get a different region of almost agreement with exp(z) or even 2sinh(z) after 2 iterations. Also , though related , notice the derivatives of 2sinh^[0.5](z) are not all positive. Btw we should also get the similar hadamard product for this F(z) as well as for 2sinh^[0.5]. But how do the plots look like ? And where are the zero's ? This requires some additional work. An intresting lemma for the 2sinh if true would be this : A := 2 pi 2sinh^[0.5](z) = a0 + a1 z + a2 z^2 + ... g(z) := |a0| + |a1| z + |a2| z^2 + ... Conjecture : |g(z)| < 1 + A 2sinh^[0.5](|z|)^A Stronger and weaker variants are also of intrest. regards tommy1729 tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 07/28/2014, 10:30 PM (07/28/2014, 12:17 PM)tommy1729 Wrote: Let f(z) be the entire fake half-iterate of exp(z). Now we do not have a sexp and hence also no theta(z) function. However at least at the positive real line there must other entire functions that satisfy the functional equation very well. Let u(z),v(z),w(z) give these other solutions f_i(z) : f_1(z) = f(z+u(z)) = f(z)(1+v(z)) = f(z) + w(z). (Here I assumed u,v,w are Always related , that is another subject) Now it seems logical to assume u,v,w also grow much slower then exp , likely like exp^[1/2]. This gives then a similar hadamard product for these functions. Let m be an integer > -1. Now U := u,v,w must satisfy U(f^[m](0)) = 0. Which btw is consistant with carlson and hadamard. So U is completely determined by the other zero's apart from U(f^[m](0)) = 0 (and a constant C of course). And that determines how well the functional equation is satisfied off the real line !! It would be amazing if one of u,v,w is somehow optimal WITHOUT other zero's then U(f^[m](0)) = 0. Of particular intrest is the cases where u,v,w have no zero's in the region where our Original f(z) satisfies the functional equation well. Because if it has a zero there (point "Q") then it must also have a zero at point f(Q) which complicates matters. regards tommy1729 It turns out 1 + v(z) cannot both be entire and grow slower then O(exp(|z|/|a|) and still satisfy the functional equation well. Hence there is no (nice) v(z) !! I can prove that if you want, its trivial. It follows from considering the zero's of both 1+v(z) and f(z). hence the attention goes to f_1(z) = f(z+u(z)). It seems u(z) changes the function in the sense that the region where f_1^(z) - exp(z) << 1 is different from f(f(z)) - exp(z) << 1. This can be explained by induction starting from the real line. However that does assume the conjectures made about f(z). It seems u(z) makes a " wobble " just like we saw with the theta functions. Im a bit puzzled by u(z). If u(z) has so many zero's this suggests (carlson , hadamard) u(z) grows at least like exp^[0.5](|z|/|a|) for z with negative real part. But then |f(z+u(z))| grows like O(exp(|z|/|a|)) ? This implies that f(f(z)) grows like exp(z) near the positive real line and like exp(exp(z)) near (but not on) the negative real line for sufficiently large z ?!? That would be one weird function. Like the half-iterate of both exp(z) and exp(exp(z)). Thus the half-iterate of a twice fake/asymptotic function itself. Fake function theory seems like a neverending story. regards tommy1729 sheldonison Long Time Fellow    Posts: 683 Threads: 24 Joined: Oct 2008 07/30/2014, 04:07 PM (This post was last modified: 08/02/2014, 02:50 PM by sheldonison.) (07/28/2014, 12:17 PM)tommy1729 Wrote: Let f(z) be the entire fake half-iterate of exp(z). Now we do not have a sexp and hence also no theta(z) function. However at least at the positive real line there must other entire functions that satisfy the functional equation very well. .... It seems u(z) makes a " wobble " just like we saw with the theta functions.We could also look at Still on vacation; enjoyed reading your posts. I generated a quick graph of . The first zero crossing is at z=137.8052295808. For smaller value of z, the error term is dominated by the 1/z laurent term I mentioned earlier; for larger values of z, it is dominated by Here is the graph. The logarithmic spikes are the zero crossings of f(f(z))-exp(z). Notice how it wobbles around exp(z). The error term is dominated by the error term for f(z), which I graphed in an earlier post. f(z) has nearly the same first error term. f(f(z))/exp(z) has a larger error than f(z)/exp^0.5(z), but they similarly. The error term for f(f(z))/exp(z) at z=1000 is f(f(z))/exp(z)-1=-1.138369235934 E-12. - Sheldon tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 08/01/2014, 11:20 PM Let f(x) = a_0 + a_1 x + a_2 x^2 + ... Let a_0 , a_1 , a_2 > 0 Assume a_2 > a_3 , a_3 > a_4 , ... > 0 Then the goal is to find a_n for n > 2. Assume a_n > (n-1) a_(n+1) Some motivation. First we try to solve for a single variable. Tetration is a difficult subject so multivariable ideas/equations are complicated in particular when we are not working with matrices. Another thing. We prefer not to solve for an equation that contains both a_(n-1) and a_n. The reason is that we get disagreement. example : F(a_10,a_11) = 0 F(a_11,a12) = 0 Now we have 2 conflicting values for a_11. Besides solving for a_(n-1) when we already had its value is unlogical. Solving for a_n and a_(n+1) makes a bit more sense since we did not yet have the value of a_(n+1). example : given a_5 it makes sense to solve for a_6,a_7. Keeping that in the back of our mind , what type of equation should we solve ? Logical would be a truncated taylor series. Sheldon approximately solved a_n x^n = exp^[0.5](x). However the truncation is extreme here. It is true that for every x , there must be an n such that a_n x^n is the most dominant term. But a_n is independant of the value of x. Hence the equation a_n x^n = exp^[0.5](x) makes some sense. However a more dominant approximation of a taylor series is a_q x^q , where q is between n and n+1 and a_q = a_n. This shows that sheldons equation is probably valid within a ratio of x^(q-n). On average that is x^(1/2). So sheldon's solution S(x) probably satisfies S(x)/x^(3/4) - C < f(x) < S(x) x^(3/4) + C for some constant C. In fact sheldon mentioned the correcting factor x^(1/2) himself. A less extreme truncation would lead to better results. How to get more dominant terms ? We need to consider the contributions of a_m x^m for a general a_m from a random taylor series with positive a_i. The contributions with respect to m look like a gaussian curve g where g(m - t) is smaller than g(q) and g(m + t2) is smaller than g(q) for suff large t and assuming m = q + o(2). the top of this curve at g(m) has growing m where m grows with x. but a gauss curve is symmetric. so a_n x^n + a_(n+1) x^(n+1) are the most dominant terms and both terms can have similar contribution ! Now perhaps you are thinking : why not solve a_n x^n + a_(n+1) x^(n+1) + a_(n+2) x^(n+2) + a_(n+3) x^(n+3) ? Its a reasonable idea ... but It violates all logic above. First there are too many variables. Second we get a lot disagreement. But most importantly its no longer likely that a_n x^n is the dominant term !! Its quite more likely to be a_(n+1) x^(n+1) + a_(n+2) x^(n+2) which then is basicly the same as what is proposed now. Notice that if a_n x^n is not one of the dominant terms , its more like the term a_(n-1) x^(n-1). But then we violate another logic of above ; we solve for values we already have. Notice the assumption "Assume a_n > (n-1) a_(n+1)" only makes this argument stronger. These consideration lead me to a_n x^n + a_(n+1) x^(n+1) = exp^[0.5](x). 2 unknowns is a bit problematic and I do not want to get disagreement. So we use the assumption "Assume a_n > (n-1) a_(n+1)". Notice this assumption makes sense since f grows more like a polynomial then an exponential. The assumption is also consistant with sheldon's equation , plots and all results sofar. a_n x^n + a_n/(n-1) x^(n+1) = exp^[0.5](x) Now we need approximations for exp^[0.5](x) without having our f(x). I use my 2sinh method for that when x is large. We continue : a_n x^n (1 + 1/(n-1) x) = exp^[0.5](x) exp(X) = x , take ln on both sides : ln(a_n) + n ln(x) + ln(1 + 1/(n-1) x) = ln(exp^[0.5](x)) replace x with exp(X) Now the strenght of my 2sinh shows : ln(exp^[0.5](exp(X))) = exp^[0.5](X) we get : ln(a_n) + n X + ln(1 + 1/(n-1) exp(X)) = exp^[0.5](X) We know from that ln(1 + "Large") is about ln("large") + 1/"large". SO we can further reduce : ln(a_n) + n X + ln(1/(n-1) exp(X)) + (n-1)/exp(X) = exp^[0.5](X) Slightly less formal but it seems the term (n-1)/exp(X) is so small we can neglet it and remove it. we then get : ln(a_n) + n X + ln(1/(n-1) exp(X)) = exp^[0.5](X) Simplify ln(a_n) + n X - ln(n-1) + X = exp^[0.5](X) Simplify even Further : ln(a_n) + (n+1) X - ln(n-1) = exp^[0.5](X) ln(a_n) = Max [ exp^[0.5](X) - (n+1) X + ln(n-1) ] = Max [exp^[0.5](X) - (n+1) X] + ln(n-1) exp^[0.5](X) - (n+1) X has one minimum value, where the derivative is zero. At the minimum, the derivative will be equal 0, so defining Now define And then finally the improved solution is : I considered further improvements but things got dubious and I stumbled upon problems with the principles used here. In other words be careful. regards tommy1729 " truth is that what does not go away when you stop believing in it " tommy1729 tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 08/01/2014, 11:36 PM Some further thought : There is good news and bad news. Basicly both are that this is in a way very similar to the Original estimate for a_n combined with our estimate for a_(n+1). Just as our estimate for a_(n+1) from a_n and the Original estimate of a_n. But maybe a refinement of the previous post helps. Am I wrong in assuming we prefer to solve for one variable ? regards tommy1729 tommy1729 Ultimate Fellow     Posts: 1,491 Threads: 355 Joined: Feb 2009 08/02/2014, 12:26 PM (This post was last modified: 08/02/2014, 11:02 PM by tommy1729.) EDIT (removed nonsense) sheldonison Long Time Fellow    Posts: 683 Threads: 24 Joined: Oct 2008 08/02/2014, 03:44 PM (08/02/2014, 12:26 PM)tommy1729 Wrote: Some further thought : Hey Tommy, I'm not able to understand the part of your equation; these are big numbers. For n=50, , perhaps you intended a ln somewhere? - Sheldon sheldonison Long Time Fellow    Posts: 683 Threads: 24 Joined: Oct 2008 08/02/2014, 04:34 PM (This post was last modified: 08/10/2014, 05:35 AM by sheldonison.) (07/17/2014, 05:46 AM)sheldonison Wrote: ...to understand the negative real axis, you mostly have to understand how Kneser's half iterate behaves for positive , due to the equation . And the definition I'm using for negative axis comes down to plus stuff, where stuff can be shown too small to create zeros anywhere else, since halfk(z) can be shown to have a large absolute value at the negative axis, so stuff winds up acting like a small perturbation of the zeros of the approximation. Before we can understand "stuff", we need an exact converging equation for a_n. Here is the equation, as a converging integral. Of course, using h_n is a pretty good approximation, but the limit converges, and gets more accurate as y goes to infinity. First, we need an equation for the trough of , at a particular real value of y, where we want the local minimum of , which is where the derivative has a zero crossing. We note that z=0 has a derivative of zero, since the function has a local maximum at zi=0, but we are interested in the local minimum, where the real part at the minimum gets arbitrarily large negative as y gets bigger. Then we get this equation for all of the Taylor series coefficients of the asymptotic half iterate, f(z). Of course, you can use y=h(n), for our previous "best behaved" radius, but the conjecture is that this integral converges as y gets arbitrarily large, for all values of a_n. This conjecture only depends on the first real minimum getting arbitrarily negative for exp^{0.5}. In this equation, exp(y) is the radius of a circle, where we are wrapping the approximation around the circle multiple times until we get to the local minimum. Each z=pi i corresponds to half way around the circle. For example, for y=2, we get mi(2)~=5.65pi, so we wrap the approximation around the circle 5.65 times, for a radius of exp(2). This limit converges rapidly as y increases. A more difficult conjecture is that this equation also converges for all negative values of n, where we are talking about a Laurent Series! Then, here are the golden values of a_n, for the asymptotic half iterate, accurate to 32 decimal digits. Code:{half= +x^ 0*  0.49910040768286480611449256148664 +x^ 1*  0.86975393764881550610513018981220 +x^ 2*  0.25464139386683457183123452615036 +x^ 3*  0.020389794621141210309883766457462 +x^ 4*  0.00057355319935562030963666319208274 +x^ 5*  0.0000066349480283811388551798440051945 +x^ 6*  0.000000035236417782822470231171202844523 +x^ 7*  9.3249390474194798654833146618092 E-11 +x^ 8*  1.3108275783181088058101316776999 E-13 +x^ 9*  1.0304474531377645338109992499820 E-16 +x^10*  4.7263089051086447203604773624283 E-20 +x^11*  1.3108531680479134461037327583388 E-23 +x^12*  2.2667203510166636785266041654942 E-27 +x^13*  2.5093571724905769180405008212467 E-31 +x^14*  1.8202710730641490150779164265302 E-35 +x^15*  8.8317492249695809130512115338466 E-40 +x^16*  2.9191797760804480951691682704407 E-44 +x^17*  6.6824837396251043572898322381552 E-49 +x^18*  1.0753580053048766166021599140359 E-53 +x^19*  1.2330918481909207056741377403551 E-58 +x^20*  1.0201122384241599346658487339891 E-63 +x^21*  6.1581919318876799901812742006051 E-69 +x^22*  2.7413713158485570640687894815224 E-74 +x^23*  9.0867443007270293626597871983267 E-80 +x^24*  2.2630247565032422317752797826787 E-85 +x^25*  4.2702840685672683263508716228771 E-91 +x^26*  6.1534101486513307395669911447273 E-97 +x^27*  6.8211133138591787310720468849488 E-103 +x^28*  5.8569044461458402614113625767890 E-109 +x^29*  3.9207619171672337067631917179388 E-115 +x^30*  2.0588139351090054627212193973763 E-121 +x^31*  8.5293478716542300913961567853655 E-128 +x^32*  2.8031084652683108006729995479402 E-134 +x^33*  7.3457996813337157532882820842756 E-141 +x^34*  1.5425863530541989149502646319825 E-147 +x^35*  2.6079689469031728136271774809838 E-154 +x^36*  3.5655969942648761764867489799646 E-161 +x^37*  3.9589823935595558073401715169891 E-168 +x^38*  3.5844087663000984825476791486429 E-175 +x^39*  2.6565482453014575880214435139086 E-182 +x^40*  1.6176901364913946434067300306779 E-189 +x^41*  8.1225821093245138453619023866767 E-197 +x^42*  3.3743842159494443045870532508263 E-204 +x^43*  1.1636380322113548204121841687534 E-211 +x^44*  3.3414117791193887342805961804906 E-219 +x^45*  8.0139082998656669432986590820899 E-227 +x^46*  1.6099984482286816262190253292747 E-234 +x^47*  2.7170111673887874757432802681464 E-242 +x^48*  3.8620399030012523700899447882433 E-250 +x^49*  4.6359068094239026501210448622201 E-258 +x^50*  4.7112860564671104284049070896713 E-266 +x^51*  4.0633888173969156147865799787716 E-274 +x^52*  2.9812864294138151331792841161239 E-282 +x^53*  1.8649814247555529976292459798576 E-290 +x^54*  9.9691545057158256900573917783738 E-299 +x^55*  4.5633385134782655342243796935439 E-307 +x^56*  1.7924480904968774321789395462320 E-315 +x^57*  6.0537133204460641567051184734009 E-324 +x^58*  1.7613854006382418674405144204076 E-332 +x^59*  4.4234937181847438236891275441742 E-341 +x^60*  9.6062199349333910553455757129229 E-350 +x^61*  1.8071357893977747411903461224328 E-358 +x^62*  2.9500793906119261031399424376955 E-367 +x^63*  4.1861438832113176863676861086587 E-376 +x^64*  5.1718420569189564848260590384058 E-385 +x^65*  5.5721295365630149500802054207483 E-394 +x^66*  5.2434610761864181816794420120788 E-403 +x^67*  4.3161290218455616775134218499144 E-412 +x^68*  3.1123772204391816644227651561834 E-421 +x^69*  1.9689619829065472069738970737448 E-430 +x^70*  1.0943094356648526668959548325101 E-439 +x^71*  5.3505230418606478400533612721597 E-449 +x^72*  2.3045540605721857919532251055566 E-458 +x^73*  8.7554739474750859144959229047230 E-468 +x^74*  2.9378444676888041787778161414704 E-477 +x^75*  8.7171763791650030139352986241477 E-487 +x^76*  2.2900763435907436360898905965880 E-496 +x^77*  5.3329621850699654818529586155360 E-506 +x^78*  1.1021414173779794198692060389354 E-515 +x^79*  2.0237236457493364339472937613359 E-525 +x^80*  3.3051734010090040785253146736889 E-535 +x^81*  4.8066246879974669772188503242664 E-545 +x^82*  6.2309299154003703910378261449910 E-555 +x^83*  7.2075276038704781939865489059551 E-565 +x^84*  7.4470678249259203410669072069184 E-575 +x^85*  6.8799542618840704189707691448556 E-585 +x^86*  5.6887123387554212717589245890207 E-595 +x^87*  4.2139464984155225143507152525240 E-605 +x^88*  2.7991159853357840719968584351765 E-615 +x^89*  1.6688289978569269076157915195546 E-625 +x^90*  8.9383291288560593726126574174143 E-636 +x^91*  4.3046852131660140981930943841115 E-646 +x^92*  1.8657210514708086914428713376622 E-656 +x^93*  7.2835803758857835339891018806736 E-667 +x^94*  2.5633138429804986238168248757730 E-677 +x^95*  8.1391004442741453908485523187915 E-688 +x^96*  2.3335731071310654588671160001735 E-698 +x^97*  6.0462083925845218287779795765013 E-709 +x^98*  1.4167776781140060144737649766706 E-719 +x^99*  3.0047713766707612115082094030781 E-730 +x^100*  5.7721916490766616272646194308725 E-741 +x^101*  1.0051073329533954091297617444478 E-751 +x^102*  1.5876101051655864714914627836334 E-762 +x^103*  2.2763915385275936321978344407162 E-773 +x^104*  2.9650262675741677534398513006778 E-784 +x^105*  3.5106724774538938342869771007255 E-795 +x^106*  3.7811947499409106123326108806872 E-806 +x^107*  3.7071205125522960714497492395807 E-817 +x^108*  3.3105531419378447681115579471314 E-828 +x^109*  2.6946559659038018653919216302630 E-839 +x^110*  2.0004294153855532650882158071798 E-850 +x^111*  1.3552960584934226299262921449794 E-861 +x^112*  8.3850528276019362718622289505176 E-873 +x^113*  4.7402759150516390261292610270314 E-884 +x^114*  2.4501297992513134302225437677246 E-895 +x^115*  1.1585632067647150421762060606285 E-906 +x^116*  5.0147490158043786508617239112848 E-918 +x^117*  1.9880504333517064850201130747949 E-929 +x^118*  7.2227159872668272720579061678065 E-941 +x^119*  2.4060865627030902597898106282254 E-952 +x^120*  7.3535769299459848259437276983085 E-964 +x^121*  2.0629960448663672546484635290444 E-975 +x^122*  5.3154842684165058726225254564594 E-987 +x^123*  1.2585225678662369543828624715412 E-998 +x^124*  2.7395468984989897619259774129850 E-1010 +x^125*  5.4855316531810155402528303195600 E-1022 +x^126*  1.0108822861008503641480002279807 E-1033 +x^127*  1.7153033576049973603789610195290 E-1045 +x^128*  2.6813496681095861799615193676774 E-1057 +x^129*  3.8632154203187923022475569144109 E-1069 +x^130*  5.1325753206376595214835382059548 E-1081 +x^131*  6.2909739014518264994302900317002 E-1093 +x^132*  7.1170348588447107922531829150789 E-1105 +x^133*  7.4349631869133816166995706604176 E-1117 +x^134*  7.1755358837079211367013832952296 E-1129 +x^135*  6.4005953851428688022188788458237 E-1141 +x^136*  5.2792126845518191923562741833628 E-1153 +x^137*  4.0280095342680852744276615532834 E-1165 +x^138*  2.8442762113656478776744296058908 E-1177 +x^139*  1.8595080050394638031227938527562 E-1189 +x^140*  1.1260355336262741141770597741831 E-1201 +x^141*  6.3185059478023237972227228368956 E-1214 +x^142*  3.2867241006516489513816376628370 E-1226 +x^143*  1.5855311005276654831679157820993 E-1238 +x^144*  7.0961524098975563551156187520556 E-1251 +x^145*  2.9476754003214298928109754468151 E-1263 +x^146*  1.1368804075315723312006909805048 E-1275 +x^147*  4.0728287290475756342684260134312 E-1288 +x^148*  1.3557812269617278007750251760013 E-1300 +x^149*  4.1952622316541024998769088280678 E-1313 +x^150*  1.2071650609336455308114509698551 E-1325 +x^151*  3.2312637072119308611078701726049 E-1338 +x^152*  8.0488654207294442338748649974483 E-1351 +x^153*  1.8664224982307982923337088202746 E-1363 +x^154*  4.0304452980864692588143080374902 E-1376 +x^155*  8.1080586405208784518167888656717 E-1389 +x^156*  1.5200322609208781016863800559477 E-1401 +x^157*  2.6565036706891443830215617307060 E-1414 +x^158*  4.3294994751289815651245367744333 E-1427 +x^159*  6.5823498635163449848555652379754 E-1440 +x^160*  9.3386689905531298836407032081739 E-1453 +x^161*  1.2367813668596923892976787072383 E-1465 +x^162*  1.5294893531302097949642806699239 E-1478 +x^163*  1.7667930505940060274348065390183 E-1491 +x^164*  1.9069936328780065176399881965592 E-1504 +x^165*  1.9238637515150748979094192059181 E-1517 +x^166*  1.8146656564578863735093597982736 E-1530 +x^167*  1.6008494118681160348902570449217 E-1543 +x^168*  1.3212007554424453549268340674563 E-1556 +x^169*  1.0204298754863582893450368551979 E-1569 +x^170*  7.3777465851474704224470487972849 E-1583 +x^171*  4.9948147871133536465003361490018 E-1596 +x^172*  3.1673619310071790051999743120795 E-1609 +x^173*  1.8818496436539240454942847710810 E-1622 +x^174*  1.0478668087781472600990531983076 E-1635 +x^175*  5.4699700108119798202092096812686 E-1649 +x^176*  2.6775897724524888137686433054764 E-1662 +x^177*  1.2294325237690728809655261657373 E-1675 +x^178*  5.2964765773031169879511934516894 E-1689 +x^179*  2.1414605685633153916863160820213 E-1702 +x^180*  8.1281477458001652530909603942617 E-1716 +x^181*  2.8969983969634489616575607196993 E-1729 +x^182*  9.6983005132099758882833894300444 E-1743 +x^183*  3.0503395905523788433339480299771 E-1756 +x^184*  9.0161117860837331206442146019198 E-1770 +x^185*  2.5050778397625380138445911053287 E-1783 +x^186*  6.5443308835195965498981496832145 E-1797 +x^187*  1.6079093717449488951345308353968 E-1810 +x^188*  3.7163735418580107519205094520222 E-1824 +x^189*  8.0825202631453331646068714950141 E-1838 +x^190*  1.6544401976104632103954118990163 E-1851 +x^191*  3.1881472586007265091828313250370 E-1865 +x^192*  5.7851408010756338004215092666134 E-1879 +x^193*  9.8873984333290892254958169505146 E-1893 +x^194*  1.5920108376659927332139990936381 E-1906 +x^195*  2.4155070121303973501240620008794 E-1920 +x^196*  3.4543821888983617522036269239819 E-1934 +x^197*  4.6572856036590416613802548902271 E-1948 +x^198*  5.9210056278933626173684610772389 E-1962 +x^199*  7.0999784656609903423552501504057 E-1976 +x^200*  8.0318487044128735388355344796493 E-1990 +x^201*  8.5736955841296876320106213166851 E-2004 +x^202*  8.6379549570131275669935590584612 E-2018 +x^203*  8.2156059203101344472355111073653 E-2032 +x^204*  7.3781705037202575754015115479386 E-2046 +x^205*  6.2579491145231710891090739884756 E-2060 +x^206*  5.0139806297441713625995837747212 E-2074 +x^207*  3.7957080997245923216418974329360 E-2088 +x^208*  2.7155259821241406171698132630248 E-2102 +x^209*  1.8363557738700768802451438375387 E-2116 +x^210*  1.1740628433421057821777973667206 E-2130 +x^211*  7.0981753502258051460817318245709 E-2145 +x^212*  4.0589288030084695873883409882588 E-2159 +x^213*  2.1957010780195810255236500187247 E-2173 +x^214*  1.1238778023538369869135766947955 E-2187 +x^215*  5.4442164801564404583051655574522 E-2202 +x^216*  2.4963628438916078728829893192368 E-2216 +x^217*  1.0837297442120598301618243963698 E-2230 +x^218*  4.4551207761364497154432210233209 E-2245 +x^219*  1.7346295851977723205655565316687 E-2259 +x^220*  6.3980216198329557226289450539372 E-2274 +x^221*  2.2359320230049242230716901184707 E-2288 +x^222*  7.4050335548082678763932071864420 E-2303 +x^223*  2.3245129551422610442911634354616 E-2317 +x^224*  6.9175729386729670176686439598172 E-2332 +x^225*  1.9519651728833711120704211891455 E-2346 +x^226*  5.2235530816524647507740249507377 E-2361 +x^227*  1.3259100688231246355692990305644 E-2375 +x^228*  3.1929646432844662917189012467867 E-2390 +x^229*  7.2959763189818210857434787158098 E-2405 +x^230*  1.5821901620262449059035467802555 E-2419 +x^231*  3.2568369220809269198845465533853 E-2434 +x^232*  6.3645976185495999608136486932789 E-2449 +x^233*  1.1810233011410470649981718073886 E-2463 +x^234*  2.0812917545843743180001137160933 E-2478 +x^235*  3.4839209979860253232978091890307 E-2493 +x^236*  5.5403573289249218512210912747724 E-2508 +x^237*  8.3717096964008922909601293473603 E-2523 +x^238*  1.2021805014275332256273568079079 E-2537 +x^239*  1.6408770246927264994411638788683 E-2552 +x^240*  2.1291408832039388518986897463955 E-2567 +x^241*  2.6267904803358885329607263614568 E-2582 +x^242*  3.0818343021308110175424084637158 E-2597 +x^243*  3.4389478802917417932763035247815 E-2612 +x^244*  3.6504256086056137433999815919077 E-2627 +x^245*  3.6866478673419651585962740782545 E-2642 +x^246*  3.5428944788828265970875357611476 E-2657 +x^247*  3.2403420506168274635327264144756 E-2672 +x^248*  2.8209589845279863091804917029551 E-2687 +x^249*  2.3379901981734771425383333003521 E-2702 +x^250*  1.8449935643976129364571679094859 E-2717 +x^251*  1.3864974449557630194049144333358 E-2732 +x^252*  9.9238610700387272943146035644498 E-2748 +x^253*  6.7661955468905308794404747181230 E-2763 +x^254*  4.3951643231706880662487885526392 E-2778 +x^255*  2.7204226590007386231018685797130 E-2793 +x^256*  1.6046927829826297494599805972699 E-2808 +x^257*  9.0220407668143541120228836622922 E-2824 +x^258*  4.8354593048047845033224625566813 E-2839 +x^259*  2.4708877100271962825258221906857 E-2854 +x^260*  1.2039611209329912178069759757049 E-2869 +x^261*  5.5947092908616561879079231541392 E-2885 +x^262*  2.4797575257941287992271451574638 E-2900 +x^263*  1.0484993197517629783940986516453 E-2915 +x^264*  4.2297484953859892645978946228906 E-2931 +x^265*  1.6282017033437962103813958871737 E-2946 +x^266*  5.9814785846579457658580487678486 E-2962 +x^267*  2.0973678925508310140708454839134 E-2977 +x^268*  7.0204513860124048128929361569859 E-2993 +x^269*  2.2435619969851113721406057440386 E-3008 +x^270*  6.8462320075794491042780947822524 E-3024 +x^271*  1.9950930497586676299541608886840 E-3039 +x^272*  5.5530277032422158108999083214016 E-3055 +x^273*  1.4764149510058814271782272044002 E-3070 +x^274*  3.7502078958721485969741426534598 E-3086 +x^275*  9.1017858622372148040390949833395 E-3102 +x^276*  2.1109499323164391347227188665436 E-3117 +x^277*  4.6791248960258165813941772872856 E-3133 +x^278*  9.9138356530228591442513442058349 E-3149 +x^279*  2.0080004529199037968113698528820 E-3164 +x^280*  3.8885278195209745127169291405925 E-3180 +x^281*  7.2004541037708460787761986112735 E-3196 +x^282*  1.2750918172412703104661669248805 E-3211 +x^283*  2.1596492273536361542217699775661 E-3227 +x^284*  3.4989531476600975459124994446056 E-3243 +x^285*  5.4232399623854837069841157428345 E-3259 +x^286*  8.0426211741222782354153066476251 E-3275 +x^287*  1.1413175498182740675299773645312 E-3290 +x^288*  1.5500194377377993981378886128387 E-3306 +x^289*  2.0148416599353000816235310923698 E-3322 +x^290*  2.5070839651532656491438274506800 E-3338 +x^291*  2.9865635421626323775665331571620 E-3354 +x^292*  3.4064341659042264496424959171075 E-3370 +x^293*  3.7205208039357170462514006673813 E-3386 +x^294*  3.8916413340839957809707676312809 E-3402 +x^295*  3.8988520048176433637470666861845 E-3418 +x^296*  3.7416642242029463512766194970626 E-3434 +x^297*  3.4400594322819498160219151491178 E-3450 +x^298*  3.0303222824401450669357416642614 E-3466 +x^299*  2.5578896958096305057557562347395 E-3482 +x^300*  2.0691549879104937795561894240776 E-3498 } I mentioned "stuff" in a previous post. The 1/x Laurent series is an excellent approximation for -stuff, where here we use a 5 term approximation. If the |z| is fairly large, then we might want to use more terms in the approximation, there is always some stuff left over. And here are the first 20 terms of the Laurent series, also generated and printed to 32 decimal digits accuracy. Generating the Laurent series is much much more difficult computationally, since there is no optimal h_n value for negative values of n, so extremely high precision calculations are required for the Kneser half iterate. I call this function the half_error term, since it is the dominant term for stuff, for small radius's. By the way, I also conjecture that even though all of these Laurent series terms converge, they don't behave very well as n gets arbitrarily large negative, and the function itself probably doesn't converge anywhere. But the truncated Laurent (1/z) series is very useful for understanding the "stuff" terms of the half iterate. A 20 term truncated Laurent series seems to work pretty well. "Stuff" in the equation above, can be approximated by -half_error(z), since at any given radius, any function can be approximated arbitrarily well by a Laurent series. We throw away the negative terms for our entire half iterate. So naturally, there is an error term generated by the negative terms we threw away. I will give a more complete thorough analysis of all of the "stuff" near some of the zeros of f(z), and at the positive real axis, in a later post, by using the mi function, the the exp^{0.5}(y+mi(y)) function, and the Taylor/Laurent series. Code:{half_error= +x^ -1*  0.0038981927072789568905822639730044 +x^ -2* -0.0017214881666811347076719510576548 +x^ -3* -0.00013712753070231088880268844304361 +x^ -4*  0.00013907417275118396879544895662740 +x^ -5*  0.000051320769626162064593356837293471 +x^ -6* -0.000022366390692218719076651284271858 +x^ -7* -0.000023771055718461525133758346306702 +x^ -8*  0.0000013731519550483503617729218008457 +x^ -9*  0.000012027038063261925914628822119819 +x^-10*  0.0000039087491754042668374627189270986 +x^-11* -0.0000057892790783328757123356666460684 +x^-12* -0.0000051218889509626574137447251389938 +x^-13*  0.0000019098976291086407044050882321080 +x^-14*  0.0000048547491203337398558987063953054 +x^-15*  0.00000074114812516008642387127350420916 +x^-16* -0.0000038309228524031537802224892859947 +x^-17* -0.0000025748868434074919176037441576404 +x^-18*  0.0000022951530139076283185894597305398 +x^-19*  0.0000036906301942910425788903290199039 +x^-20* -0.00000035885452386419832591366749556753} - Sheldon « Next Oldest | Next Newest »

 Possibly Related Threads... Thread Author Replies Views Last Post Using a family of asymptotic tetration functions... JmsNxn 15 4,100 08/06/2021, 01:47 AM Last Post: JmsNxn Reducing beta tetration to an asymptotic series, and a pull back JmsNxn 2 729 07/22/2021, 03:37 AM Last Post: JmsNxn A Holomorphic Function Asymptotic to Tetration JmsNxn 2 1,139 03/24/2021, 09:58 PM Last Post: JmsNxn An asymptotic expansion for \phi JmsNxn 1 1,011 02/08/2021, 12:25 AM Last Post: JmsNxn Merged fixpoints of 2 iterates ? Asymptotic ?  tommy1729 1 3,367 09/10/2019, 11:28 AM Last Post: sheldonison Another asymptotic development, similar to 2sinh method JmsNxn 0 3,989 07/05/2011, 06:34 PM Last Post: JmsNxn

Users browsing this thread: 1 Guest(s)