Strange result from taking logs of logs in experimental data

In summary, the conversation discusses the investigation of a dataset that shows a downward sloping shape. The speaker takes the base 10 log of the y-axis data and finds that the plot looks similar to the original. They try to manipulate the data to make the curves overlap, but are unable to do so. They then take the log of the absolute values of the logarithmic data and find that it produces the same familiar curve. The speaker wonders what type of function would have the same shape when taking the log of it. They also share some plots of the data and discuss different methods for fitting the data.
  • #1
Tcbovb
2
0
I have some experimental data (not hypothetically, this is real data), which has a downward sloping shape.

fMiDZpN.png


A sharp decrease initially, before settling to what looks almost like a downward linear slope. To investigate further, I decided to take the base 10 log of the y-axis data. The plot looked almost the same:

Dx6qXjV.png


At first I though this was a mistake, but I checked over the values and the code I used to generate this and it seems fine. If the right hand part is a slowly decaying exponential, this will appear approximately linear for both linear and logarithmic axes, so maybe that is the explanation. I tried to see if I could get the curves to overlap by scaling and shifting all the values equally, but that isn't quite possible (as one would expect). However these curves do seem to be "similar" in a sense.

So I then took the log of the absolute values of the logarithmic data. Doing this over and over again produced the same familiar curve, though sometimes it was mirrored, becoming an increasing function. After 10 logs to the original data [log(log(log ... log(Y)))], this curve finally breaks down, in part because the data before (9 logs) passes through Y = 0.

What I am wondering is what is the name of a function in which when you take the log of it, the resulting function also appears to have the same shape? What functions satisfy this condition and what is the name of such similarity?

Perhaps this is just the result of something straightforward that I haven't considered yet. At the moment it just seems so weird.

Attached are a few more plots. Data points occur every 100 in the X axis. Some consecutive data points have the same Y values, hence the square corners that seem to appear in the plots.

Hopefully this is the right forum for this, I'm not really sure where "unexpected results from data analysis leading to mathematical curiosity" fits in.

Tcb
 

Attachments

  • log_plots.zip
    180.9 KB · Views: 218
Mathematics news on Phys.org
  • #2
Tcbovb said:
What I am wondering is what is the name of a function in which when you take the log of it, the resulting function also appears to have the same shape?

Any function whose range isn't too large will have this property, at least approximately. As an example, consider the function f(x) = 1 + g(x) where g(x) is some other function, and |g(x)| is always much smaller than 1. Then

##\ln(f(x)) = \ln(1 + g(x)) = g(x) - g(x)^2/2 + g(x)^3/3 - ...##

where I used the Taylor series expansion of ln(1+x) on the right-hand side. If
|g(x)| is much smaller than 1, then all the extra terms on the far right are very small compared to g(x), so ##\ln(f(x)) \approx g(x)##.

But f(x) and g(x) have the same shape; f(x) is just shifted up by 1.

In fact, this argument is not specific to the logarithm function. You can convince yourself that essentially the same steps will work if you replace the logarithm with essentially any other function.
 
  • #3
Sure, that makes sense, but I am wondering if a Taylor series approximation fully explain what is going on here.

The data is certainly shifted, but it's not just shifted. The overall scaling of the function is different as well. Perhaps it's also due in part from the way the plots chose the y-axis scale so that the log taken plot looks more like the original data.
 
  • #4
Try the sum of a linear decay and an exponential decay. Something like A - x/B+C exp(-x/D). A will be about 4.9 and C will be about 1.6. B will be about 50000, and D will be about 7000.

Chet
 
  • #5
I have to correct my first answer. In fact, there was a mismatch between the data files to be ploted in the drawing software. The fitting is not so good than shown on the preceeding figure. It was too good to be true. I am sorry for the mistake. Now, the corrected answer :
The fit obtained with the function y = a + b xc leads to a mean squares deviation = 0.0354
On the attached figure, the computed curve (in red) is close to the original curve (in blue).
First, the original curve has been scanned in order to pick about 300 points with a convenient software.
Second, the values of the parameters a, b, c were computed thanks to the very simple method described pp.16-17 in the paper "Régressions et équations intégrales" (written in French, but the formulas to be applied are understandable in any language). Published on Scribd :
http://www.scribd.com/JJacquelin/documents
The method of regression is straigthforward (process not itterative, no need for a guessed initial value).
Of course you could as well use a more common method for non-linear regression, but you will be prompted to give an initial guess for the parameter c.
 

Attachments

  • Copy Figure.JPG
    Copy Figure.JPG
    22.1 KB · Views: 273
Last edited:
  • #6
An almost as good fitting is obtained with y = a + b*x + 1/(c*x + d) : Figure in attachment
The mean squares deviation is slightly higher and the déviations are more important in the range of small values of x. But the déviations are clearly lower in the range of large values of x.
 

Attachments

  • Copy Figure2.JPG
    Copy Figure2.JPG
    22.5 KB · Views: 307
  • #7
An even better fit is obtained with y = a0+ a1x + b xc
Mean squares deviation = 0.0107
The computation of the values of parametres was done with the method of transformation of the non-linear equation to a linear equation thanks to a convenient integral equation (paper already referenced above).
 

Attachments

  • Copy Figure3.JPG
    Copy Figure3.JPG
    22.4 KB · Views: 321
Last edited:

What is the purpose of taking logs of logs in experimental data?

Taking logs of logs in experimental data is a common practice in scientific research to transform the data into a more linear relationship. This allows for easier visualization and analysis of the data.

Why do some experiments yield strange results when taking logs of logs?

Some experiments may yield strange results when taking logs of logs due to a few reasons. One possibility is that the data may not follow a logarithmic relationship, which can skew the results. Another reason could be errors in data collection or measurement.

Can taking logs of logs alter the interpretation of the experimental data?

Yes, taking logs of logs can alter the interpretation of the experimental data. This transformation can change the shape of the data and make it easier to identify trends or patterns that may have been hidden in the original data. However, it is important to carefully consider the implications of this transformation on the interpretation of the results.

Are there any limitations to taking logs of logs in experimental data analysis?

While taking logs of logs can be a useful tool for data analysis, it is not always appropriate for all types of data. It is important to carefully assess the data and determine if this transformation is suitable for the research question at hand. Additionally, taking logs of logs can also introduce errors, so it is important to be cautious when interpreting the results.

What steps can be taken to avoid strange results when taking logs of logs in experimental data?

To avoid strange results when taking logs of logs in experimental data, it is important to carefully examine the data and ensure that it follows a logarithmic relationship. If the data does not follow a logarithmic relationship, consider using a different transformation or analyzing the data in its original form. Additionally, it is important to carefully check for any errors or inconsistencies in the data collection process.

Similar threads

Replies
4
Views
2K
Replies
24
Views
2K
  • General Math
Replies
15
Views
2K
  • General Math
Replies
2
Views
1K
  • General Math
Replies
12
Views
962
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
11
Views
755
Replies
20
Views
1K
Replies
4
Views
1K
  • STEM Educators and Teaching
Replies
5
Views
656
Back
Top