How to write measured value with error

  • Context: Undergrad 
  • Thread starter Thread starter UiOStud
  • Start date Start date
  • Tags Tags
    Accuracy Error Value
Click For Summary

Discussion Overview

The discussion revolves around how to properly express a measured value along with its associated error, particularly in the context of a lab exercise involving a teslameter. Participants explore the implications of significant figures, rounding rules, and the representation of uncertainty in measurements.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Mathematical reasoning

Main Points Raised

  • One participant measured a value of 26.3 mT with an error of ±0.0132 mT and questions whether to express this error given its smaller magnitude than the least significant digit of the measurement.
  • Another participant suggests writing the value as 26.31 to indicate uncertainty, but acknowledges the limitation of the teslameter's display.
  • A different viewpoint proposes rounding the error to 0.01, arguing that this reflects the uncertainty without neglecting it.
  • Several participants reference a "15 unit rule" for rounding errors, indicating that different rounding practices exist among users.
  • Concerns are raised about the precision of the measurement relative to the error, with some suggesting that the measuring equipment should display more significant figures to match the error's precision.
  • One participant discusses the range of possible values indicated by the teslameter's display, suggesting that the actual value could vary significantly based on the error margin.
  • Another participant emphasizes that the accuracy of digital readouts is limited by the least significant digit, suggesting that the percent error may be ignored if it is less than half of the least significant digit.
  • A later reply challenges the assumption that digital instruments round values before displaying them, proposing an alternative interpretation of how digital measurements should be expressed.

Areas of Agreement / Disagreement

Participants express differing opinions on how to represent measured values and their errors, with no consensus reached on the best approach. Various rounding rules and interpretations of measurement precision are debated.

Contextual Notes

Participants highlight limitations related to the display capabilities of the teslameter, the significance of digits in expressing uncertainty, and the potential for different rounding conventions to affect the representation of measured values.

UiOStud
Messages
9
Reaction score
2
During a labexercise we measured a value to be 26.3 mT. The teslameter was said to have an accuracy at +- 0.05%. This means we measured 26.3 +- 0.0132mT. However we are supposed to write a measured value with it's error in the last digit. But the error is smaller than the smallest digit we got from the measurement. Does this mean I shold not write 26.3 +- 0.0132mT and just write 26.3 mT without any estimation of the error?
 
Last edited by a moderator:
Physics news on Phys.org
Not sure, If I can be right, since all the significant digits in 0.0132 are uncertain, and in 26.3, all digits are certain, How about writing off a value with all certain digits and one uncertain digit at end? Like 26.31? That last digit '1' shows uncertainty in your measurement denoting error.

It is better you get this verified with someone before I misguided you.
 
Thank you, that is what I would like to do. But the Teslameter only gave me the first decimal digit so I wouldn't know what to write after 26.3. I don't think I could just add zero when I didn't measure it.
 
After referring my physics textbook regarding taking errors, You can write the measured value with error as 26.3 ± 0.01 as in 0.0132, you can consider 32 as insignificant and round them to 30 getting, 0.013 and then to 0.01. Now you can't neglect 0.01. Simply writing 26.3 will say kind of no error.
 
When I was doing undergraduate lab work, we had a "15 unit rule" which said that ##3.77\pm 0.24## is rounded to ##3.8\pm 0.3##, while ##5.171\pm 0.124## is rounded to ##5.17\pm 0.13##. Not everyone uses the same rule, though.
 
  • Like
Likes   Reactions: AlphaLearner
hilbert2 said:
When I was doing undergraduate lab work, we had a "15 unit rule" which said that ##3.77\pm 0.24## is rounded to ##3.8\pm 0.3##, while ##5.171\pm 0.124## is rounded to ##5.17\pm 0.13##. Not everyone uses the same rule, though.
Couldn't even apply any single rule regarding way you rounded off digits. Can you explain the logic behind your rounded figures hilbert2.
 
If the first nonzero digits in the error are less than "15", the error is rounded to two significant figures, otherwise it's rounded to one significant figure. And an error is always rounded upwards.
 
hilbert2 said:
If the first nonzero digits in the error are less than "15", the error is rounded to two significant figures, otherwise it's rounded to one significant figure. And an error is always rounded upwards.
Following you, should the measure come as 26 ± 0.14?
 
The result should have the same precision as the error. Here the result 26.3 mT doesn't have as good precision as the error rounded to two or one significant figures, which implies that the actual error is larger than OP thinks. I don't really get why the measuring equipment wouldn't show as many figures as the magnitude of error allows.
 
  • #10
If the meter can only display 3 digits the the meter could have measured any value between 26.25 to 26.35 and it would display 26.3.

Then the meter could be inaccurate by +/- 0.0132 so the actual value could be anywhere from 26.25-0.0132=26.2368 to 26.35+0.0132=26.3632

I think.
 
  • #11
CWatters said:
If the meter can only display 3 digits the the meter could have measured any value between 26.25 to 26.35 and it would display 26.3.

Then the meter could be inaccurate by +/- 0.0132 so the actual value could be anywhere from 26.25-0.0132=26.2368 to 26.35+0.0132=26.3632

I think.
When you have two readings one was 26.2368 and other was 26.3632, how about taking the closest value to true value by simply taking their mean? Last digit will tell error. Even probability of getting a true value from it is pretty less due to too many uncertain digits in error.
 
  • #12
UiOStud said:
During a labexercise we measured a value to be 26.3 mT. The teslameter was said to have an accuracy at +- 0.05%. This means we measured 26.3 +- 0.0132mT. However we are supposed to write a measured value with it's error in the last digit. But the error is smaller than the smallest digit we got from the measurement. Does this mean I shold not write 26.3 +- 0.0132mT and just write 26.3 mT without any estimation of the error?

With any digital readout the accuracy is never better that ±1 least-significant-digit (LSD). That's because you have no more information being given. The specifications of a device with a digital readout are given as ± a percent of either reading or full scale, ± a number of counts. If the percent error works out to less than LSD/2 then the LSD error dominates and the percent error is ignored.
 
  • Like
Likes   Reactions: pixel and AlphaLearner
  • #13
AlphaLearner said:
When you have two readings one was 26.2368 and other was 26.3632, how about taking the closest value to true value by simply taking their mean? Last digit will tell error. Even probability of getting a true value from it is pretty less due to too many uncertain digits in error.
You don't have two readings.
 
  • #14
Tom.G said:
With any digital readout the accuracy is never better that ±1 least-significant-digit (LSD). That's because you have no more information being given. The specifications of a device with a digital readout are given as ± a percent of either reading or full scale, ± a number of counts. If the percent error works out to less than LSD/2 then the LSD error dominates and the percent error is ignored.
+1

Although I don't think you can ignore the percent error unless it's much less than LSD/2.
 
  • #15
It appears that there is an assumption made in this thread (and seemingly everywhere else on the Internet) that digital instruments round their values before displaying. But I don't think that is correct.

For example, if I measure the time it takes for a ball to roll down a ramp with a timer that only measures to the whole seconds, then a typical result could read (say) 45 seconds. However, the digital display isn't going to display "45" until the measurement reaches at least 45.000000000001 seconds and will retain the "45" until the measurement reaches "45.999999999999" seconds.

It would seem to me that a digital device displaying "45" should be expressed as 45.5 \pm 0.5 seconds, not 44\pm 1 second nor 45 \pm 0.5 seconds.

Do I have this right?
 

Similar threads

  • · Replies 25 ·
Replies
25
Views
4K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 1 ·
Replies
1
Views
813
  • · Replies 2 ·
Replies
2
Views
732
  • · Replies 13 ·
Replies
13
Views
4K
Replies
4
Views
2K
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K