Dismiss Notice
Join Physics Forums Today!
The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

Standard for volt

  1. Jun 25, 2014 #1
    I've worked in calibration for a few years and I know that my equipment has to be calibrated by a standard that is a certain amount more accurate than the piece of equipment I'm calibrating, but where exactly does it end where the most accurate piece of equipment is calibrated by the absolute standard? Can a piece of equipment ever actually accurately measure 1 volt?
  2. jcsd
  3. Jun 25, 2014 #2

    jim hardy

    User Avatar
    Science Advisor
    Gold Member

    Now there's a question that opens up a whole world of its own, metrology.

    http://www.sandia.gov/psl/Dc_fact%20Sheet2008_Final.pdf [Broken]

    Last edited by a moderator: May 6, 2017
  4. Jun 26, 2014 #3
    I gotta say that's pretty damn f*ing accurate, but does that mean we are taking measurements directly from this josephson array standard. That's still not exactly accurate right?
  5. Jun 26, 2014 #4


    User Avatar
    Science Advisor
    Gold Member

    Yes it is, sort of. If the Josephson voltage was a "true" standard; it would -by definition- be exact.

    In reality it is a bit more complicated than that since the Volt is not actually a base unit (the Ampere is). Hence, what we are actually doing is using the accepted CODATA value for the Josephson constant (483.6 GHz/V). The Josephson constant can be used to relate frequency and voltage, and since we can get the frequency from our hydrogen masers ("atomic clocks") the precision of that is extremely high (probably at least one part in 10^14 if you integrate for long enough).

    This is how voltage metrology has been done in the past 25 years or so. From a practical point of view it means that we can calibrate voltage standards (in this case Zener diodes) with an accuracy much, much higher than for anything me need in practical applications.

    Btw, the 0.02 ppm is VERY conservative. Most standard labs can calibrate their secondary standard with an accuracy that is at least one order of magnitude better than advertised, this is to have some tolerance for unexpected aging etc.

    EditL I forgot to say that is possible to just calibrate say a multimeter from a Josephson array setup. This is done routinely in some demanding applications: some high-end multimeters have calibration inputs a the back, and by re-calibrating them say once a day (which can be done automatically) one can avoid problems with drift etc.
  6. Jun 26, 2014 #5

    jim hardy

    User Avatar
    Science Advisor
    Gold Member

    There's accuracy and there's resolution.

    A meter is accurate when it reports correctly to the limit of its ability.

    A three digit meter is"accurate" if all three digits are true.
    I once had access to a six digit multimeter, but that much resolution was useless - we had nothing that'd stand still enough the last two digits wasn't all over the place. Well, except a meter calibrator.

    They're talking about eight digits.. boggles my mind. That takes a lab environment.

    When you need to resolve microvolts, you struggle with effects like the temperature gradient along your measuring wires , dissimilar metals in your test prods and the like. These guys talk of ability to resolve, indeed measure, down to a tenth of a microvolt.
    100X that, around ten microvolts, is a fraction of a degreeF to a thermocouple. In forty years of practical industrial work i never had any need to resolve DC voltage any closer than that. Four digits were enough. But you need a five digit meter to check your four digit one. And so on....

    Look ma, no participles !

    old jim
Share this great discussion with others via Reddit, Google+, Twitter, or Facebook