IR Temperature Detection with sensitivty of 1 degree - possible?

Click For Summary

Discussion Overview

The discussion revolves around the feasibility of designing a system capable of detecting a temperature difference of just 1 degree Celsius between a hot object and a cold surface using infrared (IR) technology. Participants explore various sensor options, calibration needs, and potential challenges in achieving accurate readings from a distance of approximately 10 cm or more.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Experimental/applied

Main Points Raised

  • One participant seeks advice on suitable sensors or systems for detecting a 1-degree temperature difference from a distance, suggesting that calibration may be necessary.
  • Another participant mentions that handheld Fluke Thermal Imagers can detect differences down to 0.5 degrees Celsius, noting their effectiveness for relative temperature differences but high cost.
  • A participant who has used thermal imagers expresses interest in building a custom system and inquires about sensors with analog outputs, questioning the impact of IR laser thermometers on the object being measured.
  • One reply suggests that IR laser thermometers are passive sensors, primarily for aiming, and may not be very accurate but can detect hot spots effectively.
  • A participant shares their experience with infrared thermocouples, highlighting several challenges such as low signal levels, high impedance, sensitivity to case temperature, and emissivity effects, while recommending a method of taking reference readings to improve measurement accuracy.

Areas of Agreement / Disagreement

Participants express a range of opinions on the effectiveness and suitability of different IR sensors and methods, with no consensus reached on the best approach or specific technology to use for the desired application.

Contextual Notes

Participants mention various limitations and challenges associated with different sensor types, including calibration needs, accuracy issues, and environmental factors affecting readings. Specific technical details and assumptions about sensor performance are also noted but remain unresolved.

MIMSAR
Messages
5
Reaction score
0
Hi,

I'm trying to design a system which will be able to detect a hot object on a cold surface. The downside is the temperature difference is only 1 degree celsius and I'd like to use IR from a distance of about 10cm (further away preferred if possible).

Does anyone with any experience of this know a suitable sensor or system I could use please. I'm guessing some form of calibration will be needed to tell the system the temperature of the cold surface first.

Thanks,
 
Engineering news on Phys.org
The handheld Fluke Thermal Imagers I've used could detect differences down to 1/2 a degree C. They weren't that great for detecting the absolute temperature, but on a false color display, the relative differences were easy to see.

You might need to shop around. My experience was from a couple of years ago, but the cameras are very expensive. We leased rather than purchased one. We wanted a camera that could export the numeric data so we could do more quantitative analysis of the temperature gradients.
 
Thanks. I've actually used a thermal imager and found it to be quite effective. However, I'm now trying to build my own system tuned to the required temperature and the ability to interface with other components. Are there any suitable sensors/circuits, ideally giving an analogue output of the temperature at the focus?

I considered an IR laser thermometer but wondered would the beam of IR actually heat the object up in addition to reflecting to give a temperature reading?
 
You can buy thermal imaging cameras with USB interfaces like this:

http://www.infraredcamerasinc.com/fix-mounted-thermal-imaging-equipment.html"

I've never priced anything like this, but I'd imagine it is pretty expensive.

IR laser thermometers are still passive sensors. The laser is just for aiming. The aren't very accurate, but plenty good enough for detecting hot spots on circuit boards or in the AC system is working.

If you need a fairly precise infrared temperature, check the Omega catalog:

http://www.omega.com/search/esearch...ed+Sensor&submit=Search&ori=Infrared+Sensor+"
 
Last edited by a moderator:
I've done this using infrared thermocouples. You can purchase them at Omega.com. There are a couple of downsides to using these:

1. The signal level is tiny - on the order of 50uv/C
2. They have a high impedance - several k-ohm, the leakage current of many standard thermocouple meters will swamp the circuits output.
3. They're affected by their case temperature - ours was compensated by a constant temperature jacket.
4. They're affected by the emissivity of the surface. We used a rough surface to help compensate.
5. Ours had a Germanium lens. Touching the lens was enough to start it corroding.

The easiest, most sure way of taking this measurement is to point the sensor at one surface, take a reference reading, point it at the other, take another reading, and then subtract the difference. This will make up for a great many shortcomings in the system.

Good Luck,

Mike
 

Similar threads

  • · Replies 20 ·
Replies
20
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
Replies
0
Views
934
Replies
8
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 16 ·
Replies
16
Views
2K
Replies
6
Views
2K
  • · Replies 6 ·
Replies
6
Views
3K
Replies
18
Views
3K
  • · Replies 2 ·
Replies
2
Views
4K