hello, i know car battery amps range from 40 amps to whatever, i just cant understand them, the amps aren't printed on the batteries and i don't know why? i have a pretty good ideal that my car battery maybe 45 amps, what does this mean? i know the alternator produces 55 amps, if i hooked a 100 watt light bulb up to my 45 amp car battery i know that the light bulb uses 100w/12v = 8.3 amps, so how do i figure how long the 45 amp battery will run a 8 amp bulb?
i was reading something on the web and was wondering if this could be right, knowing the battery is 45 amps, and i am using around an 8 amp bulb, i would assume 8 amps is how much the bulb uses per hour??? if so, 8/45 = around 5.6 hours? could this be right?
i think you can do it this way to? take the 45 amps * 12v to get 540 watts, so we know the bulb is 100 watts, so i guess the bulb uses 100 watts per hour, which means it would last, estimating in my head, looking at the battery which is 540 watts, id say it would run little over 5 hours? does it sound like i'm on the right track?
Welcome to PF, Mountain. There must be a misunderstanding somewhere. If you have a 45 amp battery, you sure as hell aren't driving anywhere. The minimum cold cranking amps for a car battery is probably around 350. My Roadrunner draws 1,000 amps in sub-freezing temperatures.
Yes 45 amps sounds way too low. Automotive batteries supply much more current - the one I keep around the house is 650CCA (550 otherwise) but is slightly larger than what you would see in a car. Anyway, some batteries will give a rating of the full charge in Amp-hours, usually in the form of "y amps for x minutes" written somewhere on the identification label. The product of the current supplied and the discharge time is relatively constant, so if you know how many amps your lightbulb will pull then you can guess roughly how long it will take to discharge the battery.
cold cranking amps are different from amp-hours which is what you want, cca is how many amps can be delivered short term during cold weather, amp-hours is how many amps can be continuously be delivered during one hour. so if you have 45 amps-hour then you get 540 watts that can be delivered continuous for an hour, the cca number will be higher, say 120 amps that can be delivered to the starter instantaneously during winter conditions.
my friend on msn is telling me that a car battery as way more than 45 amps, and ive seen it here, i apreciate all the help, but would you not agree that a 100 watt household bulb on a 12 volt battery would be using around 8.3 amps? and how long could i exspect this bulb to run? 5.4 hours? if so, would that prove the battery was 45 amps? im sorry, i am fairly new to this and do understand i am lacking some knowledge
You're confusing the relationship between max output current (rate of change of charge / time) with the amount of charge (Q) the battery has. You can't get the max current (dQ/dt) the battery could ever output based on how long it took for the battery to discharge from loading it with a lightbulb that consumes such little power. The lightbulb will draw only so much current - period, and as long as the battery can deliver 100W then the lightbulb will shine. What you *could* do is attach 5 100W lightbulbs in parallel and confirm that they all are bright and observe upon the addition of a 6th or 7th that there would be some diminishing brightness. That would confirm more or less you have a (5 to 6) * 8.3A ~=45A of supply current.
that wouldn't work, the battery would supply more current. the way you measure a car batteries output is by it's voltage level for rough estimates. hook up the 5 light bulbs and put a voltmeter in parallel, watch as the voltage drops from 13.5 volts at no load to 12.5~13 at rated amp-hour load to 11~11.5. that should tell you more or less how much power the battery can supply over a givin time, if you want more amps increase the number of light bulbs. car batteries charge to ~13.5 volts and are considered dead at 11.
I'm no expert, but I think this is incorrect. The resistance of the bulb remains the same when it's hooked up to 12V battery, therefore the current it draws should be 0.083 amps (i.e. 1/10th of what it draws at 120V). This turns the bulb into a 1watt (12V * 0.083amps) bulb; i.e. its glow will be barely detectable.
a 60watt 120v bulb draws 7.2watts and is 20 ohms cold off of 12v so a 100watt bulb will draw about 10.2 watts cold.
MountainDew, The 45Amp your talking about is a per hour rating, meaning that your battery could give you a 45Amp current for one hour before it dies out. About the second part of your question, when you take a 120Watt bulb and connect it to a 12V car battery, you can simply calculate the current drawn by the bulb from the relation P=VI, now since it's a 120Watt bulb and the car battery voltage is 12V the current drawn will be 10Amps, now since your battery can supply 45Amp for one hour then it can Supply 10Amps for 4.5 hours.
Wow! Lot's of misinformation here. An incandescent light bulb is not a linear device. The more current that flows through the bulb, the higher the resistance of the bulb gets. Also, why do you assume that no matter what voltage source you hook a bulb to it will automatically dissipate the rated wattage? The 120 volt 120 watt bulb will dissipate 120 watts AT THE VOLTAGE IT WAS DESIGNED TO OPERATE AT! It will NOT draw 10 amps, ever.
Averagesupernova, I think most of us assumed an ideal condition, we looked at it as a HW problem rather than the practical case.
The "ideal" condition wouldn't be holding the power constant, it would be holding the resistance constant.
i googled and learned to use the formula Watts x volts = amps, how are you getting 10 watts from a 100 watt bulb if you don't mind, from what i've read i've learned to do it like this. 100 watt bulb at 120 volts draws 100/120 = .8 amps 100 watt bulb at 12 volts draws 100/12 = 8.3 amps or you could do watts like this for example: .8 amps from 120 volts draws .8 * 120 = 96 watts per hour 8 amps from 12 volts draws 8 * 12 = around 96 watts per hour the cold cranking amps on the battery is labeld 650 CCA or cold cranking amps, i would assume that would mean the battery could take a load of 6 100 watt bulbs for a very short period of time. now that the battery is 650 CCA that would mean 650 * 12 = 7800 watt hours, which they are made to take loads like that for small amounts of time? if 45 amp hours was right then 45 * 12 = 540 watt hour, or 45 amp hour, same thing in my book. so the question is whats the amp hours of the battery, i read on the internet the other day that the basic car battery holds about 45 amp hours, and i also read somewhere else last night that a basic car battery holds around 60 amp hours or 60 * 12v = 720 watt hours, meaning it could produce 720 watts in one hour and no more. i think marine batteries start at 100 amp hours and go up from there? i need to get me a ohm or amp meter and learn to measure the ohms or amps to see more about this. by the way, i like this thread, im learning lots about amps watts and volts. if my 650 CCA battery in my truck was 650 amp ours, that would mean it was 7800 watt hours, if the battery produced that over time then that would mean it would run a 100 watt light bulb 7800 watts/100 watts = 78 hours, or 100 watt bulb on 12 volts uses 100/12 = 8.3 amps, so 650 amps / 8.3 amps = 78 hours, same formula as above giving the same hours, which i dont think 650 CCA is what you go by? funny how that works out.
i have a 12v battery on the bench and i hooked up a lightbulb to to test :P, cold i got 20 ohms from a 60 watt lightbulb, 60/100 = 0.6, 0.6*20 = 12, back to ohms law, 12 volts / 12 ohms = 1 amp, 1 amp * 12 volts = 12 watts (your right about me getting the wrong number the first time, must have pressed the wrong button lol:), so your 650 amp-hour batt would run a 100 watt light bulb for 65 hours. cca is different then amp-hours, see thread. edit: let me add that because the resistance of a 100 watt house light bulb won't change the battery won't be dead for 650 hours, you won't get that 100 watts only 12.
MountainDew, resistance is a physical property of a resistor, so you need to use v=ir, find the new amperage and then plug that new amperage into p=vi to find the power at the lower voltage. A 100 watt bulb is labeled as such because the manufacturer assumes you wouldn't try to run it at any voltage other than what it was designed for, not because it would dissipate 100 watts at any voltage. Though as said above, the resistance will be temperature dependent. Also, watts is a rate already - there is no such thing as "watts per hour". You may mean watt-hours, which is watts times hours (not divided by, which is what "per" means)
russ_watters, Sorry, I didn't realize that it was a 120V bulb I thought it was a 12V bulb, now if it is a 120Volt bulb then things would be a lot different, first you would use [tex]\frac{P_1}{P_2} = \frac{V_1^2}{V_2^2}[/tex] to get the new power drawn when connected to the 12V source, in this case it would be about 1.2Watt, as you properly realized this is not enough power to light the bulb, but it WILL draw some current anyway, so using the relation V=IR you get the current which will be about 0.1Amp, and since the battery gives away 45Amps for one hour it can give 0.1Amps for 450 Hours.
I have an idea that your car battery is capable of producing a heck of a lot more than 45A. When you first turn the key on it is probably producing thousands of amps for less than a second. The first thing you have to do is create a magnetic field in the starter. To electrial engineers this is called inrush, not to be confused with starting current. This will last for less than a second but can be extremely high. Once the magnetic field is created then the normal starting current is maintained. This could be in the area of 500A or higher but I am just guessing.