answersLogoWhite

0


Best Answer

Since power is volts time amps, the current in a 60W lamp connected to 120V is 0.5A. Since a lamp is a resistive load, there is no need to consider power factor and phase angle, so that simplifies the explanation.

========================

Assuming this is an incandescent or halogen lamp (using a filament to make the light) there is a trick here: the resistance of a lamp filament varies with temperature and does not follow Ohm's law. The resistance will be much lower, thus the current will be much higher when the filament is cold, when the lamp is first connected. As the filament heats up, the resistance increases until it gets to a steady operating point of 0.5A.

For a halogen lamp, the operating temperature is about 2800-3400K, so the R at room temperature is about 16 times lower than when hot... so when connected, the current is about 8A but drops rapidly. The current could be even higher if the lamp is in a cold environment. Non-halogen lamps operate at a lower temperature and would have a lower initial current--about 5A.

And this all assumes the lamp is rated for 120V. If it is a 12V/60W lamp, the filament will probably break and create an arc, which may draw a very large current.

User Avatar

Wiki User

13y ago
This answer is:
User Avatar

Add your answer:

Earn +20 pts
Q: What is the current when a 60W lampis connected to 120v?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Continue Learning about Electrical Engineering

If you put a 60W 120V bulb in a 120V lamp that has a transformer for 240V and plug it into a 240V outlet will it work or blow the circuitry?

Off hand no but your explanations are not clear as to who is doing what to whom,,


Two lamps A and B are rated 60W 220V and 100W 220V respectively so why does the answer say that when connected in parallal bulb A is brighter?

In parallel, they both obviously have 220 v across them, so the 100 W bulb is obviously brighter than the 60 W one. The 60 W bulb has more resistance, and in series they both have to pass the same current, so the 60 W has more voltage across it and might be brighter.


A 40W bulb or a 60W bulb will glow more brighter if connected in series?

The 40-W lamp will be the brighter, because it has a greater resistance that the 60-W lamp and, therefore, will experience a greater voltage drop across its terminals. The closer the voltage drop is to the lamp's rated voltage, the brighter it will be.


100w bulb and 60w bulb connected in parallel to 220v which will glow brightly?

Total supply voltage = 220vrated power of first lamp= 100 wattso current though it, I=p/vI=100/220=.45 ampsResistance offered by first lamp=220/.45= 488 ohms( i avoid fractions)..............................................................rated power of second lamp=60 wattsvoltage is same, so current through it = 60/220=.27 ampsresistance of second lamp = 220/.27=814 ohms( i avoid fractions).........................................................................power drawn by first lamp =I12 R=(.45)2 * 488 =98.82(=100)power drawn by second lamp =I22 R=(.27)2 * 814 =59(=60)...................................................................................SO 100 WATT BULB WILL GROW BRIGHTER AS IT HAS MORE POWER


Is the current always the same in parallel circuits?

An example of a parallel circuit would be the light bulbs in track lighting. Each bulb has the same voltage applied. The current through any one light bulb equals the voltage divided by the resistance of the bulb. The current also equals the wattage of the bulb divided by the voltage. So if all the bulbs had exactly the same resistance the current would be the same. However, your question says "always" so in general the answer is no. In the case of track lighting if you had a 60 watt bulb in parallel with a 120 watt bulb, the 60W bulb would draw 1/2 amp and the 120W bulb would draw 1 amp. The sum of the current flowing in a parallel circuit equals the sum of the current in each leg of the circuit.

Related questions

Why would a high watt lamp draw more current than a low watt lamp?

Power is measured in Watts, power (Watts) = E (volts) x I (current - amps) current is determined by the internal resistance (R) of the lightbulb, the lower the resistance the more current will flow. 120v x 0.5a = 60W 120V x 0.83a = 100W the 100W lightbulb will draw more current We also have Ohm's law: E(volts) = I (amps) x R (ohms) Household voltage stays the same at 120v we have for a 100w lamp: 120v = I x R R = 120v/0.83 amps R = 144.6 ohms for a 60w lamp: 120v = I x R R = 120v/0.5 amps R = 240 ohms The higher watt lamp has lower resistance.


How does the current in a light bulb connected to a 400 V source compare to the current when this light bulb is connected to a 60 V source?

Let's examine what it means when a bulb is 100W rather than 60W. I'm assuming that you meant to state that they are 120V bulbs being connected to a 240V circuit1. With the same voltage on each, and because power is voltage times current, the current must be greater in a 100W bulb than in a 60W bulb. Since a incandescent bulb is a linear load, if you double the voltage then you double the current2. So the current through the 100W bulb is still greater than through the 60W bulb. Or you may analyze it a bit more. With both on 120V, for more current to flow in the 100W bulb, the resistance of it must be less than that of the 60W bulb. So you may generalize that under any voltage (same voltage applied to each), the 100W bulb will always have more current through it than the 60W bulb. 1Actually, if they are 120V bulbs in a 240V circuit, there is a high probability that they will blow out. But before they do, this is what will happen. 2Well, slightly less than double, because the temperature coefficient on the filament is positive, so the hotter it is, the greater the resistance. Although this may seem nonlinear, a light bulb or other temperature sensitive resistive element is still defined as linear if over the short term it obeys Ohms law at any instant of the waveform. The current in the 100 watt bulb will be greater. Power is current times voltage, so current is power divided by voltage. Voltage is the same is both cases of this question, so current is proportional to power at 240V.


If you put a 60W 120V bulb in a 120V lamp that has a transformer for 240V and plug it into a 240V outlet will it work or blow the circuitry?

Off hand no but your explanations are not clear as to who is doing what to whom,,


How do you calculate ampere given 250v AC 60W?

Divide the watts by the voltage to get current (amps). 60W / 250V = about 1/4 Amp.


Will 3 60W globes glow more brightly when they are connected in series or parallel?

In parallel they will be brighter because there will be the same voltage drop and the same current through each. In series they will each have 1/3 the voltage and current. Since the current heats up the filament lesser current is less light.


Why 60w bulb more bright than 100w bulb connected in series?

There is insufficient information in the question to answer it. In series with what? Please restate the question.


Where do you put a 60v bulb?

If you meant 60w (or watts) then it is used primarily in small table lamps and closets. If you meant 60 v (or volts) then you could use it in series with another 60v bulb on a standard household 120v circuit.


What continent is in 20s 60w?

The coordinates 20S 60W correspond to South America.


Difference between 60W and 2X30W guitar amplifier?

I believe that a 2x30w has two speakers that run twenty watts and a 60w has any number of speakers as long as the total is 60w


An electric lamp is rated at 60W the current it draws from a 230 V supply is?

P=VI so I=P/V I= 60/230 I=0.261 A


How much power is consumed by a load drawing 5 A with 12 V across it?

Power = Voltage x Current. So it will be 60W of power consumption, in your case.


What country is located 30S 60W?

The coordinates 30S 60W point to Argentina, a country located in South America.