Yes, if you are using AWG # 10 wire. If you are using AWG # 12 then the answer is no you must use a 20 amp breaker. On AWG #14 wire you must use a 15 amp breaker.
30 amps at 120 volts is 3600 watts. 30 amps at 240 volts is 7200 watts.
A #8 AWG copper conductor with an insulation factor of 90 degrees C will limit the voltage drop to 2.26% or less when supplying 30 amps for 120 feet on a 220 volt system.
For single phase 30 amps at 120 volts you would need a #8 copper wire with an insulation rating of 90 degrees C.
A 120V power supply connected to a 30 Ohm resistor will produce 120/30 or 4 amps of current.
As many as you want. It depends on the capacity of your supply. Let me break this down. Current or amps is what is drawn by a device or a load. For example, lets say you plug in a 1500 watt 120 volt space heater and turn it on. Now you want to know what kind of amperage is being drawn by this device. Take 1500 watts and divide 120 volts into it. The answer will be your amperage or current. The current draw would be 12.5 amps. (I know there are other people out there who say no no that is not entirely right. I know this. I do not wish to give a lesson in calculating impedance when this answer is close enough and is with in a safe range of accuracy. Just keeping it simple)Another fact that may be of interest to you is there is as much as a 8000 fault amp potential in the average 100 amp service panel. This means that if your two lines coming into your panel would short it would blow up. This is what is called an Arc Blast.
30 amps.
It's not that simple. The basic formula is Volts / Ohms = Amps. For 30 Volts you'd get 0.5 Amps, for 60 Volts you'd get 1 Amp, for 120 Volts you'd get 2 Amps.
30 amps at 120 volts is 3600 watts. 30 amps at 240 volts is 7200 watts.
W = Amps times Volts. A = W/V, A = 300/120 = 2.5 amps
10 AWG.
It's not that simple. The basic formula is Volts / Ohms = Amps. For 30 Volts you'd get 0.5 Amps, for 60 Volts you'd get 1 Amp, for 120 Volts you'd get 2 Amps.
Watts is volts times amps, so 12 x 30 = 360 watts
That depends on the voltage, but the residential standard is 240 volt. At that voltage you sit at around 15 amps, however it MUST be on a 20 amp circuit for national (US) or Canadian electrical code, as you can only load your circuit to 80% of it's capacity.
V=IR, V=120 VOLTS R=30 OHMS I=V/R, I=120/30, I = 4Amps
The average 17" LCD monitor will pull around 25 to 30 watts. At 120 volts this would equate to 0.2 to 0.25 amps.
A #8 AWG copper conductor with an insulation factor of 90 degrees C will limit the voltage drop to 2.26% or less when supplying 30 amps for 120 feet on a 220 volt system.
As asked, the question cannot be answered. At 1 volt, 300 Watts = 300 Amps. At 10 volts, 300 Watts = 30 Amps. At 100 volts, 300 Watts = 3 Amps. At 120 volts, 300 Watts = 2.5 Amps. At 240 volts, 300 Watts = 1.25 Amps. To calculate the relationship between Amps, Volts and Watts, use the formula: Watts = Amps * volts