# Why You Can’t Use Resistors as a Voltage Regulator

Due to WordPress’s abysmal handling of code blocks this blog post is now hosted at https://majenko.co.uk/blog/

People, all the time, use a simple resistor-based voltage divider to change 5v into 3.3v. You see it all over the place. For instance, when you want to get your 5V Arduino to communicate with a 3.3V ESP8266 – you use two resistors (10K and 22K say) to drop the 5V of the Arduino down to the 3.3V the ESP8266 expects.

And that is all fine and dandy.

However I all to often come across people attempting to do the same thing to power a 3.3V device off 5V. After all, if you can use it for changing 5V into 3.3V for the signals, surely the same thing works for the 5V power to 3.3V, right? I mean, it’s the same 5V isn’t it?

Well, no. You can’t. Absolutely not. And it’s all because of the current.

Let’s go back to basics. Ohm’s Law.

Ohm’s law states that the current through a conductor between two points is directly proportional to the voltage across the two points. Introducing the constant of proportionality, the resistance, one arrives at the usual mathematical equation that describes this relationship: Let’s apply that to the basic voltage divider: From basic principles we know that resistors in series are simply added together. So we can calculate the current that is flowing from 5V to Ground through the two resistors.

I=V/R = 5 / (10000 + 20000)

That equals 166.666µA.

We can then swivel the formula around and look at what voltage is dropped across one of the resistors – say R1. If I=V/R then V=R×I.

10000 × 0.0001666 = 1.666V

So if 1.666V is dropped across R1 then the voltage at the mid node (“3.3V Signal Out”) must be 5 – 1.666 = 3.333V. Which is correct.

And that is all fine. We have proved, with Ohm’s Law, that the voltage divider works.

But there is one important thing missing from that schematic, and that is the item the 3.3V is going into. So let’s add an ESP8266’s 3.3V power pin to it. Now an ESP8266 can draw up to 300mA when transmitting. That means the “3.3V signal out” must have up to 300mA coming out of it at times. That current has to come from somewhere, and that somewhere is the 5V supply – via R1.

So through R1 there is now 300mA plus 166.666µA, or 300.166mA.

But surely the voltage dropped across that resistor is dependant on the current flowing through it, yes? After all, V=R×I, isn’t it?

Absolutely. So let’s calculate it.

V = R×I = 10,000 × 0.300166 = 3001.66V

Whoah! Hang on there! How can three thousand volts be dropped across that resistor? There’s only 5 available! Quite right. So 5 volts is all that can be dropped across the resistor, making it 0V at the 3.3V pin instead of 3.3V. On top of that the current that flows through it is limited by Ohm’s Law: I=V/R, and V = 5V, R = 10000, so I=0.0005 or 500µA.

You see the problem here? The voltage at the output is completely dependent on the current being drawn through the output. As soon as you try to draw more than a few microamps out of the output the voltage drops. Draw more than a few hundred microamps and the voltage drops to nothing at all. Not only that the current that can possibly flow through there is severely curtailed.

And that is why you can’t use resistors to do the job of a proper voltage regulator.

## 3 thoughts on “Why You Can’t Use Resistors as a Voltage Regulator”

1. Fritz

could the use of tree diodes in series be a smart and simple solution for most projects?

Like

1. majenko Post author

Sure, if all you’re after is a rough ballpark voltage. Remember: a diode’s forward voltage changes with current and temperature. It’s not stable. Yes, it will give you “somewhere around” a 3Vf drop, then that’s fine. However that is all it gives you – a drop. If the input voltage changes then output voltage also changes. Assume at 25C and your required current the Vf of your diodes is 0.6V. That’s 1.8V drop for three of them. On a 5V supply that gives you 3.2V (great). Now the voltage coming in drops to 4.5V (batteries are going flat) – the output is now only 2.7V and your powered device may refuse to run or communicate properly. Also anything using that voltage in any calculations will now be wrong (working out the voltage of a signal on an ADC input where that voltage is the reference / VCC voltage). With a regulator the output will be 3.3V for *any* input voltage between 3.3V + the “dropout voltage” and the maximum input voltage. So if you have a 0.4V dropout voltage (for example) you can go down as low as 3.7V on the input and the output will still be 3.3V. You can go as high as the maximum input voltage (say 15V) on the input and still be 3.3V on the output. Yes, the higher your input voltage the more power gets dissipated as heat, which may be a problem as well, which is why switching (“buck”) regulators are used so much more often these days.

Like

2. Rhalf Wendel David Caacbay

i was thinking of both resistor and diode tricks but if those tricks works no one should have buildt regulators lol

Like