As we all know, AC won the “War of the Currents”. The reasoning behind this is that AC voltage is easy to convert up/down with just a ring of iron and two coils. And high voltage allows us to transport current over longer distances, with less loss.

Now, the War of the Currents happened in 1900 (approximately), and our technology has improved a lot since then. We have useful diodes and transistors now, we have microcontrollers and Buck/Boost converters. We can transform DC voltage well today.

Additionally, photovoltaics produces DC naturally. Whereas the traditional generator has an easier time producing AC, photovoltaic plants would have to transform the power into AC, which, if I understand correctly, has a massive loss.

And then there’s the issue of stabilizing the frequency. When you have one big producer (one big hydro-electric dam or coal power plant), then stabilizing the frequency is trivial, because you only have to talk to yourself. When you have 100000 small producers (assume everyone in a bigger area has photovoltaics on their roof), then suddenly stabilizing the frequency becomes more challenging, because everybody has to work in exactly the same rhythm.

I wonder, would it make sense to change our power grid from AC to DC today? I know it would obviously be a lot of work, since every consuming device would have to change what power it accepts from the grid. But in the long run, could it be worth it? Also, what about insular networks. Would it make sense there? Thanks for taking the time for reading this, and also, I’m willing to go into the maths, if that’s relevant to the discussion.

  • gandalf_der_12te@lemmy.blahaj.zoneOP
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Okay, these are short term problems. “power lines as thick as your wrist” depend on the voltage. If voltage conversion works well enough, that issue disappears.

    But, take note of how many DC voltages you use in your house. Devices in mine range from 3v to 25v and some weird one like 19v for a laptop.

    Yeah, that’s why we need some kind of standard for these things.

    • Ebby@lemmy.ssba.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 months ago

      Ha! Yes! Even today USB 5 volts is pretty sweet for low power stuff. USB PD re-complicates things, but it’s not user dependent so that’s a plus.

      And you need a loooot of copper to prevent voltage drop especially when a grid of 100 houses 1/2 mile long draw 20-80 amps each. The math starts adding up real quick.

      • bastion@feddit.nl
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        3 months ago

        I mean, you need a lot of voltage to make voltage drop irrelevant. Like, 120 or 240 volts. If distribution is voltage is the same dc/ac, we could use the same wiring (but different breakers, and everything else).

        So the wiring argument doesn’t really hold up - the question is more about efficient converters to reduce voltage once it’s at the house.

        I.e., for typical American distribution, it’s 240 in the neighborhood and drops to 120 in the house. If the dc does the same, the same amount of power can be drawn along existing wires.