As we all know, AC won the “War of the Currents”. The reasoning behind this is that AC voltage is easy to convert up/down with just a ring of iron and two coils. And high voltage allows us to transport current over longer distances, with less loss.

Now, the War of the Currents happened in 1900 (approximately), and our technology has improved a lot since then. We have useful diodes and transistors now, we have microcontrollers and Buck/Boost converters. We can transform DC voltage well today.

Additionally, photovoltaics produces DC naturally. Whereas the traditional generator has an easier time producing AC, photovoltaic plants would have to transform the power into AC, which, if I understand correctly, has a massive loss.

And then there’s the issue of stabilizing the frequency. When you have one big producer (one big hydro-electric dam or coal power plant), then stabilizing the frequency is trivial, because you only have to talk to yourself. When you have 100000 small producers (assume everyone in a bigger area has photovoltaics on their roof), then suddenly stabilizing the frequency becomes more challenging, because everybody has to work in exactly the same rhythm.

I wonder, would it make sense to change our power grid from AC to DC today? I know it would obviously be a lot of work, since every consuming device would have to change what power it accepts from the grid. But in the long run, could it be worth it? Also, what about insular networks. Would it make sense there? Thanks for taking the time for reading this, and also, I’m willing to go into the maths, if that’s relevant to the discussion.

  • SomeoneSomewhere@lemmy.nz
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    PV inverters often have around 1-2% losses. This is not very significant. You also need to convert the voltage anyway because PV output voltage varies with light level.

    Buck/boost converters work by converting the DC current to (messy) AC, then back to DC. If you want an isolating converter (necessary for most applications for safety reasons) that converter needs to handle the full power. If it’s non isolating, then it’s proportional to the voltage step.

    Frequency provides a somewhat convenient method for all parties to know whether the grid is over- or under- supplied on a sub-second basis. Operating solely on voltage is more prone to oscillation and requires compensation for voltage drop, plus the information is typically lost at buck/boost sites. A DC grid would likely require much more robust and faster real-time comms.

    The AC grid relies on significant (>10x overcurrent) short-term (<5s) overload capability. Inrush and motor starting requires small/short overloads (though still significant). Faults are detected and cleared primarily through the excess current drawn. Fuses/breakers in series will all see the same current from the same fault, but we want only the device closest to the fault to operate to minimise disruption. That’s achieved (called discrimination, coordination, or selectivity) by having each device take progressively more time to trip on a fault of a given size, and progressively higher fault current so that the devices upstream still rapidly detect a fault.

    RCDs/GFCIs don’t coordinate well because there isn’t enough room between the smallest fault required to be detected and the maximum disconnection time to fit increasingly less sensitive devices.

    Generators are perfectly able to provide this extra fault current through short term temperature rise and inertia. Inverters cannot provide 5-fold overcurrent without being significantly oversized. We even install synchronous condensers (a generator without any actual energy source) in areas far from actual generators to provide local inertia.

    AC arcs inherently self-extinguish in most cases. DC arcs do not.

    This means that breakers and expulsion type fuses have to be significantly, significantly larger and more expensive. It also means more protection is needed against arcs caused by poor connection, cable clashes, and insulation damage.

    Solid state breakers alleviate this somewhat, but it’s going to take 20+ years to improve cost, size, and power loss to acceptable levels.

    I expect that any ‘next generation’ system is likely to demand a step increase in safety, not merely matching the existing performance. I suspect that’s going to require a 100% coverage fibre comms network parallel to the power conductors, and in accessible areas possibly fully screened cable and isolated supply.

    EVs and PV arrays get away with DC networks because they’re willing to shut down the whole system in the event of a fault. You don’t want a whole neighborhood to go dark because your neighbour’s cat gnawed on a laptop charger.

    • gandalf_der_12te@lemmy.blahaj.zoneOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Thank you for this well-thought and balanced viewpoint. It took me 19 days to process all the information.

      So basically, I was wrong when I assumed that inverters had an efficiency of around 50%. That misunderstanding comes from the phrase that “filters in the inverter eliminate high-frequency components in the PWM’s output”. I thought they discard that power. But that’s apparently not the case. So the efficiency is more like >95%. So that’s good.

  • Petter1@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I learned in school, that these days, DC would in theory indeed be more efficient even if you generate a sinus at home for the then legacy devices.

  • TimeSquirrel@kbin.melroy.org
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    When you have 100000 small producers (assume everyone in a bigger area has photovoltaics on their roof), then suddenly stabilizing the frequency becomes more challenging, because everybody has to work in exactly the same rhythm.

    That’s why you have standards and codes, that ensure everybody’s equipment is capable of syncing to the grid properly before they are allowed to connect. It’s not that hard for an inverter to do. Then you have the constant background supply to stabilize it like battery farms and other energy storage technologies. And a bunch of capacitor banks to correct power factor issues.

    I don’t think we are getting away from centralized production anytime soon. Even with the move to wind and solar, although I think nuclear should be included in that mix.

  • aaaaace@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I lived un a house that had live 120v DC service.

    There was an electric fan that ran on it. The outlets were only in the basement and identical to each other.

  • gandalf_der_12te@lemmy.blahaj.zoneOP
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I’m thinking there would be a transition period, where some devices would accept AC, and the others DC.

    There will be two different types of power socket outlets, and two converters to convert between AC/DC or DC/AC. (Rectifier and Inverter)

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    DC is used for long-range transmission in high-voltage DC (HVDC) transmission lines today.

    https://en.wikipedia.org/wiki/High-voltage_direct_current

    high-voltage direct current (HVDC) electric power transmission system uses direct current (DC) for electric power transmission, in contrast with the more common alternating current (AC) transmission systems. Most HVDC links use voltages between 100 kV and 800 kV.

    HVDC lines are commonly used for long-distance power transmission, since they require fewer conductors and incur less power loss than equivalent AC lines. HVDC also allows power transmission between AC transmission systems that are not synchronized. Since the power flow through an HVDC link can be controlled independently of the phase angle between source and load, it can stabilize a network against disturbances due to rapid changes in power. HVDC also allows the transfer of power between grid systems running at different frequencies, such as 50 and 60 Hz. This improves the stability and economy of each grid, by allowing the exchange of power between previously incompatible networks.

    However, since grids are AC, it’s just to send power to a grid or pull from one.

    We also do have some increasingly beefy DC in individual households in some forms:

    • You mention solar PV systems, but more generally, 12V systems used in vehicles (and the related 24V and 48V systems that are sometimes used to push more power) are more common, with lithium batteries that can do many more charge cycles than lead-acid being available.

    • USB PD can negotiate pushing up to 240W now at 48V, which is a fair bit.

  • empireOfLove2@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Well, most all DC generators these days are actually AC alternators with the output rectified, because alternators can be run a lot more efficiently. So you’re already losing on efficiency there.

    You need to consider the consumer side as well. Dinky residential loads like your computer would be fine on DC. But most of the world, especially heavy industry, runs on synchronous or induction AC motors, big ones. Big huge tens-of-megawatts motors that often run upwards of 97% line efficiency, which is insane for any industrial process.
    The best you could replace those with would be modern brushless DC motors, which require really expensive inverter controls that die frequently due to the magnetic transients and still top out at an efficiency of only 90% if you’re lucky. And that would incur huge costs that just aren’t worth it.

  • deegeese@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    How does the efficiency and cost of buck converters compare to AC transformers? Seems like the cost and efficiency of the voltage converter should be the prime determinant, rather than specific applications of generation/consumption.

    What would a 400A 10kV utility scale DC converter look like?

    • gandalf_der_12te@lemmy.blahaj.zoneOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 months ago

      well, a large part why I asked the question is because I hope that somebody knows more about what buck/boost-converters can do today. I know they work well enough on small scales, but I have no experimental data for them on larger scales.

      I assume they would work well, but I’d like that somebody links me to the right datasheet or something.

      Edit: you have a very important point there. " Seems like the […] voltage converter should be the prime determinant, rather than specific applications of generation/consumption." YES. So, let me rephrase my question: does anybody have experience with high-power DC voltage converters?

  • Ebby@lemmy.ssba.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I heard it said many years ago that if DC won the battle, we’d have power stations every 10 miles and power lines as thick as your wrist.

    Converting local power is fairly easy, with AC inverters added for universal compatibility.

    But, take note of how many DC voltages you use in your house. Devices in mine range from 3v to 25v and some weird one like 19v for a laptop. You’d still have adapters all over the place.

    • gandalf_der_12te@lemmy.blahaj.zoneOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Okay, these are short term problems. “power lines as thick as your wrist” depend on the voltage. If voltage conversion works well enough, that issue disappears.

      But, take note of how many DC voltages you use in your house. Devices in mine range from 3v to 25v and some weird one like 19v for a laptop.

      Yeah, that’s why we need some kind of standard for these things.

      • Ebby@lemmy.ssba.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Ha! Yes! Even today USB 5 volts is pretty sweet for low power stuff. USB PD re-complicates things, but it’s not user dependent so that’s a plus.

        And you need a loooot of copper to prevent voltage drop especially when a grid of 100 houses 1/2 mile long draw 20-80 amps each. The math starts adding up real quick.

        • bastion@feddit.nl
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 months ago

          I mean, you need a lot of voltage to make voltage drop irrelevant. Like, 120 or 240 volts. If distribution is voltage is the same dc/ac, we could use the same wiring (but different breakers, and everything else).

          So the wiring argument doesn’t really hold up - the question is more about efficient converters to reduce voltage once it’s at the house.

          I.e., for typical American distribution, it’s 240 in the neighborhood and drops to 120 in the house. If the dc does the same, the same amount of power can be drawn along existing wires.

          • Quatlicopatlix@feddit.org
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            Yea have fun transmitting a decent amount of power with 240v over a meaningfull distance. Also most generators produce ac anyways so why would you recitify it at the generator instead of your device after a transformer? You still need all kinds of different voltages everywhere in your electronics and this means you still need to regulate it.

            I am not shure how the american wirering worls out but to get from 240 to 120 you still need a transformer… or is it 240v between the different phases and then 120 from phase to neutral?

            • bastion@feddit.nl
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 months ago

              240 in the neighborhood - i.e., that’s enough to distribute from the pole to a few houses. Of course you have higher voltages to go longer distances. This is equally true for AC vs DC. Thus, the idea that it takes a looot of copper for DC is erroneous.

              In fact, where conductor size is relevant is that you can use smaller conductors for DC, because of the skin effect.

              Wiring: Split phase, that is also usable as 240 for large appliances. So, the latter.

    • explore_broaden@midwest.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      But, take note of how many DC voltages you use in your house. Devices in mine range from 3v to 25v and some weird one like 19v for a laptop. You’d still have adapters all over the place.

      This is probably true, but every single one could lose the rectifier part, and instead of having to convert from pulsating DC (the output of mains rectification), you get clean DC from the wall instead, which should allow for using smaller capacitors in many places.

  • Shawdow194@fedia.io
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    Basically everything runs off 3.7-12v DC at the end of the day

    It does make sense to eventually phase out AC in most home and commercial applications

    • CameronDev@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      All the small stuff is low voltage DC, but just about every appliance requires AC (ovens, dishwashers, kettles, toasters, washing machines, aircon). Running an oven on 12v DC would be insane.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Most of those don’t actually want AC. They just want higher voltage.

        Ovens are just running resistive heating. They don’t need AC.

        Dishwashers might, if the motor is AC.

        Kettles don’t care. Same resistive heating thing as ovens.

        Toasters don’t care. Same again.

        Washing machines and air conditioners might have motors that run on AC.

        All that being said, I don’t know how problematic it’d be to use DC motors.

        Most fans in your house have fixed or several speeds, because they just run the motor off AC directly. I got a DC floor fan recently, and it was kind of nice to have fine-grained control over the speed, something that you can’t do with an AC fan.

        kagis

        https://www.tenmenelectrical.com/ac-vs-dc-ceiling-fans/

        Differences between AC and DC ceiling fans

        Now let’s dive into the differences between AC and DC ceiling fans.

        Energy Efficiency

        One of the most significant differences between AC and DC ceiling fans is their energy efficiency. DC motors are more energy-efficient than AC motors and consume up to 70% less power. DC motors also convert more of the energy they consume into rotational energy, while AC motors waste energy in the form of heat. This means that DC motors will save you money on your electricity bill in the long run.

        Speed Settings

        Both AC and DC ceiling fans can have multiple speed settings, but DC fans offer greater control over the fan’s speed. DC motors are usually controlled using a remote control, allowing you to adjust more speed settings of the fan, usually up to 6 speeds. AC fans, on the other hand, typically use a wall controller but can be fitted with a remote control. However, the settings are set to low, medium and high only, which limits the number of speed settings and the ability to adjust the speed precisely.

        Noise

        DC fans are generally quieter than AC fans due to how the motors operate. DC motors operate without the humming sound that AC motors create, making them ideal for bedrooms or other quiet spaces.

        Lifespan

        DC motors are also more durable than AC motors, as they have fewer moving parts and operate at lower temperatures. AC motors, on the other hand,  use capacitors to control fan speeds, which have a limited lifespan. When an AC fan loses the ability to change speeds, it is usually due to one of the capacitors failing.

        Price

        DC ceiling fans are more expensive than AC ceiling fans due to their higher energy efficiency, more advanced technology, and quieter operation. However, DC fans’ energy efficiency means that they will save you money on your electricity bill in the long run.

        That being said, they sell fans, so I imagine that they’d rather sell a fancier fan.