• esaru@beehaw.org
    link
    fedilink
    arrow-up
    0
    ·
    3 months ago

    I wonder if the intention could be purely financial as well: Once the backdoor is in, you can sell it for millions to a company which sells the service of breaking into systems. Or it was such a company itself that pushed the backdoor into the code.

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Looking forward, I’m still more-worried about the fact that state-backed threat actors are targeting open source projects via this social engineering route than the technical issues.

    I think that the technical issues that the attacker used can be addressed to at least some degree.

    • If autoconf is a good place to hide stuff, maybe shift away from distributing autoconf-generated files in the source tarball; I don’t know all of the problems that come up with that, but I understand that at least some of them were backwards-compatibility-breaking issues across autoconf versions in the past.

    • If maintainers are putting up bogus tarballs that differ from what’s in git, have github disallow this, and for projects that are currently doing it, have github figure out how to address their needs otherwise. If this can’t be done, at the very least, highlight the differences.

    • The ifunc hooks can probably have some kind of automated auditing added, if they’re a favored vector to hide things.

    • If, in a project takeover, attackers try to embargo any reporting of security holes and ask that nobody but them be notified, it’d be possible to have some trusted third-party notified.

    • If threat actors try to inject attacks right before freeze – I understand that Ubuntu may have been a target – then apply greater scrutiny to those changes.

    • If distros linking code into sshd with distro patches is exposing sshd to security threats that the opensshd team isn’t looking at, disallow that practice in distro policy for some set of security-sensitive projects.

    • Systemd may be too ready to link libraries into things that don’t require it.

    Maybe it makes sense to have a small number of projects that are considered “security-critical” and then require that they only rely on other projects that are also security-critical. That’s not a magic fix, but it maybe tamps down on the damage a supply-chain attack could cause. Still…my suspicion is that if an attacker could get code into something like xz, they could probably ultimately, even with only user-level privileges, figure out ways to escalate to control of a system. I mean, all you need is a user with admin privileges to run something as a user anywhere with their account. Maybe Linux and some other software projects just fundamentally don’t have enough isolation. That is, maybe the typical software package should be expected to run in a sandbox, kind of the way smartphone software or video game console software does. That doesn’t solve everything, but it at least reduces the attack surface.

    But the social side of this is a pain. We don’t want to break down the system of trust that lets open-source work well more than is necessary…but clearly, there are people being attacked by people who have a lot of time to spend on putting together tactics to attack them. I’m not sure that your typical open-source maintainer – health issues or no – can realistically constantly be on their guard against coordinated social engineering attacks.

    The attacker came via a VPN (well, unless they messed up) and had no history. The (probable) sockpuppets also had no history. It might be a good idea to look for people entering open source projects who have no history and are only visible from a VPN…but my guess is that if we rely on reputation more, attackers will just seek to subvert that as well. In this case, they probably committed bogus commits for the purpose of building reputation for years. If you’re willing to put three years into building reputation on a given project, I imagine that you can do something similar to have an account lying in wait for the next open source project to attack. And realistically, my guess is that if we trust non-VPN machines, a state-backed attacker could get ahold of one…it’s maybe more convenient for them to bounce through a VPN. It’s not something that they absolutely have to do.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        3 months ago

        desktop

        Some of it is that a lot of desktop software paradigms weren’t built to operate in that kind of environment, and you can’t just break backwards compatibility without enormous costs.

        Wayland’s been banging on that, but there’s a lot to change.

        Like, the clipboard is designed so that software packages can query its contents, rather than having the contents pushed to it.

        What’s on the screen and a lot of system state like keys that are down and where the mouse pointer is and so forth wasn’t treated as information that needed to be kept private from an application.

        I don’t think that there’s a great way to run isolated game-level 3d graphics in a VM unless you’re gonna have separate hardware.

        Something that I’ve wondered about is potential vulnerability via Steam. None of the software there is isolated in a “this might be malicious” sense – not from the rest of the system, not from other software sold via Steam. And Steam is used to distribute free software…I haven’t looked into it, but I don’t think that the bar to get something into Steam is likely super high. And then consider that there are free-to-play games that have to make money however they can, and some of that is going to be selling data, and some of how they do that may be to just offer to run whatever libraries with their game the highest bidder offers. How secure are those supply chains? And on Steam, most of the software is closed source, which makes inspecting what’s going on harder. And that’s before we even get to mods and stuff like that, which are from all over the place.

        I mean, let’s say that random library from ad company used by a free-to-play game is sending up the identity of the user on the computer. It has some functionality that slurps in a payload from the network telling it to grab credentials off the existing system, and does so for ten critical users. Would anyone notice? I have a really hard time believing that there’d be any way to pick up on that. Even if you wanted to, you can’t isolate many of these games from the network without breaking their functionality, and there’s no mechanism in in place today isolating them from the user’s storage or other identity information.

    • jmcs@discuss.tchncs.de
      link
      fedilink
      arrow-up
      0
      ·
      3 months ago

      That’s also how the most damaging attacks on proprietary software work. At some point all organizations need to trust their members and co-workers need to trust each other - I can’t think of a way to be more miserable at work than having to second guess everyone around you.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 months ago

        That’s also how the most damaging attacks on proprietary software work.

        Yeah, supply chain attacks can happen. There was that infamous SolarWinds supply chain attack recently. But I think that there are some important mitigating factors there.

        • Proprietary software companies – unless they’re using something open-source like xz upstream in their supply chain, as it’s not just a “proprietary software world” and “open-source software world” – tend to have someone’s personal information if they’re employed by them. They’re not gonna hire and pay some random name who they know only as a GitHub account through a VPN, certainly not make them maintainer of their software.

        • Many – not all – proprietary software companies mandate that employees work locally. I’s likely that if I’m working for a US company, a person is also subject to US law enforcement. In contrast, if you have a state-backed group, they’re probably targeting people elsewhere. Whoever the people from the Jia Tan group are, my guess is that it’s good odds that they will probably aim to avoid being in a country that they are targeting. Even if we expose their identities, they probably aren’t going to be directly-impacted by law enforcement. Open source projects hypothetically could do that, I suppose, but normally they’re pretty border-agnostic.

        That is, I think that this is going to be specially a challenge for the open-source world, as the attacks are targeting some things that the open-source community is notable for – border-agnosticism, a relatively-low bar to join a project, and often not a lot of personal identity validation.

        At some point all organizations need to trust their members and co-workers need to trust each other - I can’t think of a way to be more miserable at work than having to second guess everyone around you.

        Yeah, that’s kinda what I was thinking, but you put it more-frankly.

        It seems like there’s a lot of potential for this to be corrosive to the community.