• sp3ctr4l@lemmy.zip
    link
    fedilink
    arrow-up
    65
    arrow-down
    5
    ·
    14 days ago

    … Are the Feds aware that the core systems that many, many older companies (and government agencies) use are still based on COBOL?

    Is… is that not of any concern?

    • aubeynarf@lemmynsfw.com
      link
      fedilink
      arrow-up
      22
      ·
      13 days ago

      Is COBOL subject to buffer overflows and use-after-free bugs? I honestly don’t know.

      I don’t recall the COBOL code I’ve read using pointers.

      • sp3ctr4l@lemmy.zip
        link
        fedilink
        arrow-up
        13
        ·
        13 days ago

        The problem I am aware of is moreso that the number of programmers that know COBOL is vanishingly small, it … COBOL does not seem to really be taught anymore…

        …so if something goes wrong at that level, you may be SOL if you cannot find an increasingly rare programmer that knows COBOL well.

  • wewbull@feddit.uk
    link
    fedilink
    English
    arrow-up
    69
    arrow-down
    29
    ·
    14 days ago

    That sounds like policy written by somebody who has no idea what the reality of software development is.

    1 year to rewrite critical software in a new language?

    • nous@programming.dev
      link
      fedilink
      English
      arrow-up
      81
      arrow-down
      3
      ·
      13 days ago

      Did you read the article at all?

      “Putting all new code aside, fortunately, neither this document nor the U.S. government is calling for an immediate migration from C/C++ to Rust — as but one example,” he said. “CISA’s Secure by Design document recognizes that software maintainers simply cannot migrate their code bases en masse like that.”

      Companies have until January 1, 2026, to create memory safety roadmaps.

      All they are asking for by that date is a roadmap for dealing with memory safety issues, not rewrite everything.

  • mox@lemmy.sdf.org
    link
    fedilink
    arrow-up
    24
    ·
    edit-2
    13 days ago

    Don’t assume too much from the headline, folks. They’re not saying everything has to be rewritten by 2026. They’re saying new product lines serving critical infrastructure should be written in memory-safe languages, and existing ones should have a memory safety roadmap.

    If you’re about to post about how you think that’s unreasonable, I think you should explain why.

    • BlazeDaley@lemmy.world
      link
      fedilink
      arrow-up
      21
      arrow-down
      1
      ·
      13 days ago

      It’s one backed by a lot of data. One example is from the Android project.

      The percent of vulnerabilities caused by memory safety issues continues to correlate closely with the development language that’s used for new code. Memory safety issues, which accounted for 76% of Android vulnerabilities in 2019, and are currently 24% in 2024, well below the 70% industry norm, and continuing to drop.

      https://security.googleblog.com/2024/09/eliminating-memory-safety-vulnerabilities-Android.html

      There’s an argument that critical infrastructure software vendors are already meeting standards for basic, non-memory related items. Yes, there are other categories, but memory safety is one that’s harder to verify. Moving to memory safe languages is an ensure a category of correctness. This excludes usage of unsafe escape hatches.

  • perviouslyiner@lemmy.world
    link
    fedilink
    arrow-up
    26
    arrow-down
    16
    ·
    edit-2
    13 days ago

    Seems excessive to convert everything to rust when you can use std::shared_ptr and std::weak_ptr to eliminate the memory safety issue?

    • arendjr@programming.devOP
      link
      fedilink
      arrow-up
      31
      ·
      edit-2
      13 days ago

      Using smart pointers doesn’t eliminate the memory safety issue, it merely addresses one aspect of it. Even with smart pointers, nothing is preventing you from passing references and using them after they’re freed.

      • refalo@programming.dev
        link
        fedilink
        arrow-up
        2
        arrow-down
        12
        ·
        13 days ago

        To be fair, it’s entirely possible to make the same and very similar mistakes in Rust, too.

        • GetOffMyLan@programming.dev
          link
          fedilink
          arrow-up
          8
          arrow-down
          1
          ·
          edit-2
          13 days ago

          I’m fairly sure use after free isn’t possible unless you explicitly use unsafe code right?

          It’s compiler enforced is the point.

        • robinm@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          10 days ago

          Is it possible to do in Rust?

          Yes

          Is possible to do in Rust, by mistake, and not easily caught by a review?

          Definitively not.

            • robinm@programming.dev
              link
              fedilink
              arrow-up
              1
              ·
              5 days ago
              void foo() {
                  std::vector v = {0, 1, 2, 4};
                  const auto& ref = v[1];
                  add_missing_values(v);
                  std::cout << ref << "\n";
              }
              
              void add_missing_values(std::vector<int>& v) {
                  // ...
                  v.push_back(3);
              }
              

              Neither foo(), nor add_missing_values() looks suspicious. Nonetheless, if v.push_back(3) requires v to grow, then ref becomes an invalid reference and std::cout << ref becomes UB (use after free). In Rust this would not compiles.

              It is order of magnitudes easier to have lifetime errors in C++ than in Rust (use after free, double free, data races, use before initialisation, …)

                • robinm@programming.dev
                  link
                  fedilink
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  3 days ago

                  I think you have a hard time understanding the différence between “not possible” and “much harder”.

                  In Rust, the code does not compile.

                  In C++ the code compile, but

                  • if you have a test case
                  • this test case triggers the bug (it is not guarateed to properly reproduce you production environment since it depends on the parameters of the allocator of your vector)
                  • you use ubsan

                  … then the bug will be caught.

                  Yes it is possible, noone says the opposite. But you can’t deny it’s harder. And because its harder, more bugs get past review, most notably security bugs as demonstrated again and again in many studies. The

                • robinm@programming.dev
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  5 days ago

                  That’s why I did not said it was impossible, just order of magnitude harder to catch in C++ compared to Rust.

                  To have asan finding the bug, you need to have a valid unit test, that has a similar enough workload. Otherwise you may not see the bug with asan if the vector doesn’t grow (and thus ref would still be valid, not triggering UB), leading to a production-only bug.

                  Asan is a wonderfull tool, but you can’t deny it’s much harder to use and much less reliable than just running your compiler.

    • psycotica0@lemmy.ca
      link
      fedilink
      arrow-up
      12
      ·
      13 days ago

      I get what you’re saying, but I think the issue with optional memory safety features is that it’s hard to be sure you’re using it in all the places and hard to maintain that when someone can add a new allocation in the future, etc. It’s certainly doable, and maybe some static analysis tools out there can prove it’s all okay.

      Whereas with Rust, it’s built from the ground up to prove exactly that, plus other things like no memory being shared between threads by accident etc. Rust makes it difficult and obvious to do the wrong thing, rather than that being the default.

    • magic_lobster_party@fedia.io
      link
      fedilink
      arrow-up
      8
      ·
      13 days ago

      From the original document:

      Software manufacturers should build products in a manner that systematically prevents the introduction of memory safety vulnerabilities, such as by using a memory safe language or hardware capabilities that prevent memory safety vulnerabilities. Additionally, software manufacturers should publish a memory safety roadmap by January 1, 2026.

      My interpretation is that smart pointers are allowed, as long it’s systematically enforced. Switching to a memory safe language is just one example.

  • Primer - Zip@lemmy.zip
    link
    fedilink
    English
    arrow-up
    2
    ·
    13 days ago

    Just from reading the article, is the scope just critical software infrastructure? What does that encompass exactly? Banking and military software seems easy to assume - what about embedded medical device software? Or just embedded software in general?