• TotalSonic@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    Obviously floating point is of huge benefit for many audio dsp calculations, from my observations (non-programmer, just long time DAW user, from back in the day when fixed point with relatively low accumulators was often what we had to work with, versus now when 64bit floating point for processing happens more as the rule) - e.g. fixed point equalizers can potentially lead to dc offset in the results. I don’t think peeps would be getting as close to modeling non-linear behavior of analog processors with just fixed point math either.

    • ExFed@lemm.ee
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      Audio, like a lot of physical systems, involve logarithmic scales, which is where floating-point shines. Problem is, all the other physical systems, which are not logarithmic, only get to eat the scraps left over by IEEE 754. Floating point is a scam!

  • Blackmist@feddit.uk
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    I know this is in jest, but if 0.1+0.2!=0.3 hasn’t caught you out at least once, then you haven’t even done any programming.

    • dejected_warp_core@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      There are probably a lot of scientific applications (e.g. statistics, audio, 3D graphics) where exponential notation is the norm and there’s an understanding about precision and significant digits/bits. It’s a space where fixed-point would absolutely destroy performance, because you’d need as many bits as required to store your largest terms. Yes, NaN and negative zero are utter disasters in the corners of the IEEE spec, but so is trying to do math with 256bit integers.

      For a practical explanation about how stark a difference this is, the PlayStation (one) uses an integer z-buffer (“fixed point”). This is responsible for the vertex popping/warping that the platform is known for. Floating-point z-buffers became the norm almost immediately after the console’s launch, and we’ve used them ever since.

      • ripcord@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        4 months ago

        That doesn’t really answer the question, which is about the origins of the meme templete

  • dejected_warp_core@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    There are probably a lot of scientific applications (e.g. statistics, audio, 3D graphics) where exponential notation is the norm and there’s an understanding about precision and significant digits/bits. It’s a space where fixed-point would absolutely destroy performance, because you’d need as many bits as required to store your largest terms. Yes, NaN and negative zero are utter disasters in the corners of the IEEE spec, but so is trying to do math with 256bit integers.

    For a practical explanation about how stark a difference this is, the PlayStation (one) uses an integer z-buffer (“fixed point”). This is responsible for the vertex popping/warping that the platform is known for. Floating-point z-buffers became the norm almost immediately after the console’s launch, and we’ve used them ever since.

    • anton@lemmy.blahaj.zone
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      What’s the problem with -0?
      It conceptually makes sense for to negativ values to close to 0 to be represented as -0.
      In practice I have never seen a problem with -0.

      On NaN: While its use cases can nowadays be replaced with language constructs like result types, it was created before exceptions or sum types. The way it propagates kind of mirrors Haskells monadic Maybe.
      We should be demanding more and better wrapper types from our language/standard library designers.

  • Worx@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    The only reason for floating point numbers is to use your laptop as a life buoy

  • arc@lemm.ee
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    4 months ago

    The problem is, that most languages have no native support other than 32 or 64 bit floats and some representations on the wire don’t either. And most underlying processors don’t have arbitrary precision support either.

    So either you choose speed and sacrifice precision, or you choose precision and sacrifice speed. The architecture might not support arbitrary precision but most languages have a bignum/bigdecimal library that will do it more slowly. It might be necessary to marshal or store those values in databases or over the wire in whatever hacky way necessary (e.g. encapsulating values in a string).

  • NocturnalMorning@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    Uhm, I haven’t programmed in a low level language in years. I use python for my job now, and all I know are floats and ints. I don’t know what this foreign language is you speak of.

    • YTG123@feddit.ch
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      4 months ago

      Perhaps you can encode them as computation (i.e. a function of arbitrary precision)

    • Chadus_Maximus@lemm.ee
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      4 months ago

      May I propose a dedicated circuit (analog because you can only ever approximate their value) that stores and returns transcendental/irrational numbers exclusively? We can just assume they’re going to be whatever value we need whenever we need them.

      • frezik@midwest.social
        link
        fedilink
        arrow-up
        0
        ·
        4 months ago

        Wouldn’t noise in the circuit mean it’d only be reliable to certain level of precision, anyway?

    • smeg@feddit.uk
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      Do we even have a good way of encoding them in real life without computers?

  • Ephera@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    I have been thinking that maybe modern programming languages should move away from supporting IEEE 754 all within one data type.

    Like, we’ve figured out that having a null value for everything always is a terrible idea. Instead, we’ve started encoding potential absence into our type system with Option or Result types, which also encourages dealing with such absence at the edges of our program, where it should be done.

    Well, NaN is null all over again. Instead, we could make the division operator an associated function which returns a Result<f64> and disallow f64 from ever being NaN.

    My main concern is interop with the outside world. So, I guess, there would still need to be a IEEE 754 compliant data type. But we could call it ieee_754_f64 to really get on the nerves of anyone wanting to use it when it’s not strictly necessary.

    Well, and my secondary concern, which is that AI models would still want to just calculate with tons of floats, without error-handling at every intermediate step, even if it sometimes means that the end result is a shitty vector of NaNs, that would be supported with that, too.

    • RustyNova@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      While I get your proposal, I’d think this would make dealing with float hell. Do you really want to .unwrap() every time you deal with it? Surely not.

      One thing that would be great, is that the / operator could work between Result and f64, as well as between Result and Result. Would be like doing a .map(|left| left / right) operation.

    • Lmaydev@programming.dev
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      Nan isn’t like null at all. It doesn’t mean there isn’t anything. It means the result of the operation is not a number that can be represented.

      The only option is that operations that would result in nan are errors. Which doesn’t seem like a great solution.

        • Lmaydev@programming.dev
          link
          fedilink
          arrow-up
          0
          ·
          4 months ago

          Float processing is at the hardware level. It needs a way to signal when an unrepresented value would be returned.

          • Ephera@lemmy.ml
            link
            fedilink
            arrow-up
            0
            ·
            4 months ago

            My thinking is that a call to the safe division method would check after the division, whether the result is a NaN. And if it is, then it returns an Error-value, which you can handle.

            Obviously, you could do the same with a NaN by just throwing an if-else after any division statement, but I would like to enforce it in the type system that this check is done.

            • Lmaydev@programming.dev
              link
              fedilink
              arrow-up
              0
              ·
              edit-2
              4 months ago

              I feel like that’s adding overhead to every operation to catch the few operations that could result in a nan.

              But I guess you could provide alternative safe versions of float operations to account for this. Which may be what you meant thinking about it lol

      • Ephera@lemmy.ml
        link
        fedilink
        arrow-up
        0
        ·
        4 months ago

        Well, that is what I meant. That NaN is effectively an error state. It’s only like null in that any float can be in this error state, because you can’t rule out this error state via the type system.

        Why do you feel like it’s not a great solution to make NaN an explicit error?

        • CapeWearingAeroplane@sopuli.xyz
          link
          fedilink
          arrow-up
          0
          ·
          4 months ago

          Theres plenty of cases where I would like to do some large calculation that can potentially give a NaN at many intermediate steps. I prefer to check for the NaN at the end of the calculation, rather than have a bunch of checks in every intermediate step.

          How I handle the failed calculation is rarely dependent on which intermediate step gave a NaN.

          This feels like people want to take away a tool that makes development in the engineering world a whole lot easier because “null bad”, or because they can’t see the use of multiplying 1e27 with 1e-30.

          • Ephera@lemmy.ml
            link
            fedilink
            arrow-up
            0
            ·
            4 months ago

            Well, I’m not saying that I want to take tools away. I’m explicitly saying that a ieee_754_f64 type could exist. I just want it to be named annoyingly, so anyone who doesn’t know why they should use it, will avoid it.

            If you chain a whole bunch of calculations where you don’t care for NaN, that’s also perfectly unproblematic. I just think, it would be helpful to:

            1. Nudge people towards doing a NaN check after such a chain of calculations, because it can be a real pain, if you don’t do it.
            2. Document in the type system that this check has already taken place. If you know that a float can’t be NaN, then you have guarantees that, for example, addition will never produce a NaN. It allows you to remove some of the defensive checks, you might have felt the need to perform on parameters.

            Special cases are allowed to exist and shouldn’t be made noticeably more annoying. I just want it to not be the default, because it’s more dangerous and in the average applications, lots of floats are just passed through, so it would make sense to block NaNs right away.

    • xmunk@sh.itjust.works
      link
      fedilink
      arrow-up
      0
      ·
      4 months ago

      I agree with moving away from floats but I have a far simpler proposal… just use a struct of two integers - a value and an offset. If you want to make it an IEEE standard where the offset is a four bit signed value and the value is just a 28 or 60 bit regular old integer then sure - but I can count the number of times I used floats on one hand and I can count the number of times I wouldn’t have been better off just using two integers on -0 hands.

      Floats specifically solve the issue of how to store a ln absurdly large range of values in an extremely modest amount of space - that’s not a problem we need to generalize a solution for. In most cases having values up to the million magnitude with three decimals of precision is good enough. Generally speaking when you do float arithmetic your numbers will be with an order of magnitude or two… most people aren’t adding the length of the universe in seconds to the width of an atom in meters… and if they are floats don’t work anyways.

      I think the concept of having a fractionally defined value with a magnitude offset was just deeply flawed from the get-go - we need some way to deal with decimal values on computers but expressing those values as fractions is needlessly imprecise.

  • davidgro@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    Serious answer: Posits seem cool, like they do most of what floats do, but better (in a given amount of space). I think supporting them in hardware would be awesome, but of course there’s a chicken and egg problem there with supporting them in programming languages.

    • Kodiack@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      4 months ago

      I had the great honour of seeing John Gustafson give a presentation about unums shortly after he first proposed posits (type III unums). The benefits over floating point arithmetic seemed incredible, and they seemed largely much more simple.

      I also got to chat with him about “Gustafson’s Law”, which kinda flips Amdahl’s Law on its head. Parallel computing has long been a bit of an interest for me I was also in my last year of computer science studies then and we were covering similar subjects at the time. I found that timing to be especially amusing.

    • Quetzalcutlass@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      4 months ago

      Posits aside, that page had one of the best, clearest explanations of how floating point works that I’ve ever read. The authors of my college textbooks could have learned a thing or two about clarity from this writer.