• @frunch
    link
    231 year ago

    What was wrong with them? They served their purpose just fine for many years

    • @marx2k
      link
      1041 year ago

      The weighed a ton, they were limited in size, their resolution was terrible, they sucked down electricity…

      • @[email protected]
        link
        fedilink
        561 year ago

        Their screen was curved the wrong way until they released flat screen TVs

        4:3 resolution meant you lost some of the content from movies or you watched them with black bars

        • @Blue_Morpho
          link
          221 year ago

          Except movies keep changing so now if you want imax at home you need 4:3.

          Whatever isn’t available at home is what movies will change to to keep themselves unique.

          • @[email protected]
            link
            fedilink
            101 year ago

            Widescreen has been the movie industry standard for how many decades now? IMAX is its own beast but most movies aren’t filmed in real IMAX resolution and now there’s digital IMAX which is basically 19:10 which is the same as many TVs…

            • @Blue_Morpho
              link
              01 year ago

              Movies used to be all 4:3 before tv. It’s called the academy ratio. Movies now do 1.85:1 and even 2.39:1. A few even do anamorphic 2.76:1. Anything but the dominant home format.

              • @[email protected]
                link
                fedilink
                71 year ago

                Major movie studios have mostly used widescreen since the 1950s and all the different ratios you mentioned except 4:3 are better watched on a widescreen TV than a 4:3 TV.

        • @[email protected]
          link
          fedilink
          01 year ago

          4:3 resolution also means that a lot of good shows will never be watchable in the proper 16:9 format

          • @uis
            link
            21 year ago

            No, it means 4:3 IS proper format

      • @[email protected]
        link
        fedilink
        61 year ago

        We had four channels and loved it!

        And most people were lucky to have a TV. You were lucky to have a HOUSE! We used to live in one room, all hundred and twenty-six of us, no furniture. Half the floor was missing; we were all huddled together in one corner for fear of FALLING!

      • @[email protected]
        link
        fedilink
        -151 year ago

        Have you compared NES games on a CRT with the same games on a modern screen?

        CRTs just look miles better.

        • amio
          link
          fedilink
          33
          edit-2
          1 year ago

          EDIT: OK, it’s ackchually not technically “resolution” per se, I get it. :p

          That’s because the graphics were tailored to CRT resolution - which is to say, [things that just so happened to have] low/outright bad resolution.

          CRTs have advantages over more modern stuff but that’s mostly about latency.

          • @Sylvartas
            link
            291 year ago

            It’s not as much about resolution as it was about exploiting the quirks of CRT. Artists usually “squished” sprites horizontally (because crt screens would stretch them) and used the now famous “half dot” technique to have more subtle shading than what was actually possible at the pixel level. So if you just display the original sprites with no stretch and no “bleed” between pixels, it doesn’t look as good as it should.

            • Zaros
              link
              31 year ago

              I had never heard of this. Fascinating!

          • @grue
            link
            English
            81 year ago

            That’s because the graphics were tailored to CRT resolution - which is to say, low/outright bad resolution.

            No, it’s because the graphics were tailored to the analog characteristics of CRTs: things like having scanlines instead of pixels and bleed between phosphors. If they were only tailored to low resolution they’d look good on a low resolution LCD, but they don’t.

            I admit I’m quibbling, but the whole thread is that, so…

            • amio
              link
              fedilink
              31 year ago

              As a pedant, that is impressive work. Fair enough.

              • @grue
                link
                English
                11 year ago

                Thanks, but @Sylvartas’ reply (which I didn’t read until after writing mine) did a much better job, TBH.

          • @uis
            link
            21 year ago

            CRTs themselve don’t have concept of resolution.

          • @[email protected]
            link
            fedilink
            -21 year ago

            CRTs don’t have pixels so the resolution of the signal isn’t that important. It’s about the inherent softness you get from the technology. It’s better than any anti-aliasing we have today.

            • @[email protected]
              link
              fedilink
              2
              edit-2
              1 year ago

              CRTs do have pixels. If they didn’t, you could run an SVGA signal (800x600 at 60 Hz) directly into any CRT. If you tried this, it would likely damage the tube beyond repair.

              The exact mechanism varied between manufacturers and types: http://filthypants.blogspot.com/2020/02/crt-shader-masks.html

              I certainly saw aliasing problems on CRTs, though usually on computer monitors that had higher resolution and better connection standards. The image being inherently “soft” is related to limited resolution and shitty connections. SCART with RGB connections will bring out all the jagginess. The exact same display running on composite will soften it and make it go away, but at the cost of a lot of other things looking like shit.

              • @grue
                link
                English
                11 year ago

                CRTs do have pixels. If they didn’t, you could run an SVGA signal (800x600 at 60 Hz) directly into any CRT. If you tried this, it would likely damage the tube beyond repair.

                Would it, though? I’m skeptical.

                If it did, it wouldn’t be because they have “pixels,” though; it would be because overdriving the deflection yoke with higher-frequency signals would generate too much heat for the TV to handle.

                Otherwise (if it didn’t overheat), it should “work.” The result might look weird if the modulation of the signal didn’t line up with the apertures in the shadow mask right, but I don’t see any reason why sweeping the beam across faster would damage the phosphors. (Also, I’m not convinced a black & white TV would have any problem at all.)

                • @[email protected]
                  link
                  fedilink
                  1
                  edit-2
                  1 year ago

                  It will tend to turn the beam on when it’s off to the side, outside the normal range of the screen. X Windows users in the mid 90s had to put in their exact scanline information or else the screen could blow up. That went away with a combination of multiscan monitors and monitors being able to communicate their preferred settings, but those came pretty late in the CRT era.

                  Edit: in any case, color screens need to have at least bands of red/green/blue phosphor. At a minimum, there will be breaks along either the horizontal or vertical lines, if not both.

                  • @grue
                    link
                    English
                    11 year ago

                    When you say “blow up” do you mean the tube would literally explode, it would burn through phosphors, a circuit board would let the magic smoke out, or something else?

                    I remember configuring mode lines in X. Luckily, I never found out the hard way what happened if you got it wrong.

                  • @Aceticon
                    link
                    11 year ago

                    I remember doing that configuration for X.

                    You had to tell it when to change lines, when to start firing on a new line (i.e. you changed lines then waited a bit and only after started sending data), then when to stop firing and finally another wait after which a change lines (so there was some “empty” at the start and at the end of a horizontal electron gun trace and you had to tune those so that the image didn’t start or end outside the screen).

                    There was also something similar for the vertical axis - i.e. instructing it to go back to the top and some empty lines at the top and the bottom.

                    I wouldn’t say it was convenient (it was a bloody text file with weird-looking numbers and you did run the risk of blowing up the CRT), but it was kinda fun that you could create your own crazy screen resolutions for X once you understood the principle of the thing.

              • @grue
                link
                English
                11 year ago

                That image is a digital rendering of the raw data, not a photo of how a CRT would render it.

                CRTs were nothing if not the opposite of jagged.

                  • @grue
                    link
                    English
                    1
                    edit-2
                    1 year ago

                    “Blurred” is the opposite of “jagged,” though.

                    The jaggedness of the 2600 wasn’t because the TV itself was jagged; it was because the 2600 was so low-resolution (160x192, maximum) that it had to be upscaled – naively, with no antialiasing! – even just to get to NTSC (480 scanlines, give or take).

                    So yeah, when each “pixel” is three scanlines tall, of course it’s going to look jagged even after the CRT blurs it!

        • @echo64
          link
          111 year ago

          CRT filters exist now, and with HDR output (or just sending an HDR-enable signal to get tv’s to use the full brightness range) and 4k displays it honestly as good at this point. or better because the only good CRT’s you can get now are pretty small P/BVM and my tv is much bigger than those

        • @[email protected]
          link
          fedilink
          7
          edit-2
          1 year ago

          There are plenty of upscalers with minimal latency that fix that.

          There also isn’t just “CRT” in this space. Professional video monitors give a very different picture than a consumer TV with only the RF converter input.

          If one more under 25 retro fan tells me that RF tuners are the “true experience”, I’m going to drink myself to death with Malort.

          Edit: please don’t tell me you believe CRTs have zero latency. Because that’s wrong, too.

        • @marx2k
          link
          11 year ago

          Compare a PS5 on a modern day large screen 4k TV vs a CRT of your favorite brand from any year.

          If your only use case is playing old consoles, there’s filters for current emulators that fill that need adequately.

        • @Kodiack
          link
          21 year ago

          I’m 32 and I can still hear a CRT whine just like when I was a kid. Recently enough I heard a CRT in a retro exhibit before I even saw it.

      • @uis
        link
        -51 year ago

        Use better power supply

        • @Zoboomafoo
          link
          91 year ago

          Is there a better power supply than cable + wall socket?

          • @grue
            link
            English
            31 year ago

            Yes! Hire an electrical engineer to improve it for you, you pleb! \s

          • @uis
            link
            01 year ago

            If you got TV from crapufacturer, then yes.

            • @[email protected]
              link
              fedilink
              3
              edit-2
              1 year ago

              Sony Trinitrons had whine to them, and those were basically the top consumer display back then. I think my JVC PVM has power supply whine.

    • @[email protected]
      link
      fedilink
      English
      91 year ago

      Are you serious?

      • Curved (the wrong way)
      • Massively heavy
      • Noise (just from the unit itself
      • Very low resolution
      • Noticably hot (might be a benefit in the winter)
      • Small picture, especially relative to weight
      • Depending how far back you go, no/shitty remote, only has 1 port for video
        • @[email protected]
          link
          fedilink
          21 year ago

          Sometimes I think about how some technologies could have evolved if they didn’t get out of fashion. I always thought it’s a bit unfair to compare products made decades ago with new ones and use it as a comparison for the whole technology.

          In the case of crts, it would be totally possible to make them with modern aspect ratio and resolutions. The greatest challenges would probably be size, weight and power consumption.

      • @grue
        link
        English
        11 year ago

        Very low resolution

        For TVs, that’s just because they didn’t need any more resolution because the signal they were displaying was 480i (or even worse, in the case of things like really old computers/video game consoles).

        My circa-2000 19" CRT computer monitor, on the other hand, could do a resolution that’s still higher than what most similarly-sized desktop flat screen monitors can manage (it was either QXGA [2048x1536] or QSXGA [2560x2048], I forget which).

        And then, of course, there were specialized CRT displays like oscilloscopes and vector displays that actually drew with the electron beam and therefore had infinite “resolution.”

        Point is, the low resolution was not an inherent limitation of CRT technology.

    • @new_guy
      link
      51 year ago

      They were great until you had to move them. They were clunkier than a sofa because they had no place to hold and weighted as much as a refrigerator

    • Maiznieks
      link
      11 year ago

      They did break, You know? My father fixed those things, it’s that they were actually fixable back then and it was cool. Or maybe it was just russian tech that broke, we lived in one of those ussr sattellite countries.