Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse

Film Glance Forum

  1. Home
  2. The IMDb Archives
  3. Can PS3s math the audio and video quality of a Blu-Ray player?

Can PS3s math the audio and video quality of a Blu-Ray player?

Scheduled Pinned Locked Moved The IMDb Archives
10 Posts 1 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • F Offline
    F Offline
    fgadmin
    wrote on last edited by
    #1

    Archived from the IMDb Discussion Forums — Blu-ray Hi-Def Equipment


    dont_diss_dave — 14 years ago(August 07, 2011 04:04 AM)

    I just thought about the HDMI leads, and don't PS3s have one HDMI lead? With Blu-Ray players you can pay from say £10/15 to £100+ for a HDMI lead. The more you pay the higher quality the audio and video get, right?
    "
    Time to find out
    "

    1 Reply Last reply
    0
    • F Offline
      F Offline
      fgadmin
      wrote on last edited by
      #2

      mikejonas — 14 years ago(August 07, 2011 10:08 PM)

      What's an "HDMI lead"? Are you talking about an HDMI cable?
      The more you pay the higher quality the audio and video get, right?
      Not in the case of HDMI. Well, there's different versions/types of HDMI and you'll probably pay more for an HDMI cable that has the bandwidth to push through 3D Blu-ray data (i.e. an HDMI 1.4 cable) than one that'll work with a conventional Blu-ray setup (1.3 and below), but in general, a cheap HDMI cable will perform just as well as an expensive HDMI cable. That's because of the "D" in "HDMI"it's digital, which means that the signal coming through is coming through perfectly fine, or noticeably broken; there's no "softer" and "fuzzier" for weaker signals.
      I just thought about the HDMI leads, and don't PS3s have one HDMI lead?
      So?
      So picture-wise, there should be no difference between a PS3 and a standalone Blu-ray player. Audio-wise, I'm not completely up-to-date as to what different audio formats the PS3 decodes internally and/or outputs directly, but I think that for most consumers, it's really six of one and half a dozen of the other and only the most discerning audiophile might care about the difference.
      The PS3's a pretty robust computing machine, and at least back in the day (if not currently) it was more powerful than any standalone Blu-ray player. Even a PS3 from 2006 can play Blu-ray 3D discs, something dedicated Blu-ray players had to be specifically made to do.

      1 Reply Last reply
      0
      • F Offline
        F Offline
        fgadmin
        wrote on last edited by
        #3

        Ace_of_Sevens — 13 years ago(April 08, 2012 12:00 AM)

        The only thing PS3 can't do is play lossless Dolby TrueHD on 3D titles. There are only 5 3d titles with TrueHD sound, though.Almost all of them use DTS-HD Master Audio, which it handles fine.
        Like commentaries?
        http://www.ratethatcommentary.com/

        1 Reply Last reply
        0
        • F Offline
          F Offline
          fgadmin
          wrote on last edited by
          #4

          dangus — 13 years ago(April 25, 2012 08:17 PM)

          The component output is limited to 1080i for Blu-ray due to concerns about copying. Whether that's visibly worse than 1080p, I don't know. But I figure there must be some loss going from digital-to-analog and back again, compared with a straight digital to digital connection like HDMI.

          1 Reply Last reply
          0
          • F Offline
            F Offline
            fgadmin
            wrote on last edited by
            #5

            stmorgan — 13 years ago(November 23, 2012 04:31 AM)

            Hmmm, quality of cables make no difference, really? Perhaps a demo at a decent specialist (not currys!) might change your view. If not, maybe an eye test might?

            1 Reply Last reply
            0
            • F Offline
              F Offline
              fgadmin
              wrote on last edited by
              #6

              mikejonas — 13 years ago(November 23, 2012 04:44 AM)

              You evidently don't understand the concept of a digital signal. We're talking about HDMI hereit's all ones and zeroes. As long as the ones and zeroes are getting from point A to point B, it's going to look perfect. If they're not, then the picture will look noticeably broken. There's obviously going to be differences in build quality, and, as technology advances, standards (i.e. an HDMI 1.4 cable is designed to handle more data than its predecessors, to accommodate the additional bandwidth required by 3D Blu-ray), but in the realm of HDMI, it either works perfectly or it doesn't.

              1 Reply Last reply
              0
              • F Offline
                F Offline
                fgadmin
                wrote on last edited by
                #7

                stmorgan — 13 years ago(December 01, 2012 12:35 PM)

                So how is it that you can clearly see picture quality differences between different hdmi cables?

                1 Reply Last reply
                0
                • F Offline
                  F Offline
                  fgadmin
                  wrote on last edited by
                  #8

                  mikejonas — 13 years ago(December 02, 2012 06:17 AM)

                  Because you're imagining things.
                  Again, DIGITAL is DIGITAL. The devices are transmitting and receiving ONES and ZEROES; there's no such thing as receiving a ".5".
                  Saying one HDMI cable will yield better pictures than another is like saying one printer cable will yield sharper looking printouts than another. It's simply not true.

                  1 Reply Last reply
                  0
                  • F Offline
                    F Offline
                    fgadmin
                    wrote on last edited by
                    #9

                    MaximumMadness — 12 years ago(April 18, 2013 09:47 AM)

                    So how is it that you can clearly see picture quality differences between different hdmi cables?
                    Here's an article you may find interesting, stmorgan.
                    http://news.cnet.com/8301-17938_105-20056502-1/why-all-hdmi-cables-are -the-same/
                    Here are direct quotes from the end of the article:
                    "No matter what, though, there is absolutely no picture or sound quality difference between a $3.50 cable and a $1,000 cable."
                    "In the year and a half since we first published this article, the most common misunderstanding comes from those used to an analog cable mentality. They understand that over any cable, there is a high likelihood of signal degradation. As in, the signal received by the television isn't as strong or exactly the same as what leaves the source. However, unlike analog cables, there is no linear correlation between signal degradation and picture degradation."
                    "Most of you reading this only need a few feet of HDMI cabling to run from your Blu-ray player and cable/satellite box to your TV. Over these short distances, even the cheapest HDMI cables are going to work. And if they work, as you've read, it means you're getting perfect image and sound. Even over long runs, most cheap cables can do the job just fine."
                    As someone else pointed out to you, digital is digital. It's 1's and 0's. Different cables aren't going to do a better or worse job at transmitting the digital images.
                    And FURTHERMORE, this is my signature! SERIOUSLY! Did you think I was still talking about my point?

                    1 Reply Last reply
                    0
                    • F Offline
                      F Offline
                      fgadmin
                      wrote on last edited by
                      #10

                      x_jmt_x — 12 years ago(July 29, 2013 12:42 AM)

                      Common sense, really. The degradation of an analogue signal
                      may
                      result in a specific problem, like the color balance being off.
                      However, the degradation of a digital picture signal would and will result in noticable errors in the picture. The signal is just ones and zeroes, which is then 'decoded' into a visible picture. Any changes made to the picture are done when processing the signal.
                      Now, missing a few ones and zeroes may or may not result in error. If not, nothing happens. If does produce an error, the image pixelates, breaks down, or is totally lost.
                      Does any sane person really believe that a cable could rip off "some" of the ones and zeroes signifying the green colour, resulting in dull greens, but not any breaking the image.
                      While there are problems with digital signals aswell, mainly jitter (time-distortion of the signal, leading to loss of bit-depth in sound), these do not affect a picture-signal. And even jitter is more dependant of the transmitting and receiving devices (and the D/A converter), rather then cable.
                      Last 20 minutes, Pacino eats every last inch of the set and sh!ts Oscars.

                      1 Reply Last reply
                      0

                      • Login

                      • Don't have an account? Register

                      Powered by NodeBB Contributors
                      • First post
                        Last post
                      0
                      • Categories
                      • Recent
                      • Tags
                      • Popular
                      • Users
                      • Groups