1080i vs. 1080p: What's the Difference?
Progressive
(1080p) video is considered better than interlaced (1080i), but it's not always
clear why; here's what's actually happening on your TV screen.
Today's HDTVs can
display beautiful, 1,920 X 1,080-pixel video, but the actual quality of what
you're viewing depends on the source material. A lot of the time, you're not
seeing exactly 1080p. In fact, most TVs today have two modes with similar
names: 1080i and 1080p. Both have the same screen resolution, so what's the
difference between the two? Here are five things you need to know:
1080i
video is "interlaced." 1080i video plays back at 60 frames per second, but
that's a bit deceptive, because it's actually broadcast at 30 frames per
second. The TV then displays those frames twice, the first pass is
1,920 X 540 for the even scan line field, and the second pass is 1,920 X 540
for the odd scan line field. The process by which this occurs is called
interlacing. It contributes to a sense of motion and reduces perceived flicker.
1080p
video is called "progressive scan." In this format, 1,920-by-1,080-pixel high-definition
movies are progressively drawn line after line, so they're not interlaced. On
paper, that may not seem like a huge deal. But in the real world, what you end
up seeing looks sharper and more defined than 1080i, particularly during scenes
with a lot of fast motion.
Sometimes 1080p is
termed "Full HD" or "True HD," to distinguish it from 1080i
or 720p video. Blu-ray discs contain 1080p video at 24 Frames Per Second, and
then, using a method known as 3:2 pulldown, display it at 30 frames per second
on screen.
Data compression can confuse the issue. Sometimes cable companies will deliver a 1080i picture, but then compress the data significantly in order to use up less bandwidth. The result can mean smeared details or pixelated color gradations in certain scenes. It's still technically HD, and still looks better than standard-definition cable, but it's not as good as it could be.
Data compression can confuse the issue. Sometimes cable companies will deliver a 1080i picture, but then compress the data significantly in order to use up less bandwidth. The result can mean smeared details or pixelated color gradations in certain scenes. It's still technically HD, and still looks better than standard-definition cable, but it's not as good as it could be.
This also happens
with 1080p streaming Internet video, but in that case, it's usually dependent
on the speed of your data connection. In fact, Blu-ray is currently the only
practical format for watching lots of pure 1080p content. Even the latest Apple
TV, which supports 1080p streaming, does so in a compressed format that loses a
bit of quality (although it still looks quite good).
Both
formats look similar on smaller TVs. As a general rule, you need a larger TV to notice the
difference between 1080i and 1080p. Depending on your eyesight, you can
probably pick up the difference on a 32-inch LCD if you're particular about it.
But most consumers don't really see a marked difference until at least a
42-inch screen, if not larger. In fact, many people are perfectly happy with
720p HDTV sets even at higher sizes; we recently named one, the 51" Samsung PN51E490B4F, best Choice for budget
large-screen HDTVs.
1080p
isn't even the best anymore. Technology never stands
still, of course. Five years from now, you'll probably just want Ultra High
Definition (aka 4K) video instead. (For a closer look at 4K video, check out What is Ultra HD?) But for now, if you're a videophile who
appreciates a sharper picture, 1080p is definitely the way to go—both in HDTV
capability and in the source material you're viewing.