When you want to connect a monitor to a computer you will almost certainly find a connection type of VGA, DVI or HDMI cables. Let’s see the difference between each of them:
VGA is the oldest of the three standards, having been introduced in 1987. VGA is an analog signal that handles only video, no sound, and does not manage security or digital rights. Video quality is susceptible to cable quality and the distance from the computer to the monitor. If the connector has small thumbscrews on the sides of the cable it is a VGA. Because of this, VGA image quality can vary greatly depending on the type of cable brand. The VGA connector is usually blue or black in color, has fifteen pins arranged in three horizontal rows, and is trapezoidal in shape.
DVI was invented in 1999 and is similar to VGA and HDMI. DVI allows you to send video but not audio, you can send digital, analog, or both. DVDI allows handling of digital rights management and can be converted to HDMI or VGA using a converter. DVI is considered a midpoint between VGA and HDMI, being practically the successor of VGA.
HDMI first appeared in 2002, is often found in modern televisions, but is also found in most newer computer monitors. If DVI is the successor to VGA, then HDMI is a possible successor to DVI, possibly due to its appearance in HDTV. HDMI is a digital standard (1 or 0) that allows both audio and video to be transmitted. Conditions such as the quality of the cable, the distance from the machine to the monitor or the type of metal of the connector do not affect the transmitted signal. HDMI can also handle security, meaning that certain types of signals such as pay TV can be blocked before traveling through an HDMI cable.