Verizon/Nortel: 100G Trial Demos Improved Distortion Tolerance

The two companies say a trial involving 73 kilometers of fiber show 100G networking technology provides twice the tolerance for signal distortion over standard 10G.

Print Article

Verizon and Nortel say they have successfully completed a trial meant to demonstrate that 100 Gigabit optical transmission technology better tolerates signal distortion than the standard 10G wavelength.

In an announcement, Verizon said the trial involved a 73 kilometer stretch of field fiber in northeastern Texas. According to Verizon, Nortel's Optical Multiservice Edge 6500 convergence platform, which supports 40G and 100G networking, managed a rate of 92 Gbps. Verizon said the Nortel platform includes technology designed to "maintain signal integrity despite significant polarization mode dispersion (PMD)," which causes signals transmitted over optical fiber to distort in transit. Significant amounts of PMD cause limits in the amount of data that can be sent over an optical link.

Verizon has performed a number of 100G trials with partners besides Nortel. The company says it set a distance record with Nokia Siemens Networks in which data was successfully transmitted at 100G on a single wavelength for more than 1,040 kilometers. In November of of 2007, the company partnered with Alcatel-Lucent to transmit a FiOS video stream between Tampa and Miami, Florida at 100G over a live network.

The IEEE pegged 100 Gbps as the next target for Ethernet technology in late 2006. After a year with the IEEE 802.3 Higher Speed Study Group, it became the focus of a task force in December, 2007, when the IEEE established 802.3ba as the designation for 40 and 100 Gigabit Ethernet communications.

Article courtesy of Optically Networked

This article was originally published on Oct 8, 2008
Get the Latest Scoop with Networking Update Newsletter