A new technique is demonstrated that improves the performance of optical networks subject to interferometric noise. Every transmitter comprises a selected DFB laser that is NRZ ASK directly modulated with a large modulation depth. Under modulation, the center frequency of the DFB laser is found to vary over the duration of each bit, falling during mark bits and rising during space bits with an exponential-like dependency of characteristic time constant ∼20 ns. The optical frequency at the midpoint of each bit interval is not constant but depends upon the sequence of the preceding bits. Therefore, on the interference of a delayed-replica parasitic crosstalk waveform the interferometric noise generated on detection may be reduced by RF filtering according to the difference in center frequency of the interfering bits. This noise suppression which requires no additional hardware has been successfully modeled and experiment demonstrates little performance gain for a single interferer, as predicted. However, in the presence of multiple interferers significant improvement is predicted and observed at sub-Gb/s rates.