This paper investigates the exponential stability of a class of neural networks with asymmetric connection weights. By dividing the network state variables into various parts according to the characters of the neural networks, some new sufficient conditions of exponential stability are derived via constructing a Lyapunov function and using the method of the variation of constant. The new conditions are associated with the initial values and are described by some blocks of the interconnection matrix, and do not depend on other blocks. Examples are given to further illustrate the theory.