A scheme of pulse based neural circuits for object tracking is proposed. Different from the conventional frame-based methodology, the proposed design utilises parallel arrays of circuits to extract pixels with significant temporal contrast, which are indicators of moving objects. It can be implemented on an FPGA chip in full parallelism and improves the tracking performance dramatically. Moreover, its integrate-and-fire neural model can cluster concave data sets. Experimental results show that the proposed scheme outperforms conventional methods.Introduction: To automatically track multiple moving objects in a scene, conventional vision algorithms have to sequentially scan all pixels in each frame, and frame by frame. This leads to a tremendous computational burden if a high resolution camera is used, and it becomes a bottleneck for dynamic vision [1, 2]. Various methods have been proposed to cope with this issue. In [3], an optical flow algorithm is used to detect moving objects with a processing cycle time of 66 -166 ms. However, some frames may not be processed if a camera with a frame rate of higher than 25 fps is used. This would limit the response of the servo controller that is used to position the vision system. K-means algorithms are popular in clustering-based object tracking schemes [4]. The key weakness is that they cannot deal well with concave data sets in practical scenarios [5]. In [6], hardware improvement in the sensor layer is attempted to increase the processing speed. Lichtsteiner et al. propose a dynamic vision sensor (DVS), which is designed based on the addressed event representation (AER) technique and is different from the conventional frame based CCD and CMOS sensors. Although its dynamic performance is good, it has weakness in the target's detail recognition. In addition, its customised IC design is too specialised for most applications.