Having the advantages of redundancy and flexibility, various types of tight frames have already shown impressive performance in applications such as image and video processing. For example, the undecimated wavelet transform, which is a particular case of tight frames, is known to have good performance for the denoising problem. Empirically, it is widely known that higher redundancy rate of a tight frame often leads to better performance in applications. The wavelet/framelet transform is often implemented in an undecimated fashion for the purpose of better performance in practice. Though high redundancy rate of a tight frame can improve performance in applications, as the dimension increases, it also makes the computational cost skyrocket and the storage of frame coefficients increase exponentially. This seriously restricts the usefulness of such tight frames for problems in moderately high dimensions such as video processing in dimension three. Inspired by the directional tensor product complex tight framelets TP-CTFm with m≥3 in and and their impressive performance for image processing in and , in this paper we introduce directional tensor product complex tight framelets (called reduced TP-CTFm) with low redundancy. Such are particular examples of tight framelet filter banks with mixed sampling factors. In particular, we shall develop a directional tensor product complex tight framelet such that it performs nearly as well as the original TP-CTF6 in [20] for image/video denoising/inpainting but it has significantly lower redundancy rates than TP-CTF6 in every dimension. The in d dimensions not only offers good directionality as the original TP-CTF6 does but also has the low redundancy rate (e.g., the redundancy rates are 253"> and 74790f3a6fbc2a91017d12b642d"> for dimension 7474727b" title="Click to view the MathML source">d=1,…,5, respectively), in comparison with the redundancy rate of TP-CTF6 in dimension d . Moreover, our numerical experiments on image/video denoising and inpainting show that the performance using our proposed is often comparable with or sometimes better than several state-of-the-art frame-based methods which have much higher redundancy rates than that of .