In the rapidly evolving landscape of deep learning, the proliferation of large-scale neural networks has brought forth new horizons of artificial intelligence. However, their sheer size and computational demands have posed significant challenges, limiting their applicability in resource-constrained environments. The overarching problem of network compression seeks to mitigate these challenges by efficiently designing deep neural models. In this report, we embark on a journey to address this problem, exploring novel avenues for model compression. Our first contribution, CORING, introduces a pioneering filter pruning method that harnesses tensor decomposition, preserving the multidimensional essence of filters. By leveraging the power of CORING, we achieve impressive reductions in model size and computational requirements while retaining or even enhancing performance. CORING’s ability to generalize models through pruning is demonstrated across various architectures and datasets. The second contribution, NORTON, unveils a hybrid network compression technique that combines tensor decompositions with structured pruning. NORTON offers a comprehensive approach to model compression, optimizing architecture, and reducing the number of parameters. With NORTON, we attain superior compression ratios and accuracy retention, making it a versatile tool for model optimization. Looking ahead, our research sets the stage for future investigations. Potential avenues include delving deeper into the compression domain, expanding to encompass various decomposition techniques, exploring a broader spectrum of neural network architectures, and applying these efficient models to diverse applications. As we navigate the evolving landscape of deep learning, the pursuit of efficient model design remains at the forefront, driving innovation and unlocking the potential for AI in resource-constrained scenarios.