Our research focuses on enhancing the efficiency of cloud-based systems and applications. A key area of interest is the architecture and optimization of microservice-based applications. We are committed to discovering innovative architectural and software-level approaches to improve the performance, scalability, and resource utilization of these applications. Another significant effort is the development of an integrated management framework designed to ensure Service Level Objective (SLO) guarantees, providing a comprehensive solution for maintaining service quality and performance across diverse cloud services. Additionally, we are deeply involved in optimizing the performance of machine learning inference servers. Through careful analysis and design, we aim to maximize their potential, enabling rapid and accurate inferences across various applications. Our research also extends to optimizing server infrastructures for AI model training. By fine-tuning server configurations, resource allocation, and distributed processing techniques, we aim to create an optimized environment that accelerates the training of complex AI models. Ultimately, our goal is to drive the evolution of cloud computing by developing innovative solutions that streamline application deployment, management, and performance, while addressing the unique challenges of AI and machine learning workloads.