Abstract:Flexible-geometry arrays have garnered much attention in wireless communications, which dynamically adjust wireless channels to improve the system performance. In this paper, we propose a novel flexible-geometry array for a $360^\circ$ coverage, named flxible cylindrical array (FCLA), comprised of multiple flexible circular arrays (FCAs). The elements in each FCA can revolve around the circle track to change their horizontal positions, and the FCAs can move along the vertical axis to change the elements' heights. Considering that horizontal revolving can change the antenna orientation, we adopt both the omni-directional and the directional antenna patterns. Based on the regularized zero-forcing (RZF) precoding scheme, we formulate a particular compressive sensing (CS) problem incorporating joint precoding and antenna position optimization, and propose two effective methods, namely FCLA-J and FCLA-A, to solve it. Specifically, the first method involves jointly optimizing the element's revolving angle, height, and precoding coefficient within a single CS framework. The second method decouples the CS problem into two subproblems by utilizing an alternative sparse optimization approach for the revolving angle and height, thereby reducing time complexity. Simulation results reveal that, when utilizing directional radiation patterns, FCLA-J and FCLA-A achieve substantial performance improvements of 43.32\% and 25.42\%, respectively, compared to uniform cylindrical arrays (UCLAs) with RZF precoding.
Abstract:The growing emotional stress in modern society has increased the demand for Emotional Support Conversations (ESC). While Large Language Models (LLMs) show promise for ESC, they face two key challenges: (1) low strategy selection accuracy, and (2) preference bias, limiting their adaptability to emotional needs of users. Existing supervised fine-tuning (SFT) struggles to address these issues, as it rigidly trains models on single gold-standard responses without modeling nuanced strategy trade-offs. To overcome these limitations, we propose Chain-of-Strategy Optimization (CSO), a novel approach that optimizes strategy selection preferences at each dialogue turn. We first leverage Monte Carlo Tree Search to construct ESC-Pro, a high-quality preference dataset with turn-level strategy-response pairs. Training on ESC-Pro with CSO improves both strategy accuracy and bias mitigation, enabling LLMs to generate more empathetic and contextually appropriate responses. Experiments on LLaMA-3.1-8B, Gemma-2-9B, and Qwen2.5-7B demonstrate that CSO outperforms standard SFT, highlighting the efficacy of fine-grained, turn-level preference modeling in ESC.
Abstract:Role-playing enables large language models (LLMs) to engage users in immersive and personalized interactions, but it also introduces significant safety risks. Existing role-play fine-tuning techniques improve role adaptability but may degrade safety performance, particularly for villainous characters. In this work, we conduct the first comprehensive assessment of role-play fine-tuning risks by training 95 role-specific LLMs using RoleBench. Our experiments reveal that role-play fine-tuning leads to a noticeable decline in safety performance, with safety risks varying based on character traits. To tackle this challenge, we propose Safety-Aware Role-Play Fine-Tuning (SaRFT), a novel method designed to balance role-playing capabilities and safety. Extensive experiments on LLaMA-3-8B-Instruct, Gemma-2-9B-it, and Qwen2.5-7B-Instruct demonstrate that SaRFT consistently outperforms state-of-the-art baselines under both LoRA and full-parameter fine-tuning settings. Our findings highlight the necessity of role-adaptive safety measures and provide insights into mitigating role-specific safety risks in role-playing LLMs.
Abstract:As wireless communication advances toward the 6G era, the demand for ultra-reliable, high-speed, and ubiquitous connectivity is driving the exploration of new degrees-of-freedom (DoFs) in communication systems. Among the key enabling technologies, Movable Antennas (MAs) integrated into Flexible Cylindrical Arrays (FCLA) have shown great potential in optimizing wireless communication by providing spatial flexibility. This paper proposes an innovative optimization framework that leverages the dynamic mobility of FCLAs to improve communication rates and overall system performance. By employing Fractional Programming (FP) for alternating optimization of beamforming and antenna positions, the system enhances throughput and resource utilization. Additionally, a novel Constrained Grid Search-Based Adaptive Moment Estimation Algorithm (CGS-Adam) is introduced to optimize antenna positions while adhering to antenna spacing constraints. Extensive simulations validate that the proposed system, utilizing movable antennas, significantly outperforms traditional fixed antenna optimization, achieving up to a 31\% performance gain in general scenarios. The integration of FCLAs in wireless networks represents a promising solution for future 6G systems, offering improved coverage, energy efficiency, and flexibility.
Abstract:Despite the growing global demand for large language models (LLMs) that serve users from diverse linguistic backgrounds, most cutting-edge LLMs remain predominantly English-centric. This creates a performance gap across languages, restricting access to advanced AI services for non-English speakers. Current methods to enhance multilingual capabilities largely rely on data-driven post-training techniques, such as multilingual instruction tuning or continual pre-training. However, these approaches encounter significant challenges, including the scarcity of high-quality multilingual datasets and the limited enhancement of multilingual capabilities. They often suffer from off-target issues and catastrophic forgetting of central language abilities. To this end, we propose Lens, a novel approach to enhance multilingual capabilities of LLMs by leveraging their internal language representation spaces. Specially, Lens operates by manipulating the hidden representations within the language-agnostic and language-specific subspaces from top layers of LLMs. Using the central language as a pivot, the target language is drawn closer to it within the language-agnostic subspace, allowing it to inherit well-established semantic representations. Meanwhile, in the language-specific subspace, the representations of the target and central languages are pushed apart, enabling the target language to express itself distinctly. Extensive experiments on one English-centric and two multilingual LLMs demonstrate that Lens effectively improves multilingual performance without sacrificing the original central language capabilities of the backbone model, achieving superior results with much fewer computational resources compared to existing post-training approaches.