Welcome to my personal website, which serves as a platform for self-presentation and communication with others. Here, you can learn about my research progress in Intelligent Computing and other areas of interest. If you have any suggestions, please feel free to let me know.
My research centers on foundation models and collaborative intelligence between large and compact models, aiming to make powerful capabilities deployable under real-world constraints (latency, memory, privacy). I explore how large models can selectively guide, distill, and adapt lightweight models through reinforcement-driven example selection, subspace feature transfer, and curriculum-style routing.
Technically, I combine evolutionary optimization, multi-task feature selection, and reinforcement learning with representation compression to enhance generalization across heterogeneous, high-dimensional biomedical and textual datasets. This includes (1) in-context demonstration selection with diversity- and reward-aware policies, (2) evolutionary multi-task feature sharing for efficient search, (3) ensemble evolutionary pipelines for cancer pathway and biomarker discovery, and (4) adaptive distillation loops where small models iteratively absorb signal from larger counterparts while preserving efficiency.
Representative work spans large model prompting & selection, evolutionary feature learning, biomedical classification & pathway modeling, and edge/on-device inference. Publications appear in venues such as ICML, TKDE, TCBB, CSUR, KBS; several results received media coverage. I am currently pursuing synergistic LLM–edge co-training, pathway-level mechanistic modeling, and reliable evaluation protocols for collaborative small–large model systems.
Contact: xubin.wang [at] kindlab.site · 欢迎合作 / Collaboration welcome.
These two surveys establish complementary foundations for data-centric, resource-constrained intelligent systems. systematizes a closed-loop deployment paradigm for reasoning-capable large language models (LLMs) and autonomous agents across cloud–edge–device hierarchies. It organizes optimization layers (quantization, sparsity, low-rank adaptation, distillation), elastic collaboration (selective offloading, hierarchical routing, privacy-preserving personalization), and adaptive intelligence (context compression, dynamic agent/tool orchestration) into a cognition-preserving framework with standardized multi-axis evaluation (latency, energy/token, robustness, privacy, sustainability). In parallel, the survey characterizes design and evolution of lightweight yet capable models under stringent memory, power, and real-time constraints—covering data preprocessing pipelines, feature selection and curation, compression (quantization/pruning/distillation), hardware-aware co-design, and emerging foundation model influences.
We tackle the challenges associated with text classification tasks, particularly in few-shot prompting scenarios, by introducing the Reinforced Diverse Example Selector . RDES employs a reinforcement learning framework, specifically Q-learning and PPO, to optimize the selection of diverse reference examples, ensuring a balanced representation of data that enhances classification accuracy. Additionally, we explore the integration of Chain-of-Thought reasoning into the selection process, which further boosts the model's predictive performance. In parallel, we present the , an automated tool designed for the construction and management of knowledge bases within Retrieval-Augmented Generation (RAG) systems. This tool processes document data and utilizes large language models to generate high-quality question-answer pairs, facilitating the automated development of RAG system knowledge bases. Together, these contributions highlight the potential of advanced methodologies in addressing the complexities of text classification and knowledge management.
Our research at the intersection of Evolutionary Machine Learning and its applications has focused on addressing prominent challenges in complex domains like feature selection, biomarker identification and cancer classification. We have proposed novel algorithms and frameworks such as for large-scale feature selection using a self-adaptive differential evolution approach, for multi-task evolutionary learning through information sharing, for efficient biomarker gene identification from microarray data via feature weighting particle swarm optimization, for ensemble-based improved cancer screening through optimized feature selection, modeling, and classification, and which leverages a hybrid sampling technique and ant colony-based feature selection within an ensemble to enhance identification of hidden responders in imbalanced biological data. Extensive experimentation demonstrates the superior and robust performance of our proposed approaches, validating their ability to provide effective solutions for challenging machine learning problems across domains while overcoming issues like local optima, dimensionality, and generalization across datasets.