Abstract: The Mixture of Experts (MoE) model is a promising approach for handling code-switching speech recognition (CS-ASR) tasks. However, the existing CS-ASR work on MoE has yet to leverage the ...
Abstract: In the fields of intelligent transportation and multi-task cooperation, many practical problems can be modeled by colored traveling salesman problem (CTSP). When solving large-scale CTSP ...