Showing 1 - 16 results of 16 for search '"mixture-of-experts"', query time: 0.08s Refine Results
  1. 1
  2. 2

    Mixture of Expert Large Language Model for Legal Case Element Recognition by YIN Hua, WU Zihao, LIU Tingting, ZHANG Jiajia, GAO Ziqian

    Published 2024-12-01
    “…This paper introduces a conversational mixture of expert element recognition LLM. The proposed model in this paper first designs specific prompts tailored to the characteristics of cases for the ChatGLM3-6B-base model. …”
    Get full text
    Article
  3. 3

    Gated ensemble of spatio-temporal mixture of experts for multi-task learning in ride-hailing system by Md Hishamur Rahman, Shakil Mohammad Rifaat, Soumik Nafis Sadeek, Masnun Abrar, Dongjie Wang

    Published 2024-12-01
    “…Therefore, a multi-task learning architecture is proposed in this study by developing gated ensemble of spatio-temporal mixture of experts network (GESME-Net) with convolutional recurrent neural network (CRNN), convolutional neural network (CNN), and recurrent neural network (RNN) for simultaneously forecasting these spatio-temporal tasks in a city as well as across different cities. …”
    Get full text
    Article
  4. 4

    Active reinforcement learning versus action bias and hysteresis: control with a mixture of experts and nonexperts. by Jaron T Colas, John P O'Doherty, Scott T Grafton

    Published 2024-03-01
    “…In light of how bias and hysteresis function as a heuristic for efficient control that adapts to uncertainty or low motivation by minimizing the cost of effort, these phenomena broaden the consilient theory of a mixture of experts to encompass a mixture of expert and nonexpert controllers of behavior.…”
    Get full text
    Article
  5. 5
  6. 6

    Mixture of Experts Framework Based on Soft Actor-Critic Algorithm for Highway Decision-Making of Connected and Automated Vehicles by Fuxing Yao, Chao Sun, Bing Lu, Bo Wang, Haiyang Yu

    Published 2025-01-01
    “…This paper proposes a Mixture of Expert method (MoE) based on Soft Actor-Critic (SAC), where the upper-level discriminator dynamically decides whether to activate the lower-level DRL expert or the heuristic expert based on the features of the input state. …”
    Get full text
    Article
  7. 7
  8. 8
  9. 9
  10. 10
  11. 11
  12. 12
  13. 13

    The cognitive reality monitoring network and theories of consciousness by Aurelio Cortese, Mitsuo Kawato

    Published 2024-04-01
    “…The cognitive reality monitoring network (CRMN) model is derived from computational theories of mixture-of-experts architecture, hierarchical reinforcement learning and generative/inference computing modules, addressing all three levels of understanding. …”
    Get full text
    Article
  14. 14

    Research on Predicting Super-Relational Data Links for Mine Hoists Within Hyper-Relational Knowledge Graphs by Xiaochao Dang, Xiaoling Shu, Fenfang Li, Xiaohui Dong

    Published 2024-12-01
    “…This paper proposes the HyLinker model, designed to improve the representation of entities and relations through modular components, including an entity neighbor aggregator, a relation qualifier aggregator, MoE-LSTM (Mixture of Experts Long Short-Term Memory), and a convolutional bidirectional interaction module. …”
    Get full text
    Article
  15. 15

    Enhancing depression recognition through a mixed expert model by integrating speaker-related and emotion-related features by Weitong Guo, Qian He, Ziyu Lin, Xiaolong Bu, Ziyang Wang, Dong Li, Hongwu Yang

    Published 2025-02-01
    “…To tackle this challenge, we propose a Mixture-of-Experts (MoE) method that integrates speaker-related and emotion-related features for depression recognition. …”
    Get full text
    Article
  16. 16

    LoRA Fusion: Enhancing Image Generation by Dooho Choi, Jeonghyeon Im, Yunsick Sung

    Published 2024-11-01
    “…One emerging approach constructs several LoRA modules, but more than three typically decrease the generation performance of pre-trained models. The mixture-of-experts model solves the performance issue, but LoRA modules are not combined using text prompts; hence, generating images by combining LoRA modules does not dynamically reflect the user’s desired requirements. …”
    Get full text
    Article