Your browser doesn't support javascript.
loading
A Hyper-Transformer model for Controllable Pareto Front Learning with Split Feasibility Constraints.
Tuan, Tran Anh; Dung, Nguyen Viet; Thang, Tran Ngoc.
Afiliación
  • Tuan TA; Faculty of Mathematics and Informatics, Hanoi University of Science and Technology; Center for Digital Technology and Economy (BK Fintech), Hanoi University of Science and Technology, Hanoi, Vietnam. Electronic address: tuan.ta222171m@sis.hust.edu.vn.
  • Dung NV; Faculty of Mathematics and Informatics, Hanoi University of Science and Technology; Center for Digital Technology and Economy (BK Fintech), Hanoi University of Science and Technology, Hanoi, Vietnam. Electronic address: dung.nv232215m@sis.hust.edu.vn.
  • Thang TN; Faculty of Mathematics and Informatics, Hanoi University of Science and Technology; Center for Digital Technology and Economy (BK Fintech), Hanoi University of Science and Technology, Hanoi, Vietnam. Electronic address: thang.tranngoc@hust.edu.vn.
Neural Netw ; 179: 106571, 2024 Jul 26.
Article en En | MEDLINE | ID: mdl-39121789
ABSTRACT
Controllable Pareto front learning (CPFL) approximates the Pareto optimal solution set and then locates a non-dominated point with respect to a given reference vector. However, decision-maker objectives were limited to a constraint region in practice, so instead of training on the entire decision space, we only trained on the constraint region. Controllable Pareto front learning with Split Feasibility Constraints (SFC) is a way to find the best Pareto solutions to a split multi-objective optimization problem that meets certain constraints. In the previous study, CPFL used a Hypernetwork model comprising multi-layer perceptron (Hyper-MLP) blocks. Transformer can be more effective than previous architectures on numerous modern deep learning tasks in certain situations due to their distinctive advantages. Therefore, we have developed a hyper-transformer (Hyper-Trans) model for CPFL with SFC. We use the theory of universal approximation for the sequence-to-sequence function to show that the Hyper-Trans model makes MED errors smaller in computational experiments than the Hyper-MLP model.
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2024 Tipo del documento: Article

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2024 Tipo del documento: Article