Some Optimization Problems in Large Language Models

  • Hua DING
  • , Yi JIANG
  • , Zhi-Hui ZHAN*
  • , Jun ZHANG*
  • , Sam KWONG
  • *Corresponding author for this work

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

Abstract

Large language model (LLM) is a research hotspot in the field of artificial intelligence (AI). Currently, various LLMs have been designed and have demonstrated outstanding language understanding and generation capabilities in various natural language processing tasks. With the development of LLM research, a spectrum of optimization problems have emerged, posing challenges to the pursuit of further performance gains of LLMs. To enlighten future research on the optimization problems in LLMs, this paper summarizes some significant optimization problems in LLMs and proposes a classification method based on the model’s scope to categorize them. Specifically, this paper focuses on three types of cutting-edge issues namely model merging, prompt engineering, and jailbreaking attack. First, the model merging refers to the optimization problem of the integration of different pre-trained models. Second, the prompt engineering refers to the optimization problem that involves designing effective input prompts to guide LLMs to generate desired responses. Third, the jailbreaking attack refers to the optimization problem that indicates the process of crafting adversarial inputs that can circumvent the safety constraints or ethical guidelines programmed into LLMs. Also, we deeply analyze how evolutionary computation (EC) can empower LLMs with customized optimization solutions through their autonomous learning and efficient search capabilities, significantly improving the automation efficiency. Based on the above analyses, we finally discuss some potential future problem directions of LLM, aiming at promoting the deep integration and innovative development of LLMs in the field of automated optimization, leading to the next generation of AI.
Original languageEnglish
Title of host publicationProceedings: 2025 15th International Conference on Information Science and Technology, ICIST 2025
PublisherIEEE
Pages86-92
Number of pages7
ISBN (Electronic)9798331514501
ISBN (Print)9798331514518
DOIs
Publication statusPublished - Dec 2025
Event15th International Conference on Information Science and Technology - Zhanjiang, Guangdong, China
Duration: 5 Dec 20258 Dec 2025

Publication series

NameProceedings: International Conference on Information Science and Technology
PublisherIEEE
ISSN (Print)2164-4357
ISSN (Electronic)2573-3311

Conference

Conference15th International Conference on Information Science and Technology
Abbreviated titleICIST 2025
Country/TerritoryChina
CityGuangdong
Period5/12/258/12/25

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2024YFF0509600, in part by the National Natural Science Foundation of China (NSFC) under Grant 62176094 and Grant U23B2039, in part by the Tianjin Top Scientist Studio Project under Grant 24JRRCRC00030, in part by the Tianjin Belt and Road Joint Laboratory under Grant 24PTLYHZ00250, in part by the Fundamental Research Funds for the Central Universities, Nankai University (078-63253247), and in part by the National Research Foundation of Korea (NRF) Grant funded by the Korea government (MSIT) (No. RS-2025-00555463).

Keywords

  • large language models
  • optimization problems
  • evolutionary computation
  • model merging
  • prompt engineering
  • jailbreaking attack

Fingerprint

Dive into the research topics of 'Some Optimization Problems in Large Language Models'. Together they form a unique fingerprint.

Cite this