OPAL: Operator-Programmed Algorithms for Landscape-Aware Black-Box Optimization

Reading time: 5 minute
...

📝 Original Info

  • Title: OPAL: Operator-Programmed Algorithms for Landscape-Aware Black-Box Optimization
  • ArXiv ID: 2512.12809
  • Date: 2025-12-14
  • Authors: Junbo Jacob Lian, Mingyang Yu, Kaichen Ouyang, Shengwei Fu, Rui Zhong, Yujun Zhang, Jun Zhang, Huiling Chen

📝 Abstract

Black-box optimization often relies on evolutionary and swarm algorithms whose performance is highly problem dependent. We view an optimizer as a short program over a small vocabulary of search operators and learn this operator program separately for each problem instance. We instantiate this idea in Operator-Programmed Algorithms (OPAL), a landscape-aware framework for continuous black-box optimization that uses a small design budget with a standard differential evolution baseline to probe the landscape, builds a $k$-nearest neighbor graph over sampled points, and encodes this trajectory with a graph neural network. A meta-learner then maps the resulting representation to a phase-wise schedule of exploration, restart, and local search operators. On the CEC~2017 test suite, a single meta-trained OPAL policy is statistically competitive with state-of-the-art adaptive differential evolution variants and achieves significant improvements over simpler baselines under nonparametric tests. Ablation studies on CEC~2017 justify the choices for the design phase, the trajectory graph, and the operator-program representation, while the meta-components add only modest wall-clock overhead. Overall, the results indicate that operator-programmed, landscape-aware per-instance design is a practical way forward beyond ad hoc metaphor-based algorithms in black-box optimization.

💡 Deep Analysis

Figure 1

📄 Full Content

1 OPAL: Operator-Programmed Algorithms for Landscape-Aware Black-Box Optimization Junbo Jacob Lian, Mingyang Yu, Kaichen Ouyang, Shengwei Fu, Rui Zhong, Yujun Zhang, Jun Zhang, Fellow, IEEE, and Huiling Chen, Member, IEEE Abstract—Black-box optimization often relies on evolutionary and swarm algorithms whose performance is highly problem dependent. We view an optimizer as a short program over a small vocabulary of search operators and learn this operator program separately for each problem instance. We instantiate this idea in Operator-Programmed Algorithms (OPAL), a landscape- aware framework for continuous black-box optimization that uses a small design budget with a standard differential evolution baseline to probe the landscape, builds a k-nearest neighbor graph over sampled points, and encodes this trajectory with a graph neural network. A meta-learner then maps the resulting representation to a phase-wise schedule of exploration, restart, and local search operators. On the CEC 2017 test suite, a single meta-trained OPAL policy is statistically competitive with state- of-the-art adaptive differential evolution variants and achieves significant improvements over simpler baselines under nonpara- metric tests. Ablation studies on CEC 2017 justify the choices for the design phase, the trajectory graph, and the operator- program representation, while the meta-components add only modest wall-clock overhead. Overall, the results indicate that operator-programmed, landscape-aware per-instance design is a practical way forward beyond ad hoc metaphor-based algorithms in black-box optimization. Index Terms—Black-box optimization, evolutionary computa- tion, hyper-heuristics, meta-learning, graph neural networks. I. INTRODUCTION C ONTINUOUS black-box optimization (BBO) lies at the core of many engineering, control, and machine learning applications. In these settings, the objective func- tion is expensive, derivative-free, noisy, or multimodal, and This research is financially supported by the National Natural Science Foun- dation of China (Grant No. 62076185, 62301367). (Corresponding authors: Jun Zhang & Huiling Chen.) Junbo Jacob Lian is with the McCormick School of Engi- neering, Northwestern University, Evanston, IL, USA (e-mail: Jacoblian@u.northwestern.edu). Mingyang Yu is with the College of Artificial Intelligence, Nankai Univer- sity, Tianjin, China (e-mail: 1120240312@mail.nankai.edu.cn). Kaichen Ouyang is with the School of Mathematics, University of Science and Technology of China, Hefei, China (e-mail: oykc@mail.ustc.edu.cn). Shengwei Fu is with Guizhou University, Guiyang, China (e-mail: gs.swfu22@gzu.edu.cn). Yujun Zhang is with the School of New Energy, Jingchu University of Technology, Jingmen, China (e-mail: zhangyj069@gmail.com). Rui Zhong is with the Information Initiative Center, Hokkaido University, Sapporo, Japan (e-mail: zhongrui@iic.hokudai.ac.jp). Jun Zhang is with the College of Artificial Intelligence, Nankai University, Tianjin, China (e-mail: junzhang@nankai.edu.cn). Huiling Chen is with the School of Computer Science and Artificial Intelligence, Wenzhou University, Wenzhou, China (e-mail: chenhuiling.jlu@gmail.com). Source code, experiment scripts, and results are publicly available at https: //github.com/junbolian/OPAL. practitioners typically rely on population-based metaheuris- tics such as differential evolution (DE) and particle swarm optimization (PSO) [1]–[3]. Over the last two decades, the field has produced a large ecosystem of increasingly so- phisticated variants—for example L-SHADE [4] and jSO [5] in the DE family [6]—that achieve strong performance on standard benchmarks but remain highly problem-dependent. The No Free Lunch theorems for optimization make this dependence formal: averaged over all possible problems, no single algorithm or configuration can dominate all others [7]. In practice, even within a single problem instance, algorithmic needs change over time as the search progresses from global exploration to local exploitation [8]. A natural response has been to use machine learning to support or replace human design. One line of work builds per-instance algorithm selectors and configurators: a feature extractor characterizes the problem, and a model predicts which algorithm or parameter setting will perform best [9], [10]. Exploratory landscape analysis (ELA) plays a central role in this program: it maps sampled points into numerical descriptors of modality, ruggedness, conditioning, and other properties, which are then fed into supervised models for solver selection or performance prediction [10], [11]. Recent surveys document a rapid growth in such ML-assisted meta- heuristics, covering both single-objective and multiobjective settings and highlighting hybrid designs where learned models steer classical metaheuristics [12]. A second line of work treats the optimizer itself as a dynamical system and learns to adapt its parameters on- line. Automated and dynami

📸 Image Gallery

chl.jpg fig1.png flow.jpg fsw.jpg junbolian.jpg oykc.png ymy.jpg zj.jpg zr.png zyj.jpg

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut