..............................
            ..............................
            ..............................
            
Improvement of Imperialist Competitive Algorithm based on the Cosine Similarity Criterion of
        
        The  goal  of  optimizing  the  best  acceptable  answer  is  according  to  the  limitations  and  needs  of  the  problem. For  a 
problem,  there  are  several  different  answers  that  are  defined  to  compare  them  and  select  an  optimal answer; a  function  is 
called  a  target  function.  The  choice  of  this  function  depends  on  the  nature  of the  problem.  Sometimes  several  goals  are 
together  optimized; such optimization  problems  are  called  multi-objective  issues.  One  way  to  deal  with  such  problems  is  to 
form a new objective function in the form of a linear combination of the main objective functions. In the proposed approach, in 
order  to  increase  the  ability  to  discover  new  position  in  the Imperialist  Competitive  Algorithm (ICA),  its  operators  are 
combined  with  the  particle  swarm  optimization. The  colonial  competition  optimization  algorithm  has  the  ability  to  search 
global  and  has  a  fast  convergence  rate,  and  the  particle  swarm  algorithm  added  to  it  increases  the  accuracy  of  searches. In 
this  approach,  the  cosine  similarity  of  the neighboring  countries is  measured  by  the  nearest  colonies of  an  imperialist  and 
closest  competitor  country. In  the  proposed  method,  by  balancing  the  global  and  local  search,  a  method  for  improving  the 
performance  of  the  two  algorithms  is  presented.  The  simulation  results  of  the  combined  algorithm  have  been  evaluated  with 
some of the benchmark functions. Comparison of the results has been evaluated with respect to metaheuristic algorithms such 
as Differential Evolution (DE), Ant Lion Optimizer (ALO), ICA, Particle  Swarm  Optimization (PSO), and Genetic  Algorithm 
(GA).    
            [1] Abdeyazdan M., “Data Clustering Based on Hybrid Kharmonic Means and Modifier Imperialist Competitive Algorithm,” The Journal of Supercomputing, vol. 68, no. 2, pp. 574-598, 2014.
[2] Atashpaz-Gargari E. and Lucas C., “Imperialist Competitive Algorithm: An Algorithm for Optimization Inspired by Imperialistic Competition,” in Proceedings of IEEE Congress on Evolutionary Computation, Singapore, pp. 4661-4667, 2007.
[3] Ban M. and Jalal S., “A Novel Hybrid Imperialist Competitive Algorithm for Global Optimization,” Australian Journal of Basic and Applied Sciences, vol. 7, pp 330-341, 2013.
[4] Bansal J., Singh P., Saraswat M., Verma A., 268 The International Arab Journal of Information Technology, Vol. 18, No. 3, May 2021 Jadon S., and Abraham A., “Inertia Weight Strategies in Particle Swarm Optimization,” in Proceedings of 3rd World Congress on Nature and Biologically Inspired Computing, Salamanca, pp. 633-640, 2011.
[5] Branke J., Evolutionary Optimization in Dynamic Environment, Kluwer Academic Publishers, 2001.
[6] Dorigo M. and Stu¨tzle T., Handbook of Metaheuristics, Springer Link, 2003.
[7] Dorigo M., Birattari M., and Stutzle T., “Ant Colony Optimization,” IEEE Computational Intelligence Magazine, vol. 1, no. 4, pp. 28-39, 2006.
[8] Goldberg D., Genetic Algorithms in Search, Optimization and Machine Learning, Addison- Wesley Reading, 1989.
[9] Hoos H. and Stu¨tzle T., Stochastic Local Search: Foundations and Applications, Elsevier, 2004.
[10] Idoumghar L., Chérina N., Siarry P., Roche R., and Miraouic A., “Hybrid ICA-PSO Algorithm for Continuous Optimization,” Applied Mathematics, and Computation, vol. 129, no. 24, pp. 11149-11170, 2013.
[11] Jamil M. and Yang X., “A Literature Survey of Benchmark Functions for Global Optimization Problems,” Journal of Mathematical Modeling and Numerical Optimization, vol. 4, no. 2, pp. 150-194, 2013.
[12] Jin Y. and Branke J., “Evolutionary Optimization in Uncertain Environments A Survey,” IEEE Transactions Evolutionary Computation, vol. 9, no. 3, pp. 303-317, 2005.
[13] Johnson D., and Papadimitriou C., and Yannakakis M., “How Easy is Local Search?,” Journal of Computer and System Sciences, vol. 37, no. 79-100, 1988.
[14] Karaboga D. and Basturk B., “A Powerful and Efficient Algorithm for Numerical Function Optimization: Artificial Bee Colony (ABC) Algorithm,” Journal of Global Optimization, vol. 39, pp. 459-471, 2007.
[15] Kennedy J. and Eberhart R., “Particle Swarm Optimization,” in Proceedings of ICNN'95- International Conference on Neural Networks, Perth, 1995.
[16] Kirkpatrick S., “Optimization by Simulated Annealing: Quantitative Studies,” Journal of Statistical Physics, vol. 34, pp. 975-986, 1984.
[17] Li C. and Yang S., “A Clustering Particle Swarm Optimizer for Dynamic Optimization,” in Proceedings of IEEE Congress on Evolutionary Computation, Trondheim, pp. 439-446, 2009.
[18] Mirjalili S., “The Ant Lion Optimizer,” Advances in Engineering Software, vol. 83, pp. 80-98, 2015.
[19] Mirjalili S., Mirjalili S., and Hatamlou A., “Multi- Verse Optimizer: A Nature-Inspired Algorithm for Global Optimization,” Neural Computing and Applications, vol. 27, pp. 495-513, 2016.
[20] Mitchell M., Holland J., and Forrest S., “When Will A Genetic Algorithm Outperform Hill Climbing?” Advances in Neural Information Processing Systems, pp. 51-58, 1993.
[21] Pan X., Xue L., Lu Y., and Sun N., “Hybrid Particle Swarm Optimization with Simulated Annealing, Multimedia Tools Application,” Multimedia Tools and Applications, vol. 78, no. 21, pp. 29921-29936, 2019.
[22] Rocca P., Oliveri G., and Massa A., “Differential Evolution as Applied to Electromagnetic,” IEEE Antennas and Propagation Magazine, vol. 53, no. 1, pp. 38-49, 2011.
[23] Storn R. and Price K., “Differential Evolution-A Simple and Efficient Heuristic for Global Optimization Over Continuous Spaces,” Journal Global Optim, vol. 11, pp. 341-59, 1997.
[24] Van-den-Bergh F. and Engelbrecht A., “A New Locally Convergent Particle Swarm Optimizer,” in Proceedings of IEEE International Conference on Systems, Man and Cybernetics, Yasmine Hammamet, 2002.
[25] Wolpert D. and Macready W., “No Free Lunch Theorems for Optimization,” IEEE Transactions on Evolutionary Computation, vol. 1, no. 1, pp. 67-82, 1997.
[26] Yin Z. and Gao Q., “A Novel Imperialist Competitive Algorithm for Scheme Configuration Rules Mining of Product Service System,” Arabian Journal for Science and Engineering, vol. 45, pp. 3157-3169, 2020.
[27] Zhang Y., Hu X., and Wu C., “Improved Imperialist Competitive Algorithms for Rebalancing Multi-Objective Two-Sided Assembly Lines with Space and Resource Constraints,” International Journal of Production, vol. 58, no. 12, pp. 3589-3617, 2019.
[28] Zheng L., Zou J., Liu B., Hu Y., and Deng Y., “A Novel Evidence Distance in Power Set Space,” The International Arab Journal of Information Technology, vol. 17, no. 1, pp. 8-15, 2020. Improvement of Imperialist Competitive Algorithm based on the Cosine Similarity ... 269 Maryam Houtinezhad completed her B.Sc. degree in computer engineering at Azad Islamic University, her M.Sc. degree at the University of Science and Research, and her Ph.D. at the Islamic Azad University, Ferdows Branch. Her interest research areas are image processing, pattern recognition and the application of optimization and web mining algorithms. Hamid Reza Ghaffary completed his B.Sc. degree in computer science at Sharif University of Technology and his M.Sc. degree at the University of South Tehran and his Ph.D. at Ferdowsi University. He is currently a faculty member and assistant professor at the Faculty of Computer Engineering, Ferdows Azad University. His interest research areas are machine learning, pattern recognition and image processing.
