Learning from obstructions: An effective deep learning approach for minimum vertex cover

Faisal N. Abu-Khzam, Mohamed M. Abd El-Wahab, Moussa Haidous, Noureldin Yosri

Research output: Contribution to journalArticlepeer-review

Abstract

Computational intractability has for decades motivated the development of a plethora of methodologies that mainly aim at a quality-time trade-off. The use of Machine Learning has finally emerged as one of the possible tools to obtain approximate solutions to NP-hard optimization problems. Recently, Dai et al. introduced a method for computing such approximate solutions for instances of the Vertex Cover problem. In this paper we consider the effectiveness of selecting a proper training strategy by considering special problem instances called obstructions that we believe carry some intrinsic properties of the problem. Capitalizing on the recent work of Dai et al. on Vertex Cover, and using the same case study as well as 19 other problem instances, we show the utility of using obstructions for training neural networks. Experiments show that training with obstructions results in a surprisingly huge reduction in number of iterations needed for convergence, thus gaining a substantial reduction in the time needed for training the model.

Original languageEnglish
Pages (from-to)1-12
Number of pages12
JournalAnnals of Mathematics and Artificial Intelligence
DOIs
Publication statusE-pub ahead of print - 2022

Fingerprint

Dive into the research topics of 'Learning from obstructions: An effective deep learning approach for minimum vertex cover'. Together they form a unique fingerprint.

Cite this