Informations générales
Maître Assistant
Marwa Hammami received the B.Sc. degree in computer science from the National School of Engineers of Carthage, University of Carthage, Tunisia, in 2012, the M.Sc. degree in computer science from the Faculty of Sciences of Tunis, University of Tunis El Manar, Tunisia, in 2014, and the Ph.D. degree in computer science with management from the Higher Institute of Management of Tunis in 2021.
Her current research interests include feature selection and construction, mono-objective, multi-objective, bi-level optimization, as well as machine learning and computer vision. She has published several papers in international journals and conferences, including Neural Computing and Applications and Memetic Computing.
Dr. Hammami also serves as a reviewer for several international journals such as IEEE Transactions on Evolutionary Computation, Swarm and Evolutionary Computation, and Applied Soft Computing Journal.
Équipes
Publications
-
2020Marwa Hammami, Slim Bechikh, Chih-Cheng Hung, Mohamed Makhlouf, Lamjed Ben Said
Feature Construction as a Bi-level Optimization Problem.
Neural Computing and Applications, 32(1), 13783–13804, 2020
Résumé
Feature selection and construction are important preprocessing techniques in data mining. They allow not only dimensionality reduction but also classification accuracy and efficiency improvement. While feature selection consists in selecting a subset of relevant feature from the original feature set, feature construction corresponds to the generation of new high-level features, called constructed features, where each one of them is a combination of a subset of original features. Based on these definitions, feature construction could be seen as a bi-level optimization problem where the feature subset should be defined first and then the corresponding (near) optimal combination of the selected features should be found. Motivated by this observation, we propose, in this paper, a bi-level evolutionary approach for feature construction. The basic idea of our algorithm, named bi-level feature construction genetic algorithm (BFC-GA), is to evolve an upper-level population for the task of feature selection, while optimizing the feature combinations at the lower level by evolving a follower population. It is worth noting that for each upper-level individual (feature subset), a whole lower-level population is optimized to find the corresponding (near) optimal feature combination (constructed feature). In this way, BFC-GA would be able to output a set of optimized constructed features that could be very informative to the considered classifier. A detailed experimental study has been conducted on a set of commonly used datasets with varying dimensions. The statistical analysis of the obtained results shows the competitiveness and the outperformance of our bi-level feature construction approach with respect to many state-of-the-art algorithms.
Marwa Hammami, Slim Bechikh, Mohamed Makhlouf, Chih Cheng-Hung, Lamjed Ben SaidClass Dependent Feature Construction as a Bi-level Optimization Problem
IEEE Congress on Evolutionary Computation. pp 1-8, 2020
Résumé
Feature selection and construction are important
pre-processing techniques in data mining. They allow not only
dimensionality reduction but also classification accuracy and
efficiency improvement. While feature selection consists in selecting a subset of relevant features from the original feature
set, feature construction corresponds to the generation of new
high-level features, called constructed features, where each one
of them is a combination of a subset of original features. However,
different features can have different abilities to distinguish
different classes. Therefore, it may be more difficult to construct
a better discriminating feature when combining features that are
relevant to different classes. Based on these definitions, feature
construction could be seen as a BLOP (Bi-Level Optimization
Problem) where the feature subset should be defined in the upper
level and the feature construction is applied in the lower level
by performing mutliple followers, each of which generates a set
class dependent constructed features. In this paper, we propose a
new bi-level evolutionary approach for feature construction called
BCDFC that constructs multiple features which focuses on distinguishing one class from other classes using Genetic Programming
(GP). A detailed experimental study has been conducted on six
high-dimensional datasets. The statistical analysis of the obtained
results shows the competitiveness and the outperformance of our
bi-level feature construction approach with respect to many stateof-art algorithms.Marwa Hammami, Slim Bechikh, Chih Cheng-Hung, Lamjed Ben SaidClass-Dependent Weighted Feature Selection as a Bi-Level Optimization Problem.
International Conference on Neural Information Processing. pp 269-278., 2020
Résumé
Feature selection aims at selecting relevant features from the original
feature set, but these features do not have the same degree of importance. This can
be achieved by feature weighting, which is a method for quantifying the capability of features to discriminate instances from different classes. Multiple feature
selection methods have shown that different feature subset can reduce the data
dimensionality and maintain or even improve the classification accuracy. However, different features can have different abilities to distinguish instances of one
class from the other classes, which makes the feature selection process a difficult task by finding the optimal feature subset weighting vectors for each class.
Motivated by this observation, feature selection and feature weighting could be
seen as a BLOP (Bi-Level Optimization Problem) where the feature selection is
performed in the upper level, and the feature weighting is applied in the lower
level by performing mutliple followers, each of which generates a set of weighting vectors for each class. Only the optimal feature subset weighting vector is
retrieved for each class. In this paper, we propose a bi-level evolutionary approach for class-dependent feature selection and weighting using Genetic Algorithm (GA), called Bi-level Class-Dependent Weighted Feature Selection (BCDWFS). The basic idea of our BCDWFS is to exploit the bi-level model for performing upper level feature selection and lower level feature weighting with the
aim of finding the optimal weighting vectors of a subset of features for each class.
Our approach has been assessed on ten datasets and compared to three existing
approaches, using three different classifiers for accuracy evaluation. Experimental results show that our proposed algorithm gives competitive and better results
with respect to the state-of-the-art algorithms. -
2019Marwa Hammami, Slim Bechikh, Chih Cheng-Hung, Lamjed Ben Said
Weighted- Feature Construction as a Bi-level Optimization Problem
IEEE Congress on Evolutionary Computation. pp 1604-161., 2019
Résumé
Feature selection and construction are important
pre-processing techniques in machine learning and data mining.
They may allow not only dimensionality reduction but also
classifier accuracy and efficiency improvement. Feature selection
aims at selecting relevant features from the original feature set,
which could be less informative to achieve good performance.
Feature construction may work well as it creates new highlevel features, but these features do not have the same degree
of importance, which makes the use of weighted-features construction a very challenging topic. In this paper, we propose a
bi-level evolutionary approach for efficient feature selection and
simultaneous feature construction and feature weighting, called
Bi-level Weighted-Features Construction (BWFC). The basic idea
of our BWFC is to exploit the bi-level model for performing
feature selection and weighted-features construction with the
aim of finding an optimal subset of features combinations. Our
approach has been assessed on six high-dimensional datasets and
compared against three existing approaches, using three different
classifiers for accuracy evaluation. Experimental results show
that our proposed algorithm gives competitive and better results
with respect to the state-of-the-art algorithms -
2018Marwa Hammami, Slim Bechikh, Chih Cheng-Hung, Lamjed Ben Said
A Multi-objective Hybrid Filter-Wrapper Evolutionary Approach for Feature Selection.
Memetic Computing (IF: 5.9). 11(2), 193-208., 2018
Résumé
Feature selection is an important pre-processing data mining task, which can reduce the data dimensionality and improve not
only the classification accuracy but also the classifier efficiency. Filters use statistical characteristics of the data as the evaluation
measure rather than using a classification algorithm. On the contrary, the wrapper process is computationally expensive
because the evaluation of every feature subset requires running the classifier on the datasets and computing the accuracy
from the obtained confusion matrix. In order to solve this problem, we propose a hybrid tri-objective evolutionary algorithm
that optimizes two filter objectives, namely the number of features and the mutual information, and one wrapper objective
corresponding to the accuracy. Once the population is classified into different non-dominated fronts, only feature subsets
belonging to the first (best) one are improved using the indicator-based multi-objective local search. Our proposed hybrid
algorithm, named Filter-Wrapper-based Nondominated Sorting Genetic Algorithm-II, is compared against several multiobjective and single-objective feature selection algorithms on eighteen benchmark datasets having different dimensionalities.
Experimental results show that our proposed algorithm gives competitive and better results with respect to existing algorithms.Marwa Hammami, Slim Bechikh, Chih Cheng-Hung, Lamjed Ben SaidA Multi-objective Hybrid Filter-Wrapper Approach For Feature Construction On High-Dimensional Data Using GP.
Proceedings of the IEEE Congress on Evolutionary Computation. pp 1-8, 2018
Résumé
Feature selection and construction are important
pre-processing techniques in data mining. They may allow
not only dimensionality reduction but also classifier accuracy
and efficiency improvement. These two techniques are of great
importance especially for the case of high-dimensional data.
Feature construction for high-dimensional data is still a very
challenging topic. This can be explained by the large search space
of feature combinations, whose size is a function of the number of
features. Recently, researchers have used Genetic Programming
(GP) for feature construction and the obtained results were
promising. Unfortunately, the wrapper evaluation of each feature
subset, where a feature can be constructed by a combination
of features, is computationally intensive since such evaluation
requires running the classifier on the data sets. Motivated by
this observation, we propose, in this paper, a hybrid multiobjective evolutionary approach for efficient feature construction
and selection. Our approach uses two filter objectives and one
wrapper objective corresponding to the accuracy. In fact, the
whole population is evaluated using two filter objectives. However,
only non-dominated (best) feature subsets are improved using an
indicator-based local search that optimizes the three objectives
simultaneously. Our approach has been assessed on six highdimensional datasets and compared with two existing prominent
GP approaches, using three different classifiers for accuracy
evaluation. Based on the obtained results, our approach is shown
to provide competitive and better results compared with two
competitor GP algorithms tested in this study
BibTeX
@article{hammami2020feature,
author = {Marwa Hammami and Slim Bechikh and Chien-Chung Hung and Mohamed Makhlouf and Lamjed Ben Said},
title = {Feature Construction as a Bi-level Optimization Problem},
journal = {Neural Computing and Applications},
year = {2020},
volume = {32},
number = {1},
pages = {13783–13804},
doi = {10.1007/s00521-020-04784-z}
}
BibTeX
@article{Hammami2018FeatureSelection,
author = {Marwa Hammami and Slim Bechikh and Chih-Cheng Hung and Lamjed Ben Said},
title = {A Multi-objective Hybrid Filter-Wrapper Evolutionary Approach for Feature Selection},
journal = {Memetic Computing},
volume = {11},
number = {2},
pages = {193–208},
year = {2018},
doi = {10.1007/s12293-018-0271-3},
issn = {1865-9284},
publisher = {Springer},
note = {Impact Factor: 5.9}
}
BibTeX
@inproceedings{Hammami2020ClassDependent,
author = {Marwa Hammami and Slim Bechikh and Mohamed Makhlouf and Chih-Cheng Hung and Lamjed Ben Said},
title = {Class Dependent Feature Construction as a Bi-level Optimization Problem},
booktitle = {Proceedings of the IEEE Congress on Evolutionary Computation (CEC)},
pages = {1–8},
year = {2020},
doi = {10.1109/CEC48606.2020.9185723},
publisher = {IEEE},
address = {Glasgow, UK}
}
BibTeX
@inproceedings{Hammami2020ClassDependentWeighted,
author = {Marwa Hammami and Slim Bechikh and Chih-Cheng Hung and Lamjed Ben Said},
title = {Class-Dependent Weighted Feature Selection as a Bi-Level Optimization Problem},
booktitle = {Neural Information Processing – Proceedings of the 27th International Conference on Neural Information Processing (ICONIP)},
pages = {269–278},
year = {2020},
doi = {10.1007/978-3-030-63823-8_31},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
volume = {12534},
address = {Bangkok, Thailand}
}
BibTeX
@inproceedings{hammami2019weighted,
author = {Marwa Hammami and Slim Bechikh and Chih-Chung Hung and Lamjed Ben Said},
title = {Weighted-Feature Construction as a Bi-level Optimization Problem},
booktitle = {Proceedings of the IEEE Congress on Evolutionary Computation (CEC)},
year = {2019},
pages = {1604–1611}
}
BibTeX
@inproceedings{hammami2018multi,
author = {Marwa Hammami and Slim Bechikh and Chih-Chung Hung and Lamjed Ben Said},
title = {A Multi-objective Hybrid Filter-Wrapper Approach For Feature Construction On High-Dimensional Data Using GP},
booktitle = {Proceedings of the IEEE Congress on Evolutionary Computation (CEC)},
year = {2018},
pages = {1–8},
publisher = {IEEE}
}


