Profil

Dupont Stéphane

Université de Mons - UMONS > Faculté Polytechnique > Service Information, Signal et Intelligence artificielle

Université de Mons - UMONS > Faculté des Sciences > Service d'Intelligence Artificielle

Université de Mons - UMONS > Faculté Polytechnique > Information, Signal et Intelligence artificielle

ORCID
0000-0003-3674-6747

Main Referenced Co-authors
DUTOIT, Thierry  (51)
Seddati, Omar  (26)
Frisson, Christian  (25)
Ris, Christophe (22)
El Haddad, Kevin  (20)
Main Referenced Keywords
Automatic speech recognition (2); Sketch-based image retrieval (2); Software (2); 35B09, 35B32, 35B33, 35J47 (1); Artificial intelligence (1);
Main Referenced Unit & Research Centers
CRTI - Centre de Recherche en Technologie de l'Information (97)
BIOSYS - Biosys (2)
Main Referenced Disciplines
Library & information sciences (175)
Electrical & electronics engineering (32)
Computer science (27)
Mathematics (7)
Neurosciences & behavior (2)

Publications (total 214)

The most downloaded
75 downloads
Delbrouck, J.-B., & Dupont, S. (2017). Modulating and attending the source image during encoding improves Multimodal Translation. Paper presented at NIPS 2017 Workshop on Visually-Grounded Interaction and Language (ViGIL), Long Beach, United States - California. https://hdl.handle.net/20.500.12907/42056
The most cited
444 citations (Scopus®)
Dupont, S., & Juergen, L. (01 September 2000). Audio-Visual Speech Modeling for Continuous Speech Recognition. IEEE Multimedia, 2 (3), 141-151. https://hdl.handle.net/20.500.12907/41402

Seddati, O., Dupont, S., Mahmoudi, S., & Dutoit, T. (2022). Towards Human Performance on Sketch-Based Image Retrieval. In Proceedings of 19th International Conference on Content-based Multimedia Indexing, CBMI 2022. Association for Computing Machinery. doi:10.1145/3549555.3549582
Peer reviewed

Seddati, O., Dupont, S., Mahmoudi, S., & Dutoit, T. (2022). Transformers and CNNs both Beat Humans on SBIR. Eprint/Working paper retrieved from https://orbi.umons.ac.be/20.500.12907/43859. doi:10.48550/arXiv.2209.06629

Brousmiche, M., Rouat, J., & Dupont, S. (September 2022). Multimodal Attentive Fusion Network for audio-visual event recognition. Information Fusion, 85, 52 - 59. doi:10.1016/j.inffus.2022.03.001
Peer Reviewed verified by ORBi

Hammoudeh, A. T. A., & Dupont, S. (19 June 2022). How does explicit orientation encoding affect image classification of ConvNets? Paper presented at CVPR 2022 workshop: Neurovision.
Peer reviewed

Hammoudeh, A. T. A., & Dupont, S. (2022). How does orientation encoding affect image classification of ConvNets?

Bohy, H., Hammoudeh, A. T. A., Maiorca, A., Dupont, S., & Dutoit, T. (2022). Analysis of Co-Laughter Gesture Relationship on RGB videos in Dyadic Conversation Context. Proceedings of LREC.
Peer reviewed

Hammoudeh, A. T. A., Vanderplaetse, B., & Dupont, S. (30 March 2022). Deep soccer captioning with transformer. Poster session presented at Infortech' Day, Mons, Belgium.

Hammoudeh, A. T. A., Vanderplaetse, B., & Dupont, S. (2022). Soccer captioning: dataset, transformer-based model, and triple-level evaluation. Procedia Computer Science, 210 (C), 104-111. doi:10.1016/j.procs.2022.10.125
Peer reviewed

Hammoudeh, A. T. A., Maiorca, A., Dupont, S., & Dutoit, T. (2022). Are there any Body-movement Differences between Women and Men when they Laugh? Proceedings of LREC.
Peer reviewed

Hammoudeh, A. T. A., & Dupont, S. (2022). The prediction of residential building consumption using profiling and time encoding. Procedia Computer Science, 210 (C), 7-11. doi:10.1016/j.procs.2022.10.113
Peer reviewed

Parian, M., Walzer, C., Rossetto, L., Heller, S., Dupont, S., & Schuldt, H. (2021). Gesture of Interest: Gesture Search for Multi-Person, Multi-Perspective TV Footage. International Conference on Content-Based Multimedia Indexing.
Peer reviewed

Delbrouck, J.-B., Tits, N., & Dupont, S. (2020). Modulated Fusion using Transformer for Linguistic-Acoustic Emotion Recognition. In NLP Beyond Text (NLPBT) - EMNLP 2020 (2020). -.
Peer reviewed

Brousmiche, M., Dupont, S., & Rouat, J. (2020). Intra and Inter-Modality Interactions for Audio-Visual Event Detection. In ACM International Conference on Multimedia. -.
Peer reviewed

Delbrouck, J.-B., Tits, N., Brousmiche, M., & Dupont, S. (2020). A Transformer-based joint-encoding for Emotion Recognition and Sentiment Analysis. In Second Grand Challenge and Workshop on Multimodal Language - ACL 2020 (2020). -. doi:10.18653/v1/2020.challengehml-1.1
Peer reviewed

Vanderplaetse, B., & Dupont, S. (2020). Improved Soccer Action Spotting Using Both Audio and Video Streams. Conference on Computer Vision and Pattern Recognition Workshops. Proceedings.
Peer reviewed

Brousmiche, M., Rouat, J., & Dupont, S. (2020). SECL-UMONS DATABASE FOR SOUND EVENT CLASSIFICATION AND LOCALIZATION. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings. doi:10.1109/ICASSP40776.2020.9053298
Peer reviewed

Delbrouck, J.-B., & Dupont, S. (2019). Adversarial reconstruction for Multi-modal Machine Translation. Eprint/Working paper retrieved from https://orbi.umons.ac.be/20.500.12907/42275.

Delbrouck, J.-B., Maiorca, A., Hubens, N., & Dupont, S. (2019). Modulated Self-attention Convolutional Network for VQA. In NeurIPS 2019 Workshop on Visually-Grounded Interaction and Language (ViGIL) (2019). -.
Peer reviewed

Brousmiche, M., Dupont, S., & Rouat, J. (2019). Audio-Visual Fusion And Conditioning With Neural Networks For Event Recognition. International Workshop on Machine Learning for Signal Processing.
Peer reviewed

Delbrouck, J.-B., & Dupont, S. (2018). Object-oriented Targets for Visual Navigation using Rich Semantic Representations. In NIPS 2018 Workshop on Visually-Grounded Interaction and Language (ViGIL). -.
Peer reviewed

Delbrouck, J.-B., & Dupont, S. (30 October 2018). UMONS Submission for WMT18 Multimodal Translation Task. Paper presented at Third Conference on Machine Translation, Brussels, Belgium.
Peer reviewed

Delbrouck, J.-B., & Dupont, S. (2018). Bringing back simplicity and lightliness into neural image captioning. ArXiv e-prints.
Peer reviewed

Pironkov, G., Wood, S. U. N., Dupont, S., & Dutoit, T. (2018). Investigating a Hybrid Learning Approach for Robust Automatic Speech Recognition. Paper presented at International Conference on Statistical Language and Speech Processing, Mons, Belgium.

Devillers, L., Rosset, S., Dubuisson Duplessis, G., Bechade, L., Yemez, Y., Turker, B. B., Sezgin, M., El Haddad, K., Dupont, S., Deléglise, P., Estève, Y., Lailler, C., Gilmartin, E., & Campbell, N. (2018). Multifaceted Engagement in Social Interaction with a Machine: the JOKER Project. Paper presented at Workshop on Large-scale Emotion Recognition and Analysis, Xi'an, China.

El Haddad, K., Heron, L., Kim, J., Lee, M., Dupont, S., Dutoit, T., & Truong, K. (2018). A Dyadic Conversation Dataset On Moral Emotions. Paper presented at Workshop on Large-scale Emotion Recognition and Analysis, Xi'an, China.

Dupont, S., Dutoit, T., & Curcio, G. (2018). Intelligence Artificielle (IA) et Apprentissage Profond (Deep Learning). LUMONS élément - Magazine de l'Université de Mons.

Dimitropoulos, K., Manitsaris, S., Tsalakanidou, F., Denby, B., Buchman, L., Dupont, S., Nikolopoulos, S., Kompatsiaris, I., Charisis, V., Hadjileontiadis, L., Pozzi, F., Cotescu, M., Ciftci, S., Katos, A., Manitsaris, A., & Grammalidis, N. (2018). A Multimodal Approach for the Safeguarding and Transmission of Intangible Cultural Heritage: The Case of i-Treasures. IEEE Intelligent Systems.
Peer Reviewed verified by ORBi

Pironkov, G., Dupont, S., & Dutoit, T. (2017). Investigating the Impact of the Training Data Volume for Robust Speech Recognition Using Multi-Task Learning. Paper presented at IEEE International Symposium on Signal Processing and Information Technology, Bilbao, Spain.

Delbrouck, J.-B., & Dupont, S. (2017). Modulating and attending the source image during encoding improves Multimodal Translation. Paper presented at NIPS 2017 Workshop on Visually-Grounded Interaction and Language (ViGIL), Long Beach, United States - California.

Bechade, L., El Haddad, K., Bourquin, J., Dupont, S., & Devillers, L. (2017). A Corpus for Experimental Study of Affect Bursts in Human-robot Interaction. Paper presented at 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents, Glasgow, United Kingdom.

Dupont, S. (26 October 2017). L'IA pour le multimedia et au-delà: Recherches de pointe et perspectives de collaborations. Paper presented at Business & Technology Club - IA/IoT - (Infopôle & Cluster TWIST), Mons, Belgium.

Pironkov, G., Dupont, S., Wood, S. U. N., & Dutoit, T. (2017). Noise and Speech Estimation As Auxiliary Tasks for Robust Speech Recognition. Paper presented at International Conference on Statistical Language and Speech Processing, Le Mans, France.

El Haddad, K., Torre, I., Gilmartin, E., Cakmak, H., Dupont, S., Dutoit, T., & Campbell, N. (2017). Introducing AmuS: The Amused Speech Database. Paper presented at International Conference on Statistical Language and Speech Processing, Le Mans, France.

Seddati, O., Dupont, S., Mahmoudi, S., & Amiri Parian, M. (2017). Towards Good Practices for Image Retrieval Based on CNN Features. Paper presented at IEEE International Conference on Computer Vision, Venice, Italy.

Seddati, O., Ben-Lhachemi, N., Dupont, S., & Mahmoudi, S. (2017). UMONS @ MediaEval 2017: Diverse Social Images Retrieval. Paper presented at MediaEval Benchmarking Initiative for Multimedia Evaluation, Dublin, Ireland.

Delbrouck, J.-B., & Dupont, S. (2017). An empirical study on the effectiveness of images in Multimodal Neural Machine Translation. Paper presented at Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark.

Delbrouck, J.-B., Dupont, S., & Seddati, O. (2017). Visually Grounded Word Embeddings and Richer Visual Features for Improving Multimodal Neural Machine Translation. Paper presented at GLU 2017 International Workshop on Grounding Language Understanding, Stockholm, Sweden.

Salvi, G., & Dupont, S. (2017). Proceedings of the GLU 2017 International Workshop on Grounding Language Understanding. Paper presented at GLU 2017 International Workshop on Grounding Language Understanding, Stockholm, Sweden.

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2017). Amused speech components analysis and classification: Towards an amusement arousal level assessment system. Computers and Electrical Engineering.
Peer Reviewed verified by ORBi

Seddati, O., Dupont, S., & Mahmoudi, S. (2017). Triplet Networks Feature Masking for Sketch-Based Image Retrieval. Lecture Notes in Computer Science.
Peer reviewed

Rouat, J., Brousmiche, M., Brodeur, S., & Dupont, S. (21 June 2017). Scene Interpretation for Objects Recognition and Manipulation and Partnerships. Poster session presented at CHIST-ERA Conference 2017, Krakow, Poland.

Seddati, O., Dupont, S., & Mahmoudi, S. (2017). Quadruplet Networks for Sketch-Based Image Retrieval. ACM on International Conference on Multimedia Retrieval.
Peer reviewed

Seddati, O., Dupont, S., & Mahmoudi, S. (2017). DeepSketch 3: Analyzing deep neural networks features for better sketch recognition and sketch-based image retrieval. Multimedia Tools and Applications, 1-27. doi:10.1007/s11042-017-4799-2, 2017
Peer Reviewed verified by ORBi

Alivizatou, M., Kitsikidis, A., Tsalakanidou, F., Dimitropoulos, K., Chantas, G., Nikolopoulos, S., Al Kork, S., Denby, B., Buchman, L., Adda-Decker, M., Pillot-Loiseau, C., Tilmanne, J., Dupont, S., Picart, B., Pozzi, F., Ott, M., Yilmaz, E., Charisis, V., Hadjidimitriou, S., & Grammalidis, N. (2017). Intangible Cultural Heritage and New Technologies: Challenges and Opportunities for Cultural Preservation and Development. In Mixed Reality and Gamification for Cultural Heritage. Springer International Publishing.

Seddati, O., Delbrouck, J.-B., Dupont, S., & Mahmoudi, S. (25 April 2017). Deep Features for Big Data. Poster session presented at Journée scientifique du Pôle hainuyer 'Les données au coeur de notre devenir: les enjeux des big data, Tournai, e-campus, Belgium.

Wood, S. U. N., Rouat, J., Dupont, S., & Pironkov, G. (2017). Blind Speech Separation and Enhancement With GCC-NMF. IEEE/ACM Transactions on Audio, Speech and Language Processing.
Peer Reviewed verified by ORBi

Delbrouck, J.-B., & Dupont, S. (2017). Multimodal Compact Bilinear Pooling for Multimodal Neural Machine Translation. ArXiv e-prints.

Rossetto, L., Giangreco, I., Tanase, C., Schuldt, H., Dupont, S., & Seddati, O. (2017). Enhanced Retrieval and Browsing in the IMOTION System. Paper presented at Conference on Multimedia Modeling, Reykjavik, Iceland.

Pironkov, G., Dupont, S., & Dutoit, T. (2016). I-Vector Estimation as Auxiliary Task for Multi-Task Learning based Acoustic Modeling for Automatic Speech Recognition. Paper presented at Spoken Language Technology Workshop, San Diego, United States.

Pironkov, G., Dupont, S., & Dutoit, T. (2016). Speaker-Aware Multi-Task Learning for Automatic Speech Recognition. Paper presented at International Conference on Pattern Recognition, Cancun, Mexico.

El Haddad, K., Cakmak, H., Gilmartin, E., Dupont, S., & Dutoit, T. (2016). Towards a Listening Agent: A System Generating Audiovisual Laughs and Smiles to Show Interest. Paper presented at International Conference on Multimodal Interfaces, Tokyo, Japan.

Seddati, O., Dupont, S., & Mahmoudi, S. (2016). DeepSketch2Image: Deep Convolutional Neural Networks for Partial Sketch Recognition and Image Retrieval. ACM on Multimedia Conference, Amsterdam, The Netherlands, 2016, 739-741.
Peer reviewed

Tanase, C., Rossetto, L., Giangreco, I., Schuldt, H., Dupont, S., & Seddati, O. (2016). The IMOTION System at TRECVID 2016: The Ad-Hoc Video Search Task. Paper presented at TREC Video Retrieval Evaluation, .

Pironkov, G., Dupont, S., & Dutoit, T. (2016). Speaker-Aware Long Short-Term Memory Multi-Task Learning for Speech Recognition. Paper presented at European Signal Processing Conference, Budapest, Hungary.

El Haddad, K., Cakmak, H., Sulir, M., Dupont, S., & Dutoit, T. (2016). Audio Affect Burst Synthesis: A Multilevel Synthesis System for Emotional Expressions. Paper presented at European Signal Processing Conference, Budapest, Hungary.

Grammalidis, N., Dimitropoulos, K., Tsalakanidou, F., Kitsikidis, A., Roussel, P., Denby, B., Chawah, P., Buchman, L., Dupont, S., Laraba, S., Picart, B., Tits, M., Tilmanne, J., Hadjidimitriou, S., Hadjileontiadis, L., Charisis, V., Volioti, C., Stergiaki, A., Manitsaris, A., & Manitsaris, S. (2016). The i-Treasures Intangible Cultural Heritage dataset. Paper presented at IEEE Workshop on Movement and Computing, Thessaloniki, Greece.

Seddati, O., Dupont, S., & Mahmoudi, S. (2016). DeepSketch 2: Deep Convolutional Neural Networks for Partial Sketch Recognition. International Workshop on Content-based Multimedia Indexing, CBMI 2016, Bucharest, Romania, June 15-17, 2016.
Peer reviewed

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2016). Laughter and Smile Processing for Human-Computer Interactions. Paper presented at Workshop 'Just talking - casual talk among humans and machines' of LREC 2016, Portorož, Slovenia.

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2016). AVAB-DBS: an Audio-Visual Affect Bursts Database for Synthesis. Paper presented at Tenth International Conference on Language Resources and Evaluation (LREC 2016), Portorož, Slovenia.

Seddati, O., Dupont, S., & Mahmoudi, S. (2016). Réseaux de Neurones Convolutionnels Profonds pour la Reconnaissance d'Action dans les Vidéos. Paper presented at CORESA 2016 - COmpression et REprésentation des Signaux Audiovisuels, Nancy, France.

Pironkov, G., Dupont, S., & Dutoit, T. (2016). Multi-Task Learning for Speech Recognition: An Overview. Paper presented at The European Symposium on Artificial Neural Networks, Bruges, Belgium.

Çiftçi, S., Grammalidis, N., Dimitropoulos, K., Kitsikidis, A., Tsalakanidou, F., Dagnino, F., Pozzi, F., Leboullenger, C., Chawah, P., Laraba, S., Dupont, S., Manitsaris, S., Volioti, C., Charisis, V., & Hadjileontiadis, L. (2016). Final Version of 3D Visualization for Sensorimotor Learning - FP7 i-Treasures Deliverable 5.5.

Dupont, S., Cakmak, H., Curran, W., Dutoit, T., Hofmann, J., McKeown, G., Pietquin, O., Platt, T., Ruch, W., & Urbain, J. (2016). Laughter Research: A Review of the ILHAIRE Project. In Toward Robotic Socially Believable Behaving Systems. Springer International Publishing.

Tanase, C., Giangreco, I., Rossetto, L., Schuldt, H., Seddati, O., Dupont, S., Altiok, O. C., & Sezgin, M. (2016). Semantic Sketch-Based Video Retrieval with Autocompletion. Paper presented at International Conference on Intelligent User Interfaces, Sonoma, United States - California.

Schuldt, H., Dupont, S., Giangreco, I., Rossetto, L., Sahillioglu, Y., Seddati, O., Sen, C., Sezgin, M., Tanase, C., & Yildirim, D. (2016). IMOTION - Intelligent MultiModal Augmented Video Motion Retrieval System - Periodic Report 2.

Denby, B., Leboullenger, C., Hakoun, A., Roussel, P., Hadjidimitriou, S., Hadjileontiadis, L., Charisis, V., Grammalidis, N., Tsalakanidou, F., Dimitropoulos, K., Kitsikidis, A., Laraba, S., Tilmanne, J., Dupont, S., Manitsaris, S., Moutarde, F., bouzos, O., Chawah, P., Crevier-Buchman, L., & Manitsaris, A. (2016). Final Version of ICH Capture and Analysis Modules - FP7 i-Treasures Deliverable 3.4.

Seddati, O., Dupont, S., & Mahmoudi, S. (2016). Video Motion Feature Extractors, 1st Prototype - IMOTION Deliverable 2.2.

El Haddad, K., Dupont, S., & Dutoit, T. (2016). Affect bursts generation - v1 - JOKER Deliverable 5.3.

El Haddad, K., Dupont, S., & Dutoit, T. (2016). Speech Synthesis - v1 - JOKER Deliverable 5.2.

Chantas, G., Nikolopoulos, S., Kompatsiaris, I., Karavarsamis, S., Ververidis, D., Chatzilari, E., Liaros, G., Chalikias, A.-P., Nikolaidis, S., & Dupont, S. (2016). Final Version of Multimodal Analysis, Fusion and Semantic Media Interpretation - FP7 i-Treasures Deliverable 4.3.

Denby, B., Leboullenger, C., Roussel, P., Hadjidimitriou, S., Hadjileontiadis, L., Charisis, V., Grammalidis, N., Tsalakanidou, F., Dimitropoulos, K., Kitsikidis, A., Laraba, S., Tilmanne, J., Dupont, S., Manitsaris, S., Moutarde, F., bouzos, O., Chawah, P., Crevier-Buchman, L., Volioti, C., & Manitsaris, A. (2016). Final Report on ICH Capture and Analysis - FP7 i-Treasures Deliverable 3.3.

Rossetto, L., Giangreco, I., Heller, S., Tanase, C., Schuldt, H., Dupont, S., Seddati, O., Sezgin, M., Altiok, O. C., & Sahillioglu, Y. (2016). IMOTION - Searching for Video Sequences Using Multi-Shot Sketch Queries. Lecture Notes in Computer Science.
Peer reviewed

Rossetto, L., Giangreco, I., Tanase, C., Schuldt, H., Dupont, S., Seddati, O., Sezgin, M., & Sahillioglu, Y. (2016). iAutoMotion - an Autonomous Content-Based Video Retrieval Engine. Lecture Notes in Computer Science.
Peer reviewed

El Haddad, K., Dupont, S., Cakmak, H., & Dutoit, T. (2015). Shaking and Speech-Smile Vowels Classification: An Attempt at Amusement Arousal Estimation from Speech Signals. IEEE Global Conference on Signal and Information Processing.
Peer reviewed

Pironkov, G., Dupont, S., & Dutoit, T. (2015). INVESTIGATING SPARSE DEEP NEURAL NETWORKS FOR SPEECH RECOGNITION. Paper presented at Automatic Speech Recognition & Understanding, .

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2015). Towards a Level Assessment System of Amusement in Speech Signals: Amused Speech Components Classification. Paper presented at IEEE International Symposium on Signal Processing and Information Technology, .

El Haddad, K., Cakmak, H., Moinet, A., Dupont, S., & Dutoit, T. (2015). An HMM Approach for Synthesizing Amused Speech with a Controllable Intensity of Smile. Paper presented at IEEE International Symposium on Signal Processing and Information Technology, .

Dupont, S. (2015). VideoSketcher - Innovative Query Modes for Manipulating and Searching Video through Motion and Sound - Final Demo.

Devillers, L., Rossetto, S., Dubuisson Duplessis, G., Sehili, M. A., Béchade, L., Delaborde, A., Gossart, C., Letard, V., Yang, F., Yemez, Y., T¨urker, B. B., Sezgin, M., El Haddad, K., Dupont, S., Luzzati, D., Estève, Y., Gilmartin, E., & Campbell, N. (2015). Multimodal Data Collection of Human-Robot Humorous Interactions in the JOKER Project. Paper presented at Affective Computing and Intelligent Interaction, .

Manitsaris, A., Tsagkaris, A., Marchavilas, G., Veranis, G., Chawah, P., Crevier-Buchman, L., Dupont, S., Tilmanne, J., Tsalakanidou, F., Kitsikidis, A., Dimitropoulos, K., Grammalidis, N., Chantas, G., Karavarsamis, S., Denby, B., Leboullenger, C., Ceregini, A., Dagnino, F., Pozzi, F., & Yilmaz, E. (2015). Final Report on System Specification - FP7 i-Treasures Deliverable 2.4.

Seddati, O., Emre, K., Pironkov, G., Dupont, S., Mahmoudi, S., & Dutoit, T. (2015). UMons at MediaEval 2015 Affective Impact of Movies Task including Violent Scenes Detection. IEEE Multimedia.
Peer Reviewed verified by ORBi

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2015). Breath and Repeat: An Attempt at Enhancing Speech-Laugh Synthesis Quality. Paper presented at European Signal Processing Conference, .

Seddati, O., Dupont, S., & Mahmoudi, S. (2015). DeepSketch: Deep convolutional neural networks for sketch recognition and similarity search. International Workshop on Content-Based Multimedia Indexing, CBMI 2015, Prague, Czech Republic, June 10-12, 2015.
Peer reviewed

El Haddad, K., Dupont, S., D'alessandro, N., & Dutoit, T. (2015). An HMM-based Speech-smile Synthesis System: An Approach for Amusement Synthesis. Paper presented at 3rd Intl Workshop on Emotion Representation, Analysis and Synthesis in Continuous Time and Space (EmoSPACE'15), Ljubljana, Slovenia.

Picart, B., Brognaux, S., & Dupont, S. (2015). Analysis and Automatic Recognition of Human Beatbox Sounds: a Comparative Study. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

El Haddad, K., Dupont, S., Urbain, J., & Dutoit, T. (2015). Speech-Laughs: an HMM-based Approach for Amused Speech Synthesis. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

El Haddad, K., Cakmak, H., Dupont, S., & Dutoit, T. (2015). Towards a Speech Synthesis System with Controllable Amusement Levels. Paper presented at 4th Interdisciplinary Workshop on Laughter and Other Non-Verbal Vocalisations in Speech, Enschede, Netherlands.

El Haddad, K., Moinet, A., Cakmak, H., Dupont, S., & Dutoit, T. (2015). Using MAGE for Real Time Speech-Laugh Synthesis. Paper presented at 4th Interdisciplinary Workshop on Laughter and Other Non-Verbal Vocalisations in Speech, Enschede, Netherlands.

Yilmaz, E., Ugurca, D., Sahin, C., Dagnino, F., Ott, M., Pozzi, F., Dimitropoulos, K., Tsalakanidou, F., Kitsikidis, A., Al Kork, S., Xu, K., Denby, B., Roussel, P., Chawah, P., Buchman, L., Adda-Decker, M., Dupont, S., Picart, B., Tilmanne, J., & Grammalidis, N. (2015). Novel 3D Game-like Applications Driven by Body Interactions for Learning Specific Forms of Intangible Cultural Heritage. Paper presented at International Conference on Computer Vision Theory and Applications, .

Al Kork, S., Ugurca, D., Sahin, C., Chawah, P., Buchman, L., Adda-Decker, M., Xu, K., Denby, B., Roussel, P., Picart, B., Dupont, S., Tsalakanidou, F., Kitsikidis, A., Dagnino, F., Ott, M., Pozzi, F., Stone, M., & Yilmaz, E. (2015). A Novel Human Interaction Game-Like application to Learn, Perform and Evaluate Modern Contemporary Singing: Human Beat Box. Paper presented at International Conference on Computer Vision Theory and Applications, .

Seddati, O., Dupont, S., & Mahmoudi, S. (10 March 2015). DNN for action recognition in videos. Poster session presented at 8ème édition de la Matinée de Chercheurs, Mons, Belgium.

Seddati, O., Dupont, S., & Mahmoudi, S. (2015). Report on Video Motion Feature Extraction - IMOTION Deliverable 2.1.

Schuldt, H., Dupont, S., Giangreco, G., Rossetto, L., Sahillioglu, Y., Seddati, O., Sen, C., Sezgin, M., Tanase, C., & Yildirim, D. (2015). IMOTION - Intelligent MultiModal Augmented Video Motion Retrieval System - Periodic Report 1.

Rossetto, L., Giangreco, I., Schuldt, H., Dupont, S., Seddati, O., Sezgin, M., & Sahillioglu, Y. (01 January 2015). IMOTION - A Content-Based Video Retrieval Engine. Lecture Notes in Computer Science, 8936, 255-260.
Peer reviewed

Dupont, S., Urbain, J., Pelachaud, C., André, E., Wagner, J., Volpe, G., Mancini, M., Berthouze, N., Griffin, H., Curran, W., McKeown, G., Ruch, W., Hofmann, J., Platt, T., Pietquin, O., Piot, B., & Bantegnie, E. (2014). FP7 ILHAIRE Periodic Report Year 3 - Project Objectives, Work Progress and Achievements.

Dupont, S. (2014). FP7 ILHAIRE Periodic Report Year 3 - Publishable Summary (Project Nbr. 270780).

Lefebvre, D., Thunissen, M., Vandem Bemden, G., Dontaine, E., Rochette, A., Rodriguez, S., Giot, R., Frisson, C., Dupont, S., & Dutoit, T. (2014). Rapport Technique et Scientifique Année Projet RW GREENTIC SonixTrip (1317970) : Solution End-to-End intégrée (de l'authoring à la liseuse) de création de versions numériques interactives et sonorisées de bandes dessinées.

Ruch, W., Platt, T., Hofmann, J., Niewiadomski, R., Urbain, J., Mancini, M., & Dupont, S. (18 November 2014). Gelotophobia and the Challenges of Implementing Laughter into Virtual Agents Interactions. Frontiers in Human Neuroscience, 8 (928). doi:10.3389/fnhum.2014.00928
Peer reviewed

Dupont, S., Urbain, J., Pelachaud, C., André, E., Wagner, J., Volpe, G., Mancini, M., Berthouze, N., Griffin, H., Curran, W., McKeown, G., Ruch, W., Hofmann, J., Platt, T., Pietquin, O., Piot, B., & Bantegnie, E. (2014). FP7 ILHAIRE Periodic Report Year 3 - Project Management and Use of Resources.

Frisson, C., Dupont, S., Yvart, W., Riche, N., Siebert, X., & Dutoit, T. (2014). A Proximity Grid Optimization Method to Improve Audio Search for Sound Design. Paper presented at International Symposium/Conference on Music Information Retrieval, .

Urbain, J., Dupont, S., Pelachaud, C., & McKeown, G. (2014). Laughter Similarity Browsing - FP7 ILHAIRE Deliverable 1.5.

Frisson, C., Dupont, S., Yvart, W., Riche, N., Siebert, X., & Dutoit, T. (2014). AudioMetro: Directing Search for Sound Designers Through Content-based Cues. Paper presented at Audio Mostly Conference, . doi:10.1145/2636879.2636880

Nixon, L., Baltussen, L. B., Grisard, F., Dupont, S., & Patz, N. (2014). Scenario Demonstrators - FP7 LinkedTV Deliverable 6.2.

Giangreco, G., Rossetto, L., Schuldt, H., Sezgin, M., Sahillioglu, Y., Dupont, S., & Seddati, O. (2014). Initial Requirements and System Specification - IMOTION Deliverable 1.1.

Frisson, C., Riche, N., Coutrot, A., Delestage, C.-A., Dupont, S., Ferhat, O., Guyade, N., Mahmoudi, S., Mancas, M., Mital, P. K., Ech aniz, A. P., Rocca, F., Rochette, A., & Yvart, W. (2014). Auracle: how are salient cues situated in audiovisual content? The 10th International Summer Workshop on Multimedia Interfaces.

Mancini, M., Ach, L., Bantegnie, E., Baur, T., Berthouze, N., Datta, D., Ding, Y., Dupont, S., Griffin, H., Lingenfelser, F., Niewiadomski, R., Pelachaud, C., Pietquin, O., Piot, B., Urbain, J., Volpe, G., & Wagner, J. (2014). Laugh When You're Winning. IFIP Advances in Information and Communication Technology.
Peer Reviewed verified by ORBi

Grisard, F., Kayalar, C., Alacam, S., Özgün, A., Ipek, Y., & Dupont, S. (2014). Social Documentary: An interactive and evolutive installation to explore crowd-source media content. Paper presented at 2nd International Workshop on Interactive Content Consumption at TVX 2014 (WSICC2014), Newcastle upon Tyne, United Kingdom.

Grisard, F., Kayalar, C., Alaçam, S., Balaban, O., Ipek, Y., & Dupont, S. (24 June 2014). Social Documentary: An interactive and evolutive installation to explore crowd-sourced media content. Poster session presented at 2nd International Workshop on Interactive Content Consumption at TVX 2014 (WSICC2014), Brussels, Belgium.

Yvart, W., Dutoit, T., & Dupont, S. (2014). Une approche info-communicationnelle des librairies musicales en ligne. Paper presented at Proc. SFSIC'14. XIXe congrès SFSIC, Toulon, France.

Urbain, J., Cakmak, H., Charlier, A., Denti, M., Dutoit, T., & Dupont, S. (2014). Arousal-Driven Synthesis of Laughter. IEEE Journal of Selected Topics in Signal Processing.
Peer Reviewed verified by ORBi

Denby, B., Al Kork, S., Hakoun, A., Xu, K., Roussel, P., Stone, M., Manitsaris, A., Kourvoulis, G., Katos, A., Glushkova, A., Gatziaki, V., Volioti, C., Grammalidis, N., Dimitropoulos, K., Tsalakanidou, F., Kitsikidis, A., Adda-Decker, M., Crevier-Buchman, L., Pillot-Loiseau, C., & Manitsaris, S. (2014). First Version of ICH Capture and Analysis Modules - FP7 i-Treasures Deliverable 3.2.

Denby, B., Al Kork, S., Hakoun, A., Xu, K., Roussel, P., Stone, M., Manitsaris, A., Kourvoulis, G., Katos, A., Glushkova, A., Gatziaki, V., Volioti, C., Grammalidis, N., Dimitropoulos, K., Tsalakanidou, F., Kitsikidis, A., Adda-Decker, M., Crevier-Buchman, L., Pillot-Loiseau, C., & Manitsaris, S. (2014). First Report on ICH Capture and Analysis - FP7 i-Treasures Deliverable 3.1.

Frisson, C., Rocca, F., Dupont, S., Dutoit, T., Grobet, D., Giot, R., El Brouzi, M., Bouaziz, S., Yvart, W., & Leleu-Merviel, S. (2014). Tangible Needle, Digital Haystack: Tangible Interfaces for Reusing Media Content Organized by Similarity. Paper presented at Tangible and Embedded Interaction, . doi:10.1145/2540930.2540983

Dupont, S. (22 January 2014). FP7 FET ILHAIRE Project - Testimony of Belgian participation to the FET Programme. Paper presented at H2020 FET Infomation Day Organized by FNRS/FWO, Brussels, Belgium.

Sabato, C., Giraudet, A., Delattre, V., Desnos, Y., Frisson, C., Giot, R., Yvart, W., Rocca, F., Dupont, S., Vandem Bemden, G., Leleu-Merviel, S., & Dutoit, T. (10 January 2014). Scenarizing CADastre Exquisse: A Crossover between Snoezeling in Hospitals/Domes, and Authoring/Experiencing Soundful Comic Strips. Lecture Notes in Computer Science, 8326, 22-33. doi:10.1007/978-3-319-04117-9_3
Peer reviewed

Dimitropoulos, K., Manitsaris, S., Tsalakanidou, F., Nikolopoulos, S., Denby, B., Al Kork, S., Crevier-Buchman, L., Pillot-Loiseau, C., Adda-Decker, M., Dupont, S., Tilmanne, J., Ott, M., Alivizatou, M., Yilmaz, E., Hadjileontiadis, L., Charisis, V., Deroo, O., Manitsaris, A., Kompatsiaris, I., & Grammalidis, N. (2014). Capturing the Intangible: An Introduction to the i-Treasures Project. Paper presented at International Conference on Computer Vision Theory and Applications, .

Frisson, C., Keyaerts, G., Grisard, F., Dupont, S., Ravet, T., Zajega, F., Colmenares Guerra, L., Todoroff, T., & Dutoit, T. (30 November 2013). MashtaCycle: On-Stage Improvised Audio Collage by Content-Based Similarity and Gesture Recognition. Lecture Notes in Computer Science, 124, 114-123.
Peer reviewed

Reboursière, L., & Dupont, S. (30 November 2013). EGT: Enriched Guitar Transcription. Lecture Notes in Computer Science, 124, 163-168.
Peer reviewed

Dupont, S., Urbain, J., Pelachaud, C., André, E., Wagner, J., Volpe, G., Mancini, M., Berthouze, N., Griffin, H., Curran, W., McKeown, G., Ruch, W., Hofmann, J., Platt, T., Pietquin, O., Piot, B., & Ach, L. (2013). FP7 ILHAIRE Periodic Report Year 2 - Project Objectives, Work Progress and Achievements.

Dupont, S., & Ravet, T. (2013). Improved Audio Classification using a Novel Non-Linear Dimensionality Reduction Ensemble Approach. Paper presented at 14th International Society for Music Information Retrieval Conference, ISMIR, Curitiba, Brazil.

Dupont, S. (2013). FP7 ILHAIRE Periodic Report Year 2 - Publishable Summary (Project Nbr. 270780).

Nixon, L., Patz, N., Baltussen, L. B., Verbruggen, E., Blom, J., Dupont, S., Hildebrand, M., Ebermann, J., & Milanova, V. (2013). FP7 LinkedTV Deliverable 6.2 - Scenario Demonstrators.

Frisson, C., Schayes, E., Uyttenhove, S., Dupont, S., Giot, R., & Dutoit, T. (2013). Designing artfully-mediated interactive surfaces organizing media collections. Paper presented at ACM Interactive Tabletops and Surfaces (ITS), Workshop: Collaboration meets Interactive Surfaces: Walls, Tables, Tablets and Phones, St Andrews, United Kingdom.

Mancini, M., Niewiadomski, R., Volpe, G., Aung, H., Berthouze, N., Griffin, H., Dupont, S., Urbain, J., Hofmann, J., Platt, T., Ruch, W., Wagner, J., Lingenfelser, F., André, E., & Pelachaud, C. (2013). FP7 ILHAIRE Deliverable 2.2 - Improved Multimodal Analysis and Recognition.

Urbain, J., Niewiadomski, R., Huang, J., Ding, Y., Datta, D., Bantegnie, E., Ach, L., Mancini, M., Pelachaud, C., Dupont, S., & Dutoit, T. (2013). FP7 ILHAIRE Deliverable 3.2 - Improved Multimodal Audio-Facial-Gestural Laugh Generation and Synthesis.

Dupont, S., Ravet, T., picard-limpens, C., & Frisson, C. (2013). Nonlinear Dimensionality Reduction Approaches applied to Music and Textural Sounds. Paper presented at IEEE International Conference on Multimedia and Expo (ICME 2013), San Jose, United States - California.

Frisson, C., Keyaerts, G., Grisard, F., Dupont, S., Ravet, T., Zajega, F., Colmenares Guerra, L., Todoroff, T., & Dutoit, T. (2013). MashtaCycle: On-Stage Improvised Audio Collage by Content-Based Similarity and Gesture Recognition. Paper presented at INTETAIN 2013, Mons, Belgium.

Niewiadomski, R., Hofmann, J., Urbain, J., Platt, T., Wagner, J., Piot, B., Cakmak, H., Pammi, S., Baur, T., Dupont, S., Geist, M., Lingenfelser, F., McKeown, G., Pietquin, O., & Ruch, W. (2013). Laugh-aware virtual agent and its impact on user amusement. Paper presented at Twelfth International Conference on Autonomous Agents and Multiagent Systems (AAMAS2013), Saint Paul, United States - Minnesota.

Frisson, C., Dupont, S., Moinet, A., Picard, C., Ravet, T., Siebert, X., & Dutoit, T. (2013). Video­Cy­cle: user-friendly nav­i­ga­tion by sim­i­lar­ity in video data­bases. Paper presented at Pro­ceed­ings of the Mul­ti­me­dia Mod­el­ing Con­fer­ence (MMM), Video Browser Show­down ses­sion, Huangshan, China.

Frisson, C., Dupont, S., Moinet, A., Picard, C., Ravet, T., Dutoit, T., & Siebert, X. (07 January 2013). VideoCycle: User-Friendly Navigation by Similarity in Video Databases. Lecture Notes in Computer Science, 7733, 550-553.
Peer reviewed

Dupont, S. (2012). FP7 ILHAIRE Periodic Report Year 1 Publishable Summary (Project Nbr. 270780).

Urbain, J., Niewiadomski, R., Hofmann, J., Bantegnie, E., Baur, T., Berthouze, N., Cakmak, H., Cruz, R. T., Dupont, S., Geist, M., Griffin, H., Lingenfelser, F., Mancini, M., Miranda, M., McKeown, G., Pammi, S., Pietquin, O., Piot, B., Platt, T., & Wagner, J. (2012). Laugh Machine. Paper presented at eNTERFACE'12, Metz, France.

Frisson, C., Dupont, S., Leroy, J., Moinet, A., Ravet, T., Siebert, X., & Dutoit, T. (2012). LoopJam: turning the dance floor into a collaborative instrumental map. Paper presented at Proceedings of the 12th Conference on New Interfaces for Musical Expression (NIME'12), Ann Arbor, Michigan, United States - Michigan.

Reboursière, L., Lähdeoja, O., Drugman, T., Dupont, S., Picard-Limpens, C., & Riche, N. (2012). Left and right-hand guitar playing techniques detection. Paper presented at 12th Conference on New Interfaces for Musical Expression (NIME'12), Ann Arbor, Michigan, United States - Michigan.

Frisson, C., Dupont, S., Leroy, J., Moinet, A., Ravet, T., Siebert, X., & Dutoit, T. (2012). LoopJam : une Carte Musicale Collaborative sur la Piste de Danse. Paper presented at Actes des Journées d'Informatique Musicale (JIM 2012), Mons, Belgium.

Lähdeoja, O., Reboursière, L., Drugman, T., Dupont, S., Picard-Limpens, C., & Riche, N. (2012). Detection des Techniques de Jeu de la Guitare. Paper presented at Journées d'Informatique Musicale (JIM 2012), Mons, Belgium.

Stanoevska-Slabeva, K., Ebermann, J., Köhler, J., Pauer, A., Nixon, L., Mezaris, V., Svatek, V., Dupont, S., Hardman, L., Troncy, R., Fricke, R., Oomen, J., Gural, J., & de Abreu Pereira, N. (2012). FP7 LinkedTV Deliverable 8.1 Exploitation Plan for the Project.

Reboursière, L., Lähdeoja, O., Chessini Bose, R., Drugman, T., Dupont, S., Picard-Limpens, C., & Riche, N. (2011). Guitar As Controller. Quarterly Progress Scientific Report of the Numediart Research Program.

Dupont, S., Frisson, C., Leroy, J., Moinet, A., Ravet, T., Siebert, X., & Urbain, J. (2011). LoopJam. Paper presented at NEM 2011, Torino, Italy.

Frisson, C., Dupont, S., Siebert, X., & Dutoit, T. (2011). Similarity in media content: digital art perspectives. Paper presented at ISEA 2011, Istanbul, Turkey.

Pietquin, O., & Dupont, S. (2011). Incorporating Laughter into Human-Avatar Interactions: Research and Evaluation - The ILHAIRE project.

Frisson, C., Dupont, S., Leroy, J., Moinet, A., Ravet, T., & Siebert, X. (01 June 2011). LoopJam: a collaborative musical map on the dance floor. Quarterly Progress Scientific Report of the Numediart Research Program, 4 (2), 37-40.

Dupont, S., Frisson, C., Urbain, J., Mahmoudi, S., & Siebert, X. (01 March 2011). MEDIABLENDER : INTERACTIVE MULTIMEDIA SEGMENTATION. Quarterly Progress Scientific Report of the Numediart Research Program, 4 (1), 1-6.

Tardieu, D., Siebert, X., Mazzarino, B., Chessini Bose, R., Dubois, J., Dupont, S., Varni, G., & Visentin, A. (01 September 2010). Browsing a dance video collection: dance analysis and interface design. Journal on Multimodal User Interfaces, 4 (1), 37-46.

Tardieu, D., Dupont, S., Siebert, X., Mazzarino, B., & Blumenthal, B. (2010). An Interactive installation for browsing a Dance Video Database. Paper presented at IMIDA2010, Workshop on Interactive Multimedia Installations and Digital Art, at ICME2010, Singapore, Singapore.

Frisson, C., Dupont, S., Siebert, X., Tardieu, D., Dutoit, T., & Macq, B. (2010). DeviceCycle: rapid and reusable prototyping of gestural interfaces, applied to audio browsing by similarity. Paper presented at Conference on New Interfaces for Musical Expression, Sydney, Australia.

Dupont, S., Frisson, C., Siebert, X., & Tardieu, D. (2010). Browsing sound and music librairies by similarity. Paper presented at 128th Convention of the Audio Engineering Society, .

Dutoit, T., & Dupont, S. (2010). Speech processing. In Multimodal signal processing - Theory and applications for human-computer interaction (pp. 25-61). Elsevier.

Siebert, X., Dupont, S., Frisson, C., & Tardieu, D. (2009). MultiMediaCycle: consolidating the hyforge framework towards improved scalability and usability. Paper presented at QPSR of the numediart research program, .

Tardieu, D., Dubois, J., Chessini Bose, R., Dupont, S., Mazzarino, B., Hidot, S., Moinet, A., Siebert, X., Varni, G., & Visentin, A. (01 September 2009). Video Navigation Tool: Application to browsing a database of dancers' performances. Quarterly Progress Scientific Report of the Numediart Research Program, 2 (3), 85-90.

Tardieu, D., Chessini Bose, R., Dubois, J., Dupont, S., Hidot, S., Mazzarino, B., Siebert, X., Varni, G., & Visentin, A. (2009). Video Navigation Tool: Application to browsing a database of dancers performances. Paper presented at 5th International Summer Workshop on Multimodal Interfaces - eNTERFACE'09, Genova, Italy.

Urbain, J., Dubuisson, T., Dupont, S., Frisson, C., Sebbe, R., & D'alessandro, N. (2009). AudioCycle : a similarity-based visualization of musical libraries. Paper presented at International Conference on Multimedia and Expo, New-York, United States - New York.

Dupont, S., Dubuisson, T., Mills, J., Moinet, A., Siebert, X., Tardieu, D., & Urbain, J. (26 June 2009). LaughterCycle. Quarterly Progress Scientific Report of the Numediart Research Program, 2 (2), 23-32.

Dupont, S., Dubuisson, T., Urbain, J., Sebbe, R., D'alessandro, N., & Frisson, C. (2009). Audiocycle : browsing musical loop libraries. Paper presented at International workshop on Content-Based Multimedia Indexing, Chania, Greece.

Siebert, X., Dupont, S., Fortemps, P., & Tardieu, D. (01 March 2009). MediaCycle: browsing and performing with sound and image libraries. Quarterly Progress Scientific Report of the Numediart Research Program, 2 (1).

Urbain, J., Dupont, S., Niewiadomski, R., Dutoit, T., & Pelachaud, C. (27 February 2009). Towards a virtual agent using similarity-based laughter production. Paper presented at Interdisciplinary workshop on laughter and other Interactional Vocalisations in speech, Berlin, Germany.

Dupont, S., D'alessandro, N., Dubuisson, T., Frisson, C., Sebbe, R., & Urbain, J. (26 December 2008). AudioCycle: Browing Musical Loop Libraries. Quarterly Progress Scientific Report of the Numediart Research Program, 1 (4), 119-127.

Couvreur, L., Bettens, F., Drugman, T., Dubuisson, T., Dupont, S., Frisson, C., Jottrand, M., & Mancas, M. (01 June 2008). Project # 2.3 : audio thumbnailing. Quarterly Progress Scientific Report of the Numediart Research Program, 1 (2), 67-85.

DeMori, R., Deroo, O., Dupont, S., Fissore, L., Laface, P., Mertins, A., & Wellekens, C. (01 October 2007). Introduction to the Special Issue on Intrinsic Speech Variation. Speech Communication, 49 (10-11).
Peer Reviewed verified by ORBi

De Mori, R., Deroo, O., Dupont, S., Fissore, L., Laface, P., Mertins, A., Wellekens, C., De Mori, R. (Ed.), Deroo, O. (Ed.), Dupont, S. (Ed.), Fissore, L. (Ed.), Laface, P. (Ed.), Mertins, A. (Ed.), & Wellekens, C. (Ed.). (2007). Intrinsic Speech Variation and Speech Recognition. Elsevier.

Benzeghiba, M., De Mori, R., Deroo, O., Dupont, S., Erbes, T., Jouvet, D., Fissore, L., Laface, P., Mertins, A., Ris, C., Rose, R., Tyagi, V., & Wellekens, C. (01 October 2007). Automatic Speech Recognition and Speech Variability: a Review. Speech Communication, 49 (10-11), 763-786.
Peer Reviewed verified by ORBi

Dupont, S. (2007). FP6 DIVINES Project Final Activity Report Publishable Executive Summary - February 2004-January 2007 (Project Nbr. 002034).

Dupont, S. (2007). FP6 DIVINES Project Periodic Activity Report Publishable Executive Summary - February 2006-January 2007 (Project Nbr. 002034).

Benzeghiba, M., DeMori, R., Deroo, O., Dupont, S., Jouvet, D., Fissore, L., Laface, P., Mertins, A., Ris, C., Rose, R., Tyagi, V., & Wellekens, C. (2006). Impact of variabilities on speech recognition. Paper presented at SPECOM 2006, 11th International Conference Speech and Computer, Saint Petersburg, Russia.

Dupont, S. (2006). Proceedings of Workshop on Speech Recognition and Intrinsic Variation (SRIV2006). Paper presented at International Workshop on Speech Recognition and Intrinsic Variation (SRIV'2006), Toulouse, France.

Benzeghiba, M., De Mori, R., Deroo, O., Dupont, S., Erbes, T., Jouvet, D., Fissore, L., Laface, P., Mertins, A., Ris, C., Rose, R., Tyagi, V., & Wellekens, C. (2006). Automatic speech recognition and intrinsic speech variation. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

Dupont, S., & Ris, C. (2006). Multiple acoustic and variability estimation models for ASR. Paper presented at International Workshop on Speech Recognition and Intrinsic Variation (SRIV'2006), Toulouse, France.

Dupont, S. (2006). FP6 DIVINES Project Periodic Activity Report Publishable Executive Summary - February 2005-January 2006 (Project Nbr. 002034).

Dupont, S., Ris, C., Deroo, O., & Poitoux, S. (2005). Feature Extraction and Acoustic Modeling: an Approach for Improved Generalization across Languages and Accents. Paper presented at IEEE workshop on Automatic Speech Recognition and Understanding (ASRU), San Juan , Unknown/unspecified.

Couvreur, L., Boite, J.-M., Dupont, S., & Ris, C. (2005). Confidence Measure Normalization for Robust Selection of ASR Agents. Paper presented at International Conference on Speech and Computer, Patras, Greece.

Dupont, S. (2005). FP6 DIVINES Project Periodic Activity Report Publishable Executive Summary (Project Nbr. 002034): February 2004-January 2005.

Dupont, S., & Woodruff, P. (2005). Bimodal combination of speech anf handwriting for improved word recognition. Paper presented at EUSIPCO 2005, Antalya, Turkey.

Dupont, S., & Ris, C. (2005). A study of implicit and explicit modeling of coarticulation and pronunciation variation. Paper presented at Interspeech 2005, Lisbonne, Portugal.

Dupont, S., & Ris, C. (2004). Combined use of close-talk and throat microphones for improved speech recognition under non-stationary background noise. Paper presented at Robust 2004 ITRW on Robustness Issues in Conversational, .

Garudadri, H., Sivadas, S., Hermansky, H., Morgan, N., Wooters, C., Adami, A., Benitez, C., Burget, L., Dupont, S., Grezl, F., Jain, P., Kajarekar, S., & Motlicek, P. (01 October 2003). Distributed voice recognition system utilizing multistream network feature processing (PCT Patent WO03094152/US2003204394).

Dupont, S. (2003). FPMs et Multitel battent un record du monde. Polytech-News (Num. 24).

Dupont, S. (29 January 2003). Robust Parameters for Noisy Speech Recognition (PCT patent WO2001BE00072 20010425).

Dupont, S., & Ris, C. (2003). Robust Feature Extraction and Acoustic Modeling at Multitel: Experiments on the Aurora Databases. Paper presented at Eurospeech 2003, Genève, Switzerland.

Garudadri, H., Hermansky, H., Burget, L., Jain, P., Kajarekar, S., Sivadas, S., Dupont, S., Benitez, C., & Morgan, N. (01 January 2003). System and method for computing and transmitting parameters in a distributed voice recognition system (PCT Patent Patent WO02061727/US2003004720).

Adami, A., Burget, L., Dupont, S., Garudadri, H., Grezl, F., Hermansky, H., Jain, P., Kajarekar, S., Morgan, N., & Sivadas, S. (2002). Qualcomm-icsi-ogi features for ASR. Paper presented at International Conference on Speech and Language Computing (ICSLP'2002), Denver, United States - Colorado.

Segura, J., Benitez, M., de la Torre, A., Dupont, S., & Rubio, A. (2002). VTS residual noise compensation. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

Benitez, C., Burget, L., Chen, B., Dupont, S., Garudadri, H., Hermansky, H., Jain, P., Kajarekar, S., & Sivadas, S. (2001). Robust ASR front-end using spectral-based and discriminant features: experiments on the aurora tasks. Paper presented at Eurospeech 2001, Aalborg Congress and Culture Centre, Denmark.

Couvreur, C., Couvreur, L., Dupont, S., & Ris, C. (2001). Fast Adaptation for Robust Speech Recognition in Reverberant Environments. Paper presented at International Workshop on Adaptation Methods for Speech Recognition, Sophia Antipolis, France.

Dupont, S., & Ris, C. (2001). Multiband with Contaminated Training Data. Paper presented at CRAC Workshop (Eurospeech 2001 Satelitte event), Aalborg, Denmark.

Dupont, S., & Juergen, L. (01 September 2000). Audio-Visual Speech Modeling for Continuous Speech Recognition. IEEE Multimedia, 2 (3), 141-151.
Peer Reviewed verified by ORBi

Dupont, S. (2000). Etude et développement d'architectures multi-bandes et multi-modales pour la reconnaissance robuste de la parole. Unpublished doctoral thesis, Université de Mons.
Jury: .

Amdal, I., Blomberg, M., Boite, J.-M., Bourlard, H., Brendborg, M., Christensen, H., Cochard, J.-L., Cernocky, J., Deroo, O., Dupont, S., Elenius, K., Fiser, R., Fontaine, V., Hanzl, V., Hutter, H. P., Hogberg, J., Imperl, B., Johansen, F. T., Kacic, Z., & Wellekens, C. (2000). Final Report of COST Action 249 Continuous Speech Recognition over the Telephone: Chapter 3 Acoustic Modelling.

Dupont, S., & Ris, C. (2000). Assessing Local Noise Level Estimation Methods: Application to Noise Robust ASR. Speech Communication, (34), 141-158.
Peer Reviewed verified by ORBi

Dupont, S., & Cheboub, L. (2000). Fast Speaker Adaptation of Artificial Neural Networks for Automatic Speech Recognition. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

Dupont, S., & Cheboub, L. (2000). Fast Speaker Adaptation of HMM/ANN automatic speech recognition systems. Paper presented at IEEE Signal Processing Symposium, Hilvarenbeek, Netherlands.

Dupont, S., & Ris, C. (1999). Assessing Local Noise Level Estimation Methods. Paper presented at Workshop on Robust Methods For Speech Recognition in Adverse Conditions (Nokia, COST249, IEEE), Tampere, Finland.

Deroo, O., Dupont, S., & Ris, C. (1999). Context Dependent Hybrid HMM/ANN Systems for large vocabulary continuous speech recognition system. Paper presented at Eurospeech'99, Budapest, Hungary.

Dupont, S., & Luettin, J. (1998). Using the Multi-Stream Approach for Continuous Audio-Visual Speech Recognition: Experiments on the M2VTS Database. Paper presented at International Conference on Speech and Language Computing (ICSLP'98), Sydney, Australia.

Dupont, S. (1998). Missing Data Reconstruction for Robust Automatic Speech Recognition in the Framework of Hybrid HMM/ANN Systems. Paper presented at International Conference on Speech and Language Computing (ICSLP'98), Sydney, Australia.

Dupont, S. (1998). Reconstruction de Données Manquantes pour la Reconnaissance Robuste de la Parole dans le Cadre des Systèmes Hybrides HMM/ANN. Paper presented at XXIIèmes Journées d'Etudes sur la Parole, Martigny, Switzerland.

Dupont, S., & Luettin, J. (02 June 1998). Continuous Audio-Visual Speech Recognition. Lecture Notes in Computer Science, 1406.
Peer reviewed

Luettin, J., & Dupont, S. (1998). Continuous Audio-Visual Speech Recognition: Rapport de Recherche Idiap-RR-02-1998.

Bourlard, H., Dupont, S., & Ris, C. (17 December 1997). Multi-Stream Speech Recognition. CCAI: the Journal for the Integrated Study of Artificial Intelligence, Cognitive Science and Applied Epistemology, 15 (3), 215-234.
Peer reviewed

Dupont, S., Boite, J.-M., Ris, C., Deroo, O., Fontaine, V., & Zanoni, L. (1997). Context Independent and Context Dependent Hybrid HMM/ANN Systems for Training Independent Tasks. Paper presented at EUROSPEECH'97, Rhodes, Greece.

Dupont, S., & Bourlard, H. (1997). Using Multiple Time Scales in a Multi-Stream Recognition System. Paper presented at EUROSPEECH'97, Rhodes, Greece.

Dupont, S. (1997). Using the Multi-Stream Approach for Continuous Audio-Visual Speech Recognition: Experiments on the M2VTS Database.

Dupont, S., Boite, J.-M., Bourlard, H., Deroo, O., & Fontaine, V. (1997). Hybrid HMM/ANN Systems for Training Independent Tasks : Experiments on Phonebook and Related Improvements. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

Bourlard, H., & Dupont, S. (1997). Subband-based Speech Recognition. IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings.
Peer reviewed

Dupont, S., Ris, C., & Bourlard, H. (1997). Robust Speech Recognition Based on Multi-Stream Features. Paper presented at ESCA/NATO Workshop on Robust Speech Recognition for Unknown Communication Channels, Pont-à-Mousson, France.

Dupont, S., Bourlard, H., & Ris, C. (1997). Robust Speech Recognition based on Multi-Stream Features: Rapport de Recherche Idiap-RR-01-1997.

Dupont, S., Bourlard, H., & Ris, C. (1997). Using the Multi-Stream Approach for Continuous Audio-Visual Speech Recognition: Rapport de Recherche Idiap-RR-14-1997.

Boite, J.-M., Dupont, S., Ris, C., Bataille, F., Deroo, O., Fontaine, V., & Zanoni, L. (1997). STRUT : Un logiciel complet pour l'entrainement et la reconnaissance de la parole. Paper presented at Premières Journées Scientifiques et Techniques FRANCIL, Avignon, France.

Dupont, S., & Bourlard, H. (1996). Multiband approach for speech recognition. Paper presented at ProRisc 8th Annual WorkShop on Circuits, System and Signal Processing, Mierlo, Netherlands.

Deroo, O., Leich, H., Boite, J.-M., Dupont, S., Ris, C., & Fontaine, V. (1996). Hybrid HMM/ANN systems for Speaker Independent Continuous Speech Recognition In French. In ProRisc 8th Annual WorkShop on Circuits, System and Signal Processing (1996). -.
Peer reviewed

Bourlard, H., & Dupont, S. (1996). A new ASR approach based on independent processing and recombination of partial frequency bands. In International Conference on Speech and Language Processing (ICSLP'96). -.
Peer reviewed

Bourlard, H., Dupont, S., Hermansky, H., & Morgan, N. (1996). Towards sub-band-based speech recognition. In EUSIPCO (1996). -.
Peer reviewed

Bourlard, H., Dupont, S., & Ris, C. (1996). Multi-Stream Speech Recognition: Rapport de Recherche Idiap-RR-07-1996.

Dupont, S. (17 June 1996). Multi-Band Speech Recognition. Paper presented at COST Action 249 Continuous Speech Recognition over the Telephone 6th MCM and WGMs, Stockholm, Sweden.

Dupont, S. (1995). Etude et Développement d'un Convertisseur Pitch-to-MIDI. Unpublished master thesis, Université de Mons.
Jury: .

Contact ORBi