Shilpa et al., 2026 - Google Patents

AI Interview Assistant with Voice Analysis

Shilpa et al., 2026

Document ID
6217244358403213271
Author
Shilpa K
Sathwika O
Sekhar G
Anand J
Bhaskar N
Maheswari K
Publication year
Publication venue
2026 International Conference on Computing, Electronics & Communications Engineering (ICCECE)

External Links

Snippet

AI Interview Assistant with Voice Analysis is a smart interview helper of the people that is based on natural language processing, speech emotion recognition, and artificial intelligence to assist in review by the data. It involves analyzing audio data of several …
Continue reading at ieeexplore.ieee.org (other versions)

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • G10L17/26Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/66Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computer systems utilising knowledge based models
    • G06N5/02Knowledge representation
    • G06N5/022Knowledge engineering, knowledge acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/20Handling natural language data
    • G06F17/27Automatic analysis, e.g. parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/065Adaptation
    • G10L15/07Adaptation to the speaker
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06KRECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K9/00Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
    • G06K9/62Methods or arrangements for recognition using electronic means
    • G06K9/6217Design or setup of recognition systems and techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/003Changing voice quality, e.g. pitch or formants
    • G10L21/007Changing voice quality, e.g. pitch or formants characterised by the process used
    • G10L21/013Adapting to target pitch

Similar Documents

Publication Publication Date Title
Hema et al. Emotional speech recognition using cnn and deep learning techniques
Hassan et al. Enhanced dysarthria detection in cerebral palsy and ALS patients using WaveNet and CNN-BiLSTM models: A comparative study with model interpretability
CN110956142A (en) Intelligent interactive training system
CN119478525A (en) A method and system for analyzing ideological and political classroom interaction based on multimodal fusion
Zhao et al. A multimodal teacher speech emotion recognition method in the smart classroom
Dwivedi et al. Analysing the impact of lstm and mfcc on speech emotion recognition accuracy
Rao et al. Deep learning structure for emotion prediction using MFCC from native languages
Harish et al. Human emotion recognition by audio signals using MLP classifier
Nagarajan et al. Neutrosophic speech recognition Algorithm for speech under stress by Machine learning
Dehbozorgi et al. Affective computing: A topic-based SER approach on collaborative discussions in academic setting
Shilpa et al. AI Interview Assistant with Voice Analysis
Kalra LSTM based feature learning and CNN based classification for speech emotion recognition
Thalor et al. Voice based answer evaluation system for physically disabled students using natural language processing and machine learning
Wang et al. Automatic English-Speaking Fluency Scoring System Based on Automatic Speech Assessment
Jadhav et al. AI-based multimodal emotion and behavior analysis of interviewee
Alkhamali et al. Combining Transformer, CNN, and LSTM Architectures: A Novel Ensemble Learning Technique That Leverages Multi-acoustic Features for Speech Emotion Recognition in Distance Education Classrooms
Roken et al. Arabic multimodal emotion recognition using deep learning
Huang et al. Dynamic analysis of classroom engagement sentiment based on multilevel feature extraction and Transformer CNN-LSTM integrated model: personalized behavior prediction for non-English learners in a cross-modal adversarial learning framework
CN118070777B (en) A multi-dimensional eloquence improvement and collaborative creation method, system, device and medium
Pathak et al. Emotion-Aware Text to Speech: Bridging Sentiment Analysis and Voice Synthesis
Jayanth et al. MASCCA: A Multi-modal AI System for Comprehensive Confidence Assessment
Anurag et al. Voice based answer evaluation system for physically disabled students using natural language processing and machine learning
Sajitha et al. Speech emotion recognition for adaptive learning experiences using LSTM & MFCC
Ali et al. Impact of accent on urdu speech emotion recognition: Deep learning based an experimental study
Qiu et al. Machine Learning in Human Emotion Detection from the Speech