Shilpa et al., 2026 - Google Patents
AI Interview Assistant with Voice AnalysisShilpa et al., 2026
- Document ID
- 6217244358403213271
- Author
- Shilpa K
- Sathwika O
- Sekhar G
- Anand J
- Bhaskar N
- Maheswari K
- Publication year
- Publication venue
- 2026 International Conference on Computing, Electronics & Communications Engineering (ICCECE)
External Links
Snippet
AI Interview Assistant with Voice Analysis is a smart interview helper of the people that is based on natural language processing, speech emotion recognition, and artificial intelligence to assist in review by the data. It involves analyzing audio data of several …
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
- G10L17/26—Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/66—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/20—Handling natural language data
- G06F17/27—Automatic analysis, e.g. parsing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/065—Adaptation
- G10L15/07—Adaptation to the speaker
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6217—Design or setup of recognition systems and techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/003—Changing voice quality, e.g. pitch or formants
- G10L21/007—Changing voice quality, e.g. pitch or formants characterised by the process used
- G10L21/013—Adapting to target pitch
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Hema et al. | Emotional speech recognition using cnn and deep learning techniques | |
| Hassan et al. | Enhanced dysarthria detection in cerebral palsy and ALS patients using WaveNet and CNN-BiLSTM models: A comparative study with model interpretability | |
| CN110956142A (en) | Intelligent interactive training system | |
| CN119478525A (en) | A method and system for analyzing ideological and political classroom interaction based on multimodal fusion | |
| Zhao et al. | A multimodal teacher speech emotion recognition method in the smart classroom | |
| Dwivedi et al. | Analysing the impact of lstm and mfcc on speech emotion recognition accuracy | |
| Rao et al. | Deep learning structure for emotion prediction using MFCC from native languages | |
| Harish et al. | Human emotion recognition by audio signals using MLP classifier | |
| Nagarajan et al. | Neutrosophic speech recognition Algorithm for speech under stress by Machine learning | |
| Dehbozorgi et al. | Affective computing: A topic-based SER approach on collaborative discussions in academic setting | |
| Shilpa et al. | AI Interview Assistant with Voice Analysis | |
| Kalra | LSTM based feature learning and CNN based classification for speech emotion recognition | |
| Thalor et al. | Voice based answer evaluation system for physically disabled students using natural language processing and machine learning | |
| Wang et al. | Automatic English-Speaking Fluency Scoring System Based on Automatic Speech Assessment | |
| Jadhav et al. | AI-based multimodal emotion and behavior analysis of interviewee | |
| Alkhamali et al. | Combining Transformer, CNN, and LSTM Architectures: A Novel Ensemble Learning Technique That Leverages Multi-acoustic Features for Speech Emotion Recognition in Distance Education Classrooms | |
| Roken et al. | Arabic multimodal emotion recognition using deep learning | |
| Huang et al. | Dynamic analysis of classroom engagement sentiment based on multilevel feature extraction and Transformer CNN-LSTM integrated model: personalized behavior prediction for non-English learners in a cross-modal adversarial learning framework | |
| CN118070777B (en) | A multi-dimensional eloquence improvement and collaborative creation method, system, device and medium | |
| Pathak et al. | Emotion-Aware Text to Speech: Bridging Sentiment Analysis and Voice Synthesis | |
| Jayanth et al. | MASCCA: A Multi-modal AI System for Comprehensive Confidence Assessment | |
| Anurag et al. | Voice based answer evaluation system for physically disabled students using natural language processing and machine learning | |
| Sajitha et al. | Speech emotion recognition for adaptive learning experiences using LSTM & MFCC | |
| Ali et al. | Impact of accent on urdu speech emotion recognition: Deep learning based an experimental study | |
| Qiu et al. | Machine Learning in Human Emotion Detection from the Speech |