1. bookAHEAD OF PRINT
Dettagli della rivista
License
Formato
Rivista
eISSN
2444-8656
Prima pubblicazione
01 Jan 2016
Frequenza di pubblicazione
2 volte all'anno
Lingue
Inglese
Accesso libero

Application research of bel canto performance based on artificial intelligence technology

Pubblicato online: 30 Nov 2022
Volume & Edizione: AHEAD OF PRINT
Pagine: -
Ricevuto: 09 May 2022
Accettato: 08 Aug 2022
Dettagli della rivista
License
Formato
Rivista
eISSN
2444-8656
Prima pubblicazione
01 Jan 2016
Frequenza di pubblicazione
2 volte all'anno
Lingue
Inglese
Introduction

The World Conference on Artificial Intelligence (AI) in Music (Summit on Music Intelligence [SOMI]: Music IntelLigence) was held in Beijing from 22 October to 24 October 2021. The conference was jointly organised by the Central Conservatory of Music and the China Association for Artificial Intelligence. International Nobel physicist Li Zhengdao wished this conference thus: ‘science and technology integration, the combination of science and technology, I wish the AI Music Conference a complete success!’ The development of human society and the development of science and technology will have a great impact on music. In the industrial age, the smelting technology invented the copper pipe and realised the symphony. The rise of electronic technology and computer applications has brought about the prosperity of intelligent music. Nowadays, with the rise of new information industry applications, such as fifth-generation wireless technology (5G), virtual reality (VR), augmented reality (AR), AI, cloud computing and big data, AI music is also rising [1].

In 2019, China Ping, an AI research institute, debuted the symphonic variations on ‘Me and My Motherland’ by AI at Shenzhen Concert Hall [2]. The work adopts the self-changing mode of ‘I and my Country,’ with other classical music elements added to the variations. The whole piece is divided into five sections, with Me and My Motherland as the theme, showing the grand picture of the development of new China. The music director of Shenzhen Philharmonic Orchestra, Lin Daye, highly praised the quality of its music and proposed that AI will be handed down in 5–10 years. In fact, AI music is not a new research direction, it has been developed over several years; especially in recent years, AI music has aroused great attention [3].

AI is generally defined as ‘machine-displayed intelligence,’ while AI music refers to music produced through operations based on systems such as computer neural networks (NNs). At present, AI technology is still in its infancy and far from real maturity. But some of its intelligence is already impressive, such as Google's AlphaGo, which beat the world's most famous Go player. The basic principle of AI and chess is basically the same; AI uses the genetic algorithm, NN, Markov chain, hybrid algorithm and other technologies (through the rules of the computer) and establishes a massive database and then applies deep learning and its rules, structure and other analyses [4].

From the traditional point of view, the advantage of machines is that they can help people to do repetitive mechanical work, rather than creative work. However, with the continuous development of AI technology, its application in music creation, production, analysis, education and other aspects is increasingly becoming extensive [5]. With the development of AI technology, the efficiency of music creation has not only been greatly improved, but it has also solved the previous work that was difficult to be effectively completed by humans. In the present reasonable method, rules and the corresponding algorithm model are used to create music. AI technology can solve the difficult problems in the traditional music industry, including ‘creation standards and evaluation of AI music’, ‘deep learning’, ‘the potential of AI virtual singers’, and the application of AI in various fields. Machine learning can be said to be an important branch of AI, and it has a profound impact on other aspects of AI [6]. Using AI technology to realise automatic singing bel canto is not new research; related research has been done for many years, but there have been technical limitations.

Kenneth Philles, director of the Canadian Institute of Telecommunications and Media Arts and professor of the Department of Music Artificial Intelligence and Information Technology at the Central Conservatory of Music, suggested that music is needed to solve the delay problem caused by cross-regional co-operation using the Internet [7]. The AI automatically handles the delay curve and timing settings at each node. Academician Guan Xiaohong discussed the mathematical characteristics of three kinds of musical melodies under the title of Quantitative Rules in Musical Melodies and constructed a mathematical model in order to find out the extent to which they change and then obtain their variance [8].

General Secretary Xi Jinping said, ‘Fine arts, art, science and technology complement and promote each other [9]. At present, music has become the forefront of the development of technology and art. With the continuous development of AI technology, from a single technology to the integrated technology, from a single intelligence to swarm intelligence, from data-driven to scenario-driven tasks, AI and cross-border integration of music art and the modern media, each note and every melody of music are likely to achieve the best effect in terms of hearing, sight and touch, apart from creating a sound–scene blend of artistic conception, letting the feast of hearing sublimate into a variety of sensory enjoyment, so that the immersion of music continues to convey the emotion contained in music. This paper focusses on the application of AI technology in bel canto, analyses the problems encountered at present and discusses how to use AI technology to better serve music.

Analysis of relevant technical means
Types of NNs

Traditional AI thinks in the top–down manner; its essential characteristic is to simulate the human brain neuron. This method has two characteristics: first, calculation of the weight of the neighbouring neuron through the corresponding output function and its further processing; second, the weight value is used to determine the information transmission relationship between neurons, and the weight is constantly optimised and adjusted in this process [10]. At the same time, the NN also depends on a large amount of data during processing. Therefore, the NN shows non-linearity, distributed configuration, parallel structure, adaptability, self-organisation and other characteristics. Take people's music creation as an example: usually through the perception of music (appreciation), music imitation and writing, they finally achieve independent creation. The process of composition also includes the study of composition techniques, harmony theory and other aspects. Learners constantly improve their creative ideas in practice and under the guidance of teachers. This learning method can be simulated by using the structure of NN to a large extent, thus laying a foundation for the application of this technology [11].

The operation process of NN requires the architecture of input, output, weight and multi-level perception (Figure 1). NNs can be regarded as a ‘black box.’ As long as there are enough training sets, the desired value Y can be obtained after X is given at the input end [12]. Specifically, the workflow of an NN is to determine the inputs and outputs, then to find a set or sets of algorithms that can be input and output, and then to find a set of known solutions to train the model; and then, over and over again, repeat the process; thus, conclusions can be drawn and the process can be constantly modified.

Fig. 1

Schematic diagram of a neural network

Music is an art of time, and much of its message is based on time. On this basis, a variety of methods can be adopted to complete the processing of the timeline information, namely recurrent NN (RNN) [13, 14], which is an NN that can learn on the basis of existing and generated data. RNN is a time-shifted NN, which measures depth based on time. Circular networks generally have the same input layer and output layer. This is because the circular network expects the next entry to be the input of the next step, so as to generate the order; so RNN is a good implementation method [15].

Long- and short-term memory units

Long short-term memory (LSTM) is a special RNN structure that belongs to feedback NN. LSTM is an NN formed to solve the problem of gradient disappearance or burst in an RNN cyclic network. RNN technology can process time dimension information at the same time, but if the data is stored for too long, it becomes very difficult to store the data, which becomes a big problem. Therefore, LSTM is a special hidden unit whose natural property is long-time storage. The main change to LSTM is the addition of three gates, namely the input gate, the output gate and the forget gate. In practice, LSTM has been proved to be superior to RNN [16]. It has been widely used in machine translation, conversation generation and compilation. LSTM can describe the logical thinking and cognitive activities of more complex people, so it is the most significant research field at present.

Variational autoencoder (VAE)

The autoencoder (AE) is shown in Figure 2. It uses an NN to convert an image or sound into a set of numbers. The goal is to make the image and sound searchable and then use that number to reconstruct the image and sound.

Fig. 2

Autoencoder (AE)

The VAE is an improvement of the AE. Its structure is similar to that of the AE, but it also includes an encoder and a decoder [17]. The hidden vector generated by the variational autocoding method must satisfy the standard normal distribution. The VAE mode was first used in The Google Music VAE mode, where the user can input two videos, which are then inserted by the model to produce a continuous transit.

Transformer

Transformer is the first machine translation model introduced by Google Brain in 2017. The advent of this model has shaken the RNN in the field of deep learning. In many experiments, the transformer performed better than the RNN. Google also made a name for itself in 2018 by publishing an article on Music Transformer to solve the music generation problem [18]. The transformer model mechanism is shown in Figure 3.

Fig. 3

Transformer model mechanism

Application of AI technology in bel canto performance

This overall network structure is shown in Figure 4, including four parts: musical score encoder; duration predictor; length regulator and decoder.

Fig. 4

Overall network structure diagram. BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

Musical score encoder

It encodes the sequence of phonemes, note durations and pitches into a sequence of dense spatial vectors. Musical score = musical score; Phoneme ID = Phoneme to ID – encoded by phoneme embedding; Note duration = duration – encoded by duration embedding; Note pitch = pitch – encoded by pitch embedding; These three input vectors are then ‘stacked’ and thrown to multi-layer transformer encoder layers. As one can see, the position encoding information is also incorporated, followed by each FFT block [19, 20, 21]. The structure of the score encoder is shown in Figure 5.

Fig. 5

Input: ‘score’ encoder

The sheet music usually contains the lyrics, pitch and length of the note so that people can get hold of the sheet music and sing it. As shown in Figure 5, the first step is to convert the word form into phonetic phonemes (such as Chinese characters to pinyin), where each syllable (such as fang->f + ang, a vowel + consonant form, so that both f and ang exist as phonemes) is subdivided into several phonemes. Each pitch is converted to a pitch ID according to musical instrument digital interface (MIDI) standards [22, 23, 24, 25].

In addition, the length is converted to frame count (the number of phoneme frames) via music Tempo. Note pitch and Note duration are repeated to fit the length of each phoneme. Thus, it can be considered that the shape of the input score is SRN×3 S \in {R^{N \times 3}}

Type: N is the number of phonemes.

For example: WO ai ni Zhong Guo - wo ai n I zhong guo.

After three embedding layers respectively, the corresponding vectors of phoneme, duration and pitch will be added [26].

An FFT block consists of a multi-head self-attention layer, a two-layer 1D convolution layer (similar to the feed-forward linear layer) and the rectified linear unit (ReLU) activation function [27, 28]. There is no layer norm or residual connection in the paper as shown in Figure 6:

Fig. 6

Example of music notation

I - > Wu O O three phonemes; - converts to phoneme ID; And - > h e e; - > pitch, a quantified pitch number; You - > n I I three phonemes. - > beats, first converted to a specific time length (according to MIDI tempo) and then converted to 10 ms as a cut. For example: 1 s = 1000 ms - > 1000 ms/10 ms = 100.

Duration predictor

According to the vector sequence obtained by the encoder, obtain phoneme duration, i.e. the duration of each phoneme (one phoneme, how long should be sung in the song) [29, 30]. What it yields is the duration of each phoneme (for example, a sequence of integers). This predictor includes several 1D convolutional networks, similar to those in FastSpeech. In addition to the duration of phonemes, the duration of syllables also plays an important role in learning the rhythmic patterns of ‘singing speech synthesis.’ So, in an attempt to learn better rhythm, we propose to add syllable-based level (pronunciation of individual characters: Wu o o = I, now it is one: wo, the control). A syllable can correspond to one or more notes. Therefore, the duration loss for a syllable (de-syllable level), loss of syllable duration (L-SD) is designed to reinforce the similarity between the ‘syllable duration for reference answer’ and ‘the sum of the predicted duration of all phonemes in that syllable as predicted by the model [31].’ These are as follows: Ldur=wpd*Lpd+wsd*Lsd {L_{dur}} = {w_{pd}}*{L_{pd}} + {w_{sd}}*{L_{sd}}

This is the loss of duration prediction. L-pd refers to the loss of phoneme duration and L-sd refers to the loss of syllable duration. W-pd and W-sd are their respective losses.

Length regulator

Its main work is to ‘extend’ the length of the ‘vector sequence obtained by the encoder’ according to the predicted tone length [32]. For example, if a phoneme lasts for 5 s (‘Ah... Ahh’), then the regulator expands it several times according to how many time slices correspond to 5 s (say, 20 ms time slices) (here 5000/20=250 times). The results for the encoding vector sequence after length extension are as shown in Figure 7.

Fig. 7

Length predictor and duration regulator

Decoder

The decoder, shown in Figure 8, is responsible for generating acoustic features from an expanded sequence of coding vectors [32, 33].

Fig. 8

Decoder decoder.

BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

In this paper, the World Vocoder is used, which requires the decoder to be able to predict mel-generalised coefficient (MGC) and band aperiodicity (BAP), not mel spectrum. The definition of loss here is as follows: Lspec=wm*Lm+wb*Lb {L_{spec}} = {w_m}*{L_m} + {w_b}*{L_b}

The left side represents the loss in spectral parameters. On the right, L-M stands for loss of MGC; L-b indicates loss of BAP [34].

W-m and W-b are their respective weights. Compared with speech, singing has a more complex and sensitive F0 profile. For example, it has a wider range from 100 Hz to 3500 Hz. The dynamics of F0 movements, such as trills and overtones, help convey emotions more expressively [35].

Some prior research has shown that even a slight deviation from standard pitch can seriously damage the listening experience. On the other hand, it is difficult to cover all pitch ranges in sufficient cases using training data. This means that F0 prediction can be problematic if the pitch of the input note is not displayed or is rarely present in the training data. Data enhancement solves this problem by implementing pitch conversion on training data. But this is not economical and can lead to longer training sessions. Instead, we propose a residual connection between the input and output pitches (here, we use logF0, the logarithmic scale of F0) so that the decoder only needs to predict human bias based on the standard note pitch, which is more robust for rare or invisible data. Later experiments confirmed this [36, 37].

The prediction of F0, as usual, is accompanied by the voiced/unvoiced (V/UV) decision. Since the V/UV decision is binary, a logistic regression is used here. Thus, the loss function of the final decoder is as follows: Ldec=Lspec+wf*Lf+wu*Lu {L_{dec}} = {L_{spec}} + {w_f}*{L_f} + {w_u}*{L_u}

Here, L-f stands for loss of logF0; L-u stands for loss of V/UV decision and W-f and W-u are their respective weights.

Fig. 9

Residual connection of pitch. BAP, band aperiodicity; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

Experimental verification
Preparation

The data included 1000 pieces of Chinese Mandarin bel canto music, with one girl singing, in a professional recording studio at 48 Hz and under 16-bit quantisation. Then, it underwent cuts for no more than 10 s. We ended up with 10,360 segments with 70 h of data. From this pool, 8756 fragments are used as training sets, and the remaining 1032 are used as validation sets and test sets. Acoustic features were extracted by World: 15 ms frame shift, 60-D MGC, 5-D BAP, 1-D logF0 and 1-D V/UV flag. The duration label (integer) of the phoneme is obtained by a hidden Markov model (HMM) forced alignment (probably something like HM-based speech synthesis system [H-triple-S or HTS]). In XiaoiceSing (Figure 1), six layers of FFT blocks are shared. The phoneme word list size is 72. Encoder output is a 384-D vector (one vector per phoneme); decoder output is 67-D acoustic signature. The loss of MGC, BAP and logF0 is calculated separately using L1 regularisation. The V/UV decision uses binary cross entropy loss, Adam optimiser NVIDIA P100 GPU, a batch size= 3230K iteration step convergence.

Mean opinion score (MOS) comparison

In this experiment, baseline and XiaoIceSing systems were compared to evaluate overall performance. We first conducted MOS test on the two systems. For each system, we prepared 30 audio samples, each completed in about 10 s. We asked 10 listeners to rate the MOS score of each sample based on articulation accuracy, tone quality and naturalness. The MOS score is the average of all samples to obtain the final score (see Table 1). The MOS score for pronunciation accuracy was 4.52, close to the recorded score of 4.89 and 1.28 higher than the baseline. The sound quality and naturalness also improved by 1.54 and 1.48, respectively, from the baseline. The standard deviation on the MOS was smaller than the baseline, indicating better performance stability across all test cases.

Mean opinion score test results

Evaluation itemRecordingXiaoiceSingBaseline

Pronununciation accuracy4.81±0.564.42±0.683.84±1.20
Sound quality4.64±0.503.80±0.742.56±0.85
Naturalness4.88±0.553.71±0.772.43±0.83
Error, correlation and distortion

Some objective indicators were further calculated (Table 2). The quality of the synthesised song was measured by root mean square error of duration (Dur RMSE), duration correlation (Dur CORR), F0 RMSE, F0 CORR, mel cepstral distortion (MCD), band aperiodic distortion (BAPD) and voiced/unvoiced error rate (V/UV error). It shows that XiaoiceSing achieved lower RMSE and higher CORR in both duration and F0 than the baseline. Meanwhile, MCD, BAPD and V/UV errors were lower than baseline for XiaoiceSing. We can see that small ice stars have a greater ability to generate accurate phoneme duration, F0 and spectral features.

Objective evaluation results of different systems

Evaluation itemXiaoiceSingBaseline

DUR RMSE21.0523.39
DUR CORR0.810.79
F0 RMSE(Hz)11.4512.74
F0 CORR0.890.81
MCD (dB)4.426.89
BAPD (dB)24.1227.39
V/UV error (%)2.865.27

BAPD, band band aperiodic distortion; Dur RMSE, root mean square error of duration; Dur CORR, duration correlation; F0 RMSE, xxx; F0 CORR, yyy; MCD, mel cepstral distortion; V/UV error, voiced/unvoiced error

Experimental results show that compared with the baseline system, it has a great advantage in terms of sound quality, pronunciation accuracy and naturalness. In particular, F0 perception performs very well due to the residual linkage between note pitch and predicted F0, and the improvement in duration prediction is significant due to the addition of a syllable duration constraint between the expected syllable duration and the predicted phoneme duration.

Conclusion

In this experiment, we compare XiaoiceSing's duration (duration) model with a separate LSTM-based duration (duration) model to evaluate the advantages of the proposed duration modelling. Similarly, another A/B test was conducted to assess the duration of rhythm preference, with 88.7% supporting XiaoiceSing and only 16.5% supporting the baseline. A single LSTM-based duration model is very unstable and predicts a very long duration for the last phoneme. Moreover, its cumulative error is much greater than that of XiaoiceSing. The results of the two A/B tests in the last two sections confirm the significant advantages of our proposed system in terms of duration and F0 prediction. The conclusion is consistent with the above objective evaluation.

‘The significance of scientific exploration lies in the pursuit of truth, the value of artistic creation lies in beauty, and the common pursuit of both lies in the development and enhancement of human creativity and appreciation. The rapid development of emerging technologies represented by AI has brought science and art closer together. To further promote communication between scientists and artists, thinking collusions and achievements, technology elements throughout the whole process produce works of art creation with the creative transformation of science and technology for art resources, innovative development, producing more perceptible, public art achievements of cognitive and cultural products so that the public can get more beautiful experience and knowledge harvest.’ In terms of AI technology itself, the biggest challenge is how to understand creative artistic thinking. At the moment, it is difficult for computers to programme understanding, and AI composing systems cannot replace human composers, but it does affect the production and composition process of music. The application of AI technology reduces the threshold of music education, greatly improves the efficiency of music education and solves the problem of music copyright to a certain extent. In the future, with the increase of relevant research at home and abroad and the attention of all sectors of society, more applications will emerge, thus spawning more suitable theories.

Fig. 1

Schematic diagram of a neural network
Schematic diagram of a neural network

Fig. 2

Autoencoder (AE)
Autoencoder (AE)

Fig. 3

Transformer model mechanism
Transformer model mechanism

Fig. 4

Overall network structure diagram. BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced
Overall network structure diagram. BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

Fig. 5

Input: ‘score’ encoder
Input: ‘score’ encoder

Fig. 6

Example of music notation
Example of music notation

Fig. 7

Length predictor and duration regulator
Length predictor and duration regulator

Fig. 8

Decoder decoder.BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced
Decoder decoder.BAP, band aperiodicity; F0, xxx; FFT, yyy; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

Fig. 9

Residual connection of pitch. BAP, band aperiodicity; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced
Residual connection of pitch. BAP, band aperiodicity; MGC, mel-generalised coefficient; V/UV, voiced/unvoiced

Mean opinion score test results

Evaluation item Recording XiaoiceSing Baseline

Pronununciation accuracy 4.81±0.56 4.42±0.68 3.84±1.20
Sound quality 4.64±0.50 3.80±0.74 2.56±0.85
Naturalness 4.88±0.55 3.71±0.77 2.43±0.83

Objective evaluation results of different systems

Evaluation item XiaoiceSing Baseline

DUR RMSE 21.05 23.39
DUR CORR 0.81 0.79
F0 RMSE(Hz) 11.45 12.74
F0 CORR 0.89 0.81
MCD (dB) 4.42 6.89
BAPD (dB) 24.12 27.39
V/UV error (%) 2.86 5.27

Intelligence C. CDC #289: Can there be Artificial Intelligence?[J]. Totse Com. Intelligence C CDC #289: Can there be Artificial Intelligence?[J] Totse Com Search in Google Scholar

Chaudhri, K Vinay, Gunning. Intelligent Learning Technologies Part 2: Applications of Artificial Intelligence to Contemporary and Emerging Educational Challenges.[J]. AI Magazine, 2013. ChaudhriK Vinay Gunning Intelligent Learning Technologies Part 2: Applications of Artificial Intelligence to Contemporary and Emerging Educational Challenges.[J] AI Magazine 2013 10.1609/aimag.v34i4.2518 Search in Google Scholar

Zhang L K. The Usage of the Bel Canto Singing Skills in the Performance of “Huaer”[J]. Journal of Qinghai Nationalities University (Social Sciences), 2011. ZhangL K The Usage of the Bel Canto Singing Skills in the Performance of “Huaer”[J] Journal of Qinghai Nationalities University (Social Sciences) 2011 Search in Google Scholar

Aissani N, Beldjilali B, Trentesaux D. Dynamic scheduling of maintenance tasks in the petroleum industry: A reinforcement approach[J]. Engineering Applications of Artificial Intelligence, 2009, 22(7):1089–1103. AissaniN BeldjilaliB TrentesauxD Dynamic scheduling of maintenance tasks in the petroleum industry: A reinforcement approach[J] Engineering Applications of Artificial Intelligence 2009 22 7 1089 1103 10.1016/j.engappai.2009.01.014 Search in Google Scholar

Kirschenbaum M G. The Poetics of Artificial Intelligence[J]. KirschenbaumM G The Poetics of Artificial Intelligence[J] Search in Google Scholar

Wei E. Intonation Characteristics of Singing Based on Artificial Intelligence Technology and Its Application in Song-on-Demand Scoring System. 2021. WeiE Intonation Characteristics of Singing Based on Artificial Intelligence Technology and Its Application in Song-on-Demand Scoring System 2021 10.1155/2021/5510401 Search in Google Scholar

Bel-Enguix G, M.D. Jiménez-López. Genetic code and verbal language: Syntactic and semantic analogies[J]. Frontiers in Artificial Intelligence & Applications, 2011, 228:85–103. Bel-EnguixG Jiménez-LópezM.D. Genetic code and verbal language: Syntactic and semantic analogies[J] Frontiers in Artificial Intelligence & Applications 2011 228 85 103 Search in Google Scholar

Beliakov G, James S, Smith L. Single-preference consensus measures based on models of ecological evenness.[C] Modeling Decisions for Artificial Intelligence. 2014. BeliakovG JamesS SmithL Single-preference consensus measures based on models of ecological evenness.[C] Modeling Decisions for Artificial Intelligence 2014 10.1007/978-3-319-12054-6_5 Search in Google Scholar

Vinod B. Artificial Intelligence and Emerging Technologies in Travel[J]. Management for Professionals, 2021. VinodB Artificial Intelligence and Emerging Technologies in Travel[J] Management for Professionals 2021 10.1007/978-3-030-70424-7_11 Search in Google Scholar

Intelligence C. CDC #289: Can there be Artificial Intelligence?[J]. Totse Com. Intelligence C CDC #289: Can there be Artificial Intelligence?[J] Totse Com Search in Google Scholar

Chaudhri, K Vinay, Gunning. Intelligent Learning Technologies Part 2: Applications of Artificial Intelligence to Contemporary and Emerging Educational Challenges.[J]. AI Magazine, 2013. ChaudhriK Vinay Gunning Intelligent Learning Technologies Part 2: Applications of Artificial Intelligence to Contemporary and Emerging Educational Challenges.[J] AI Magazine 2013 10.1609/aimag.v34i4.2518 Search in Google Scholar

Zhang L K. The Usage of the Bel Canto Singing Skills in the Performance of “Huaer”[J]. Journal of Qinghai Nationalities University (Social Sciences), 2011. ZhangL K The Usage of the Bel Canto Singing Skills in the Performance of “Huaer”[J] Journal of Qinghai Nationalities University (Social Sciences) 2011 Search in Google Scholar

Aissani N, Beldjilali B, Trentesaux D. Dynamic scheduling of maintenance tasks in the petroleum industry: A reinforcement approach[J]. Engineering Applications of Artificial Intelligence, 2009, 22(7):1089–1103. AissaniN BeldjilaliB TrentesauxD Dynamic scheduling of maintenance tasks in the petroleum industry: A reinforcement approach[J] Engineering Applications of Artificial Intelligence 2009 22 7 1089 1103 10.1016/j.engappai.2009.01.014 Search in Google Scholar

Kirschenbaum M G. The Poetics of Artificial Intelligence[J]. KirschenbaumM G The Poetics of Artificial Intelligence[J] Search in Google Scholar

Wei E. Intonation Characteristics of Singing Based on Artificial Intelligence Technology and Its Application in Song-on-Demand Scoring System. 2021. WeiE Intonation Characteristics of Singing Based on Artificial Intelligence Technology and Its Application in Song-on-Demand Scoring System 2021 10.1155/2021/5510401 Search in Google Scholar

Bel-Enguix G, M.D. Jiménez-López. Genetic code and verbal language: Syntactic and semantic analogies[J]. Frontiers in Artificial Intelligence & Applications, 2011, 228:85–103. Bel-EnguixG Jiménez-LópezM.D. Genetic code and verbal language: Syntactic and semantic analogies[J] Frontiers in Artificial Intelligence & Applications 2011 228 85 103 Search in Google Scholar

Beliakov G, James S, Smith L. Single-preference consensus measures based on models of ecological evenness.[C] Modeling Decisions for Artificial Intelligence. 2014. BeliakovG JamesS SmithL Single-preference consensus measures based on models of ecological evenness.[C] Modeling Decisions for Artificial Intelligence 2014 10.1007/978-3-319-12054-6_5 Search in Google Scholar

Vinod B. Artificial Intelligence and Emerging Technologies in Travel[J]. Management for Professionals, 2021. VinodB Artificial Intelligence and Emerging Technologies in Travel[J] Management for Professionals 2021 10.1007/978-3-030-70424-7_11 Search in Google Scholar

Russell, Stuart J. Artificial Intelligence: A Modern Approach[J]. Posts and Telecommunications Press, 2002. RussellStuart J Artificial Intelligence: A Modern Approach[J] Posts and Telecommunications Press 2002 Search in Google Scholar

Russell S J, Norvig P. Artificial Intelligence: A Modern Approach[M]. 2003. RussellS J NorvigP Artificial Intelligence: A Modern Approach[M] 2003 Search in Google Scholar

Ligeza A. Artificial Intelligence: A Modern Approach[J]. Applied Mechanics & Materials, 2009, 263(2):2829–2833. LigezaA Artificial Intelligence: A Modern Approach[J] Applied Mechanics & Materials 2009 263 2 2829 2833 10.1016/0925-2312(95)90020-9 Search in Google Scholar

Nilsson N J. Artificial intelligence: A modern approach[J]. Applied Mechanics & Materials, 2003, 263(5):2829–2833. NilssonN J Artificial intelligence: A modern approach[J] Applied Mechanics & Materials 2003 263 5 2829 2833 10.1016/0004-3702(96)00007-0 Search in Google Scholar

Kohavi R, John G. Artificial Intelligence Wrappers for feature subset selection[J]. Cleve, 1997. KohaviR JohnG Artificial Intelligence Wrappers for feature subset selection[J] Cleve 1997 10.1016/S0004-3702(97)00043-X Search in Google Scholar

Russell S J, Norvig P N. Artificial Intelligence: A Modern Approach. Prentice Hall[J]. applied mechanics & materials, 2009. RussellS J NorvigP N Artificial Intelligence: A Modern Approach. Prentice Hall[J] applied mechanics & materials 2009 Search in Google Scholar

Nilsson N J. Principle of artificial intelligence[M]. IEEE, 1982. NilssonN J Principle of artificial intelligence[M] IEEE 1982 10.1007/978-3-662-09438-9 Search in Google Scholar

F Rossi, Beek P V, Walsh T, et al. Handbook of constraint programming: foundations of artificial intelligence[J]. Elsevier Science Inc. 2006. RossiF BeekP V WalshT Handbook of constraint programming: foundations of artificial intelligence[J] Elsevier Science Inc. 2006 Search in Google Scholar

Zhu X, Goldberg A B. Introduction to Semi-Supervised Learning (Synthesis Lectures on Artificial Intelligence and Machine Learning)[J]. Synthesis Lectures on Artificial Intelligence & Machine Learning, 2009, 3(1):1–130. ZhuX GoldbergA B Introduction to Semi-Supervised Learning (Synthesis Lectures on Artificial Intelligence and Machine Learning)[J] Synthesis Lectures on Artificial Intelligence & Machine Learning 2009 3 1 1 130 10.1007/978-3-031-01548-9 Search in Google Scholar

Norvig P, Russell S. Artificial Intelligence: A Modern Approach (All Inclusive), 3/E[J]. Applied Mechanics & Materials, 1995, 263(5):2829–2833. NorvigP RussellS Artificial Intelligence: A Modern Approach (All Inclusive), 3/E[J] Applied Mechanics & Materials 1995 263 5 2829 2833 Search in Google Scholar

Russell S, Norvig P. Artificial Intelligence: A Modern Approach, 3rd Edition[J]. Applied Mechanics & Materials, 1995, 263(5):2829–2833. RussellS NorvigP Artificial Intelligence: A Modern Approach, 3rd Edition[J] Applied Mechanics & Materials 1995 263 5 2829 2833 Search in Google Scholar

Negnevitsky M. Artificial Intelligence: A Guide to Intelligent Systems[J]. Information & Computing Sciences, 2005, 48(48):284–300. NegnevitskyM Artificial Intelligence: A Guide to Intelligent Systems[J] Information & Computing Sciences 2005 48 48 284 300 Search in Google Scholar

Wang F. Proceedings of the National Conference on Artificial Intelligence[J]. Springer Verlag, 2011, 36(5):823–834. WangF Proceedings of the National Conference on Artificial Intelligence[J] Springer Verlag 2011 36 5 823 834 Search in Google Scholar

D Lehmann, M Magidor. Erratum: “What does a conditional knowledge base entail?” Artificial Intelligence 55 (1992), no. 1, 1–60; MR1167275 (93i:03035).[J]. artificial intelligence, 1994(2):411. LehmannD MagidorM Erratum: “What does a conditional knowledge base entail?” Artificial Intelligence 55 (1992), no. 1, 1–60; MR1167275 (93i:03035).[J] artificial intelligence 1994 2 411 10.1016/0004-3702(94)90073-6 Search in Google Scholar

Angel G. Logical Foundations Of Artificial Intelligence[M]. Broad Research - international virtual organization, 2010. AngelG Logical Foundations Of Artificial Intelligence[M] Broad Research - international virtual organization 2010 Search in Google Scholar

Fox M, College U. Handbook of Logic in Artificial Intelligence and Logic Programming: Volume 1: Logic Foundations[M]. Oxford University Press, Inc. 1994. FoxM CollegeU Handbook of Logic in Artificial Intelligence and Logic Programming: Volume 1: Logic Foundations[M] Oxford University Press, Inc. 1994 Search in Google Scholar

Haugeland J. Artificial Intelligence: The Very Idea[J]. Philosophical Review, 1985, 7:3–11. HaugelandJ Artificial Intelligence: The Very Idea[J] Philosophical Review 1985 7 3 11 10.7551/mitpress/1170.001.0001 Search in Google Scholar

Kojima K, Yanagida M, Nakayama I. Variability of Vibrato - A Comparative Study between Japanese Traditional Singing and Bel Canto -[J]. speech prosody, 2004. KojimaK YanagidaM NakayamaI Variability of Vibrato - A Comparative Study between Japanese Traditional Singing and Bel Canto -[J] speech prosody 2004 Search in Google Scholar

Chen Y Q. The rationality of Traditional National Vocal Singing under the Bel Canto[J]. Journal of Chuxiong Normal University, 2009. ChenY Q The rationality of Traditional National Vocal Singing under the Bel Canto[J] Journal of Chuxiong Normal University 2009 Search in Google Scholar

Articoli consigliati da Trend MD

Pianifica la tua conferenza remota con Sciendo