OAK

AI MUSIC : Deep Network-Based Music Composition and Singing Voice Synthesis

Metadata Downloads
Abstract
In this paper, we propose a novel deep network model for music composition and singing voice synthesis being studied in the artificial intelligence music field. In the first chapter, we propose a Multi- Objective Evolutionary Algorithm (MOEA) method that applies a deep network generation model to solve high-dimensional and complex problems such as musical composition. The proposed method demonstrates the scalability of the model and the validity of the proposed technique through two experiments on the knapsack problem and music generation. In the second chapter, we propose a non-autoregressive transformer based model suitable for Singing voice synthetics (SVS), which requires a long sequence. This paper proposes an improved SVS model compared to the existing Transformer model using LHS attention. The proposed method performs extensive experiments on sequences of various lengths, showing that it is memory efficient, has a faster inference speed, and generates good quality speech compared to conventional SVS models.
Author(s)
Eunbin Lee
Issued Date
2023
Type
Thesis
URI
https://scholar.gist.ac.kr/handle/local/18849
Alternative Author(s)
이은빈
Department
대학원 AI대학원
Advisor
Ahn, Chang Wook
Degree
Master
Appears in Collections:
Department of AI Convergence > 3. Theses(Master)
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.