concatenative synthesis

{{Short description|Method of synthesizing sounds}}

Concatenative synthesis is a technique for synthesising sounds by concatenating short samples of recorded sound (called units). The duration of the units is not strictly defined and may vary according to the implementation, roughly in the range of 10 milliseconds up to 1 second. It is used in speech synthesis and music sound synthesis to generate user-specified sequences of sound from a database (often called a corpus) built from recordings of other sequences.

In contrast to granular synthesis, concatenative synthesis is driven by an analysis of the source sound, in order to identify the units that

best match the specified criterion.

In speech

{{main|Speech synthesis#Concatenative synthesis}}

{{Empty section|date=April 2015}}

In music

Concatenative synthesis for music started to develop in the 2000s in particular through the work of Schwarz{{Citation

|last=Schwarz

|first=Diemo

|title=Data-Driven Concatenative Sound Synthesis

|url=http://recherche.ircam.fr/equipes/analyse-synthese/schwarz/thesis/

|accessdate=2010-01-15

|date=2004-01-23

}}

and Pachet{{Citation

|author1=Zils, A.

|author2=Pachet, F.

|title=Musical Mosaicing

|url=http://www.csl.sony.fr/downloads/papers/2001/zils-dafx2001.pdf

|accessdate=2011-04-27

|periodical=Proceedings of the COST G-6 Conference on Digital Audio Effects (DaFx-01), University of Limerick

|pages=39–44

|year=2001

|archive-url=https://web.archive.org/web/20110927012331/http://www.csl.sony.fr/downloads/papers/2001/zils-dafx2001.pdf

|archive-date=2011-09-27

|url-status=dead

}} (so-called musaicing).

The basic techniques are similar to those for speech, although with differences due to the differing nature of speech and music: for example, the segmentation is not into phonetic units but often into subunits of musical notes or events.{{Citation

| last =Schwarz

| first =D.

| year =2005

| title =Current research in Concatenative Sound Synthesis

| periodical =Proceedings of the International Computer Music Conference (ICMC)

| url =http://recherche.ircam.fr/equipes/analyse-synthese/schwarz/publications/icmc2005/Schwarz_ICMC2005_Current-Research.pdf

| doi =

}}{{Citation

| author =Maestre, E. and Ramírez, R. and Kersten, S. and Serra, X.

| year =2009

| title =Expressive Concatenative Synthesis by Reusing Samples from Real Performance Recordings

| periodical =Computer Music Journal

| volume =33

| issue =4

| pages =23–42

| doi =10.1162/comj.2009.33.4.23

| citeseerx =10.1.1.188.8860

| s2cid =1078610

}}

Zero Point, the first full-length album by Rob Clouth (Mesh 2020), features self-made concatenative synthesis software called the 'Reconstructor' which "chops sampled sounds into tiny pieces and rearranges them to replicate a target sound. This allowed Clouth to use and manipulate his own beatboxing, a technique used on 'Into' and 'The Vacuum State'."{{Cite web |title=Zero Point, by Rob Clouth |url=https://robclouth.bandcamp.com/album/zero-point |access-date=2022-07-23 |website=Rob Clouth |language=en}} Clouth's concatenative synthesis algorithm was adapted from 'Let It Bee — Towards NMF-Inspired Audio Mosaicing' by Jonathan Driedger, Thomas Prätzlich, and Meinard Müller.{{Citation |title=Sónar+D CCCB 2020 Talk: "Journey to the Center of the Musical Brain" | date=29 October 2020 |url=https://www.youtube.com/watch?v=O7IpRDuvZJQ |language=en |access-date=2022-07-23}}{{Cite web |title=AudioLabs - Let it Bee - Towards NMF-inspired Audio Mosaicing |url=https://www.audiolabs-erlangen.de/resources/MIR/2015-ISMIR-LetItBee |access-date=2022-07-23 |website=www.audiolabs-erlangen.de}}

Clouth's work on Zero Point was cited as an inspiration for recent innovations in concatenative synthesis as outlined in "The Concatenator: A Bayesian Approach to Real Time Concatenative Musaicing" by Chris Tralie and Ben Cantil (ISMIR 2024), which improved on the speed, accuracy, and playability of prior realtime concatenative synthesis methods.{{Cite web |url=https://www.ctralie.com/TheConcatenator/supplementary/ |access-date=2025-01-06 |website=www.ctralie.com}}{{Cite web |last=LLC |first=Conference Catalysts |date=2024-09-16 |title=Accepted Papers |url=https://ismir2024.ismir.net/accepted-papers |access-date=2025-01-06 |website=ISMIR 2024 |language=en}} The new algorithm serves as the engine behind the Concatenator plugin by DataMind Audio, which is currently still in beta.{{Cite web |title=DataMind Audio {{!}} Unparalleled Real-time Neural Synthesis |url=https://datamindaudio.ai/ |access-date=2025-01-06 |language=en-US}}

See also

References

{{Speech synthesis}}

Category:Speech synthesis

Category:Sound synthesis types

{{Technology-stub}}