Interesting People mailing list archives
Re JFK video: hear Kennedy's 'lost' Dallas speech in his own voice
From: "Dave Farber" <farber () gmail com>
Date: Sun, 25 Mar 2018 11:12:34 -0400
Begin forwarded message:
From: "Wendy M. Grossman" <wendyg () pelicancrossing net> Date: March 25, 2018 at 9:20:16 AM EDT To: dave () farber net Subject: Re: [IP] JFK video: hear Kennedy's 'lost' Dallas speech in his own voice For IP if you wish: This story I wrote for Scientific American goes into more detail about how the same company made a synthetic voice for Roger Ebert: http://www.scientificamerican.com/article.cfm?id=new-speech-synthesis-could-make-robert-ebert-sound-more-like-himself wgOn 2018-03-25 13:58, Dave Farber wrote: Begin forwarded message:*From:* Dewayne Hendricks <dewayne () warpspeed com <mailto:dewayne () warpspeed com>> *Date:* March 25, 2018 at 8:12:03 AM EDT *To:* Multiple recipients of Dewayne-Net <dewayne-net () warpspeed com <mailto:dewayne-net () warpspeed com>> *Subject:* *[Dewayne-Net] JFK video: hear Kennedy's 'lost' Dallas speech in his own voice* *Reply-To:* dewayne-net () warpspeed com <mailto:dewayne-net () warpspeed com> JFK video: hear Kennedy’s ‘lost’ Dallas speech in his own voice On November 22, 1963 John F Kennedy was silenced. Fifty-five years later we have made it possible for him to be heard again By Aaron Rogan Mar 15 2018 <https://www.thetimes.co.uk/edition/news/jfk-video-hear-kennedys-lost-dallas-speech-in-his-own-voice-xtkvhm255> It took eight weeks to bring to life the 2,590 words John F Kennedy never got to speak. Sound engineers pulled 116,777 sound units from 831 of his speeches and radio addresses. These units were then split in half and analysed for pitch and energy. The half units, known as phones, were each about 0.04 seconds long and had to be tested next to each other to ensure that they did not clash. The W sound in weapons, for example, is not the same as the W sound in words. “There are only 40/45 phones in English so once you’ve got that set you can generate any word in the English language. The problem is that it would not sound natural because one sound merges into the sound next to it so they’re not really independent. You really need the sounds in the context of every other sound and that makes the database big,” Chris Pidcock, co-founder and chief voice engineer at CereProc in Edinburgh, said. How the speech was recreated Mr Pidcock’s company specialises in text-to-voice technology. It is used by brands to bring characters and products to life and also to allow people who are losing the power of speech from motor neurone disease or other conditions to maintain their own voice, something that is also part of their character. He has previously recreated Roger Ebert’s voice, at a time when the late American film critic had lost his speech because of cancer. That project involved trawling hundreds of hours of film commentaries from DVDs. Recreating Kennedy’s voice and unique cadence for the JFK Unsilenced project, to produce the speech he was due to give on the day he was assassinated in Dallas in 1963, proved a more difficult challenge. “Because of the old analogue recording devices used, it appeared as if it was a different person speaking each time. Trying to harmonise the environment and manipulate the audio so that it ran together was quite difficult,” Mr Pidcock said. He estimated that of the final 20-minute speech, fewer than half the audio units were beside each other when they were pulled from recordings. “Getting to that point is pretty challenging based on the variable audio quality, as well as the speech itself having different qualities and noise levels. One of the things we needed to do is get a very accurate transcription of the audio so that things like ‘umms’ and ‘ehs’ could be labelled and we could make sure the phonetic pieces we got were correct. If you label them incorrectly you might pick the wrong piece and the whole sentence will sound wrong.” The project involved a painstaking two-month process to pick the best-quality recordings of Kennedy’s voice and cross reference them with the written speech so that any crosstalk or static was not selected. Once a database of the cleanest sound units was built, a new computer system was employed to recognise and recreate Kennedy’s oratorical style. This required feeding data from his speeches into a computer until it learnt the patterns in his delivery. Then, when the speech was put in, the system could tweak it and make it more natural. [snip] Dewayne-Net RSS Feed: http://dewaynenet.wordpress.com/feed/ Twitter: https://twitter.com/wa8dzpArchives <https://www.listbox.com/member/archive/247/=now> | Modify <https://www.listbox.com/member/?&> Your Subscription | Unsubscribe Now <https://www.listbox.com/unsubscribe/?&&post_id=20180325085847:3FC8451C-302C-11E8-9483-C8555F67D5ED> [Powered by Listbox] <http://www.listbox.com>-- www.pelicancrossing.net <-- all about me Twitter: @wendyg
------------------------------------------- Archives: https://www.listbox.com/member/archive/247/=now Modify Your Subscription: https://www.listbox.com/member/?member_id=18849915&id_secret=18849915-aa268125 Unsubscribe Now: https://www.listbox.com/unsubscribe/?member_id=18849915&id_secret=18849915-32545cb4&post_id=20180325111241:F49937F0-303E-11E8-82C8-D8C866047D21 Powered by Listbox: http://www.listbox.com
Current thread:
- Re JFK video: hear Kennedy's 'lost' Dallas speech in his own voice Dave Farber (Mar 25)