Konferans bildirisi Açık Erişim

Database construction for speech to lip-readable animation conversion

acs, Gyorgy Ta; Tihanyi, Atilla; Bardi, Tamas; Feldhoffer, Gergo; Srancsi, Balint


DataCite XML

<?xml version='1.0' encoding='utf-8'?>
<resource xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns="http://datacite.org/schema/kernel-4" xsi:schemaLocation="http://datacite.org/schema/kernel-4 http://schema.datacite.org/meta/kernel-4.1/metadata.xsd">
  <identifier identifierType="URL">https://aperta.ulakbim.gov.tr/record/40797</identifier>
  <creators>
    <creator>
      <creatorName>acs, Gyorgy Ta</creatorName>
      <givenName>Gyorgy Ta</givenName>
      <familyName>acs</familyName>
      <affiliation>Peter Pazmany Catholic Univ, Fac Informat Technol, Prater U 50-a, H-1083 Budapest, Hungary</affiliation>
    </creator>
    <creator>
      <creatorName>Tihanyi, Atilla</creatorName>
      <givenName>Atilla</givenName>
      <familyName>Tihanyi</familyName>
      <affiliation>Peter Pazmany Catholic Univ, Fac Informat Technol, Prater U 50-a, H-1083 Budapest, Hungary</affiliation>
    </creator>
    <creator>
      <creatorName>Bardi, Tamas</creatorName>
      <givenName>Tamas</givenName>
      <familyName>Bardi</familyName>
      <affiliation>Peter Pazmany Catholic Univ, Fac Informat Technol, Prater U 50-a, H-1083 Budapest, Hungary</affiliation>
    </creator>
    <creator>
      <creatorName>Feldhoffer, Gergo</creatorName>
      <givenName>Gergo</givenName>
      <familyName>Feldhoffer</familyName>
      <affiliation>Peter Pazmany Catholic Univ, Fac Informat Technol, Prater U 50-a, H-1083 Budapest, Hungary</affiliation>
    </creator>
    <creator>
      <creatorName>Srancsi, Balint</creatorName>
      <givenName>Balint</givenName>
      <familyName>Srancsi</familyName>
      <affiliation>Peter Pazmany Catholic Univ, Fac Informat Technol, Prater U 50-a, H-1083 Budapest, Hungary</affiliation>
    </creator>
  </creators>
  <titles>
    <title>Database Construction For Speech To Lip-Readable Animation Conversion</title>
  </titles>
  <publisher>Aperta</publisher>
  <publicationYear>2006</publicationYear>
  <dates>
    <date dateType="Issued">2006-01-01</date>
  </dates>
  <resourceType resourceTypeGeneral="Text">Conference paper</resourceType>
  <alternateIdentifiers>
    <alternateIdentifier alternateIdentifierType="url">https://aperta.ulakbim.gov.tr/record/40797</alternateIdentifier>
  </alternateIdentifiers>
  <relatedIdentifiers>
    <relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.81043/aperta.40796</relatedIdentifier>
    <relatedIdentifier relatedIdentifierType="DOI" relationType="IsIdenticalTo">10.81043/aperta.40797</relatedIdentifier>
  </relatedIdentifiers>
  <rightsList>
    <rights rightsURI="http://www.opendefinition.org/licenses/cc-by">Creative Commons Attribution</rights>
    <rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights>
  </rightsList>
  <descriptions>
    <description descriptionType="Abstract">The training database was one of the critical element in our speech to facial animation conversion system. This system was developed as a communication aid for deaf people. The specific database was constructed from audio and visual records of professional lip-speakers. The standardized MPE -4 system was used to animate the talking head model. The trained neural net is able to calculate with acceptable error the principal component weights of feature points from the speech frames. The feature point coordinates are, calculated from PC weights. The whole system can be implemented in mobile phones. Deaf persons were able to recognize about 50 of words from the speech driven animation in the final test.</description>
  </descriptions>
</resource>
42
9
görüntülenme
indirilme
Görüntülenme 42
İndirme 9
Veri hacmi 1.5 kB
Tekil görüntülenme 40
Tekil indirme 9

Alıntı yap