...A Planet of Partners™

  • Increase font size
  • Default font size
  • Decrease font size


Encoding device and decoding device
Tue, 22 Jul 2014 08:00:00 EDT
An encoding device (200) includes an MDCT unit (202) that transforms an input signal in a time domain into a frequency spectrum including a lower frequency spectrum, a BWE encoding unit (204) that generates extension data which specifies a higher frequency spectrum at a higher frequency than the lower frequency spectrum, and an encoded data stream generating unit (205) that encodes to output the lower frequency spectrum obtained by the MDCT unit (202) and the extension data obtained by the BWE encoding unit (204). The BWE encoding unit (204) generates as the extension data (i) a first parameter which specifies a lower subband which is to be copied as the higher frequency spectrum from among a plurality of the lower subbands which form the lower frequency spectrum obtained by the MDCT unit (202) and (ii) a second parameter which specifies a gain of the lower subband after being copied.
Method and apparatus for the provision of information signals based upon speech recognition
Tue, 22 Jul 2014 08:00:00 EDT
A wireless comprises at least one subscriber unit in wireless communication with an infrastructure. Each subscriber unit implements a speech recognition client, and the infrastructure comprises a speech recognition server. A given subscriber unit takes as input an unencoded speech signal that is subsequently parameterized by the speech recognition client. The parameterized speech is then provided to the speech recognition server that, in turn, performs speech recognition analysis on the parameterized speech. Information signals, based in part upon any recognized utterances identified by the speech recognition analysis, are subsequently provided to the subscriber unit. The information signals may be used to control the subscriber unit itself; to control one or more devices coupled to the subscriber unit, or may be operated upon by the subscriber unit or devices coupled thereto.
Temporal video tagging and distribution
Tue, 22 Jul 2014 08:00:00 EDT
A method and apparatus for temporal tagging of videos in response to identification of locations on the video where persons in a contact list of a user are identified in response to facial recognition. The temporal tags can be retained within the video, or contained separately in a metadata file having fields to indicate time locations in the video and identified persons seen at those locations. Use of the temporal tags allows for automatic distribution of content relevant to the recipient, based on their presence in the video, or of persons in the video of which they are interested.
Converting non-natively executable programs to downloadable executable programs
Tue, 22 Jul 2014 08:00:00 EDT
In an embodiment, a data processing method comprises receiving, from a first computer, and storing at a service provider computer, a copy of a non-natively-executable computer program; generating and distributing a download link that is configured, when activated, to cause downloading of a dynamically generated natively executable installer program from the service provider computer; receiving a request based upon the download link, and in response, the service provider computer generating a natively executable installer program that comprises the non-natively-executable computer program file and, optionally a player or other support software and/or one or more bundled external software offers, and providing the installer program to the end user computer; receiving, from the installer program, a request for the non-natively-executable computer program; providing the non-natively-executable computer program from the service provider computer to the end user computer; and the method is performed by one or more computing devices.
Web service support for a multimodal client processing a multimodal application
Tue, 22 Jul 2014 08:00:00 EDT
Web service support for a multimodal client processing a multimodal application, the multimodal client providing an execution environment for the application and operating on a multimodal device supporting multiple modes of user interaction including a voice mode and one or more non-voice modes, the application stored on an application server, includes: receiving, by the server, an application request from the client that specifies the application and device characteristics; determining, by a multimodal adapter of the server, modality requirements for the application; selecting, by the adapter, a modality web service in dependence upon the modality requirements and the characteristics for the device; determining, by the adapter, whether the device supports VoIP in dependence upon the characteristics; providing, by the server, the application to the client; and providing, by the adapter to the client in dependence upon whether the device supports VoIP, access to the modality web service for processing the application.
Fuzzy inference methods, and apparatuses, systems and apparatus using such inference apparatus
Tue, 22 Jul 2014 08:00:00 EDT
A health monitoring system for complex networked apparatus includes a number of neuro-fuzzy inference apparatuses feeding inference results into a data fusion hierarchy. At each level in the hierarchy, fuzzy inference is applied to generate a desired output signal by processing selected input signals in accordance with a knowledge base defining fuzzy membership functions and fuzzy inference rules defined in advance. The knowledge base includes alternative definitions of membership functions and/or inference rules. The apparatus selects which definition to use according to environmental or other conditions, and predetermined selection criteria.
Apparatus and methods for processing a signal using a fixed-point operation
Tue, 22 Jul 2014 08:00:00 EDT
Apparatus and methods for processing compression encoded signals are provided. In some embodiments, a signal processing method is provided that includes receiving a subband of a compression encoded signal at a subband processor, generating envelope information regarding the subband of the compression encoded signal to provide changes in the dynamic range of the compression encoded signal for fixed-point digital signal processing, processing the compression encoded signal with a fixed-point companding digital signal processor using the envelope information, and producing a processed compression encoded signal at the output of the subband processor.
Apparatus and method for calculating bandwidth extension data using a spectral tilt controlled framing
Tue, 22 Jul 2014 08:00:00 EDT
An apparatus for calculating bandwidth extension data of an audio signal in a bandwidth extension system, in which a first spectral band is encoded with a first number of bits and a second spectral band different from the first spectral band is encoded with a second number of bits, the second number of bits being smaller than the first number of bits, has a controllable bandwidth extension parameter calculator for calculating bandwidth extension parameters for the second frequency band in a frame-wise manner for a sequence of frames of the audio signal. Each frame has a controllable start time instant. The apparatus additionally includes a spectral tilt detector for detecting a spectral tilt in a time portion of the audio signal and for signaling the start time instant for the individual frames of the audio signal depending on spectral tilt.
Decoding method and apparatus for an audio signal through high frequency compensation
Tue, 22 Jul 2014 08:00:00 EDT
A decoding apparatus decodes a first encoded data that is encoded from a low-frequency component of an audio signal, and a second encoded data that is used when creating a high-frequency component of an audio signal from a low-frequency component and encoded in accordance with a certain bandwidth, into the audio signal. In the decoding apparatus, a high-frequency component detecting unit divides the high-frequency component into bands with a certain interval range correspondingly to the certain bandwidth, and detects magnitude of the high-frequency components corresponding to each of the bands. A high-frequency component compensating unit compensates the high-frequency components based on the magnitude of the high-frequency components corresponding to each of the bands detected by the high-frequency component detecting unit. A decoding unit that decodes the low-frequency component decoded from the first encoded data, and the high-frequency components compensated by the high-frequency component compensating unit, into the audio signal.
Language converter and transmitting system
Tue, 22 Jul 2014 08:00:00 EDT
A language converter and transmitting system includes a processor adapted for converting the first language into a plurality of secondary languages. An audio input is operationally coupled to the processor. An audio output of a media source may be operationally coupled to the audio input. A wireless transmitter is operationally coupled to the processor for transmitting each of the secondary languages. Each of the secondary languages is transmitted on one of a plurality of frequencies. Each of a plurality of receiving devices includes a receiver for receiving signals from the wireless transmitter. A tuner for selecting one of the plurality of frequencies is operationally coupled to the receiver. A sound emitter for emitting an audible sound is operationally coupled to the tuner. The tuner may be used to select one of said plurality of frequencies such that one of the plurality of secondary languages is emitted through the sound emitter.
Method for quick scroll search using speech recognition
Tue, 22 Jul 2014 08:00:00 EDT
A method for a computing device to search for data entails receiving first user input that initiates a quick scrolling action and activates a speech recognition subsystem, receiving second user input by recognizing voice input using the speech recognition subsystem to determine a search query, and searching for data that corresponds to the search query. The quick scrolling action and activation of the speech recognition subsystem may be triggered, for example, by a swiping gesture on an optical jog pad, on a touch screen, or on a touch-sensitive mouse, or by a contactless three-dimensional gesture.
Systems and methods for editing telecom web applications through a voice interface
Tue, 22 Jul 2014 08:00:00 EDT
Systems and associated methods for editing telecom web applications through a voice interface are described. Systems and methods provide for editing telecom web applications over a connection, as for example accessed via a standard phone, using speech and/or DTMF inputs. The voice based editing includes exposing an editing interface to a user for a telecom web application that is editable, dynamically generating a voice-based interface for a given user for accomplishing editing tasks, and modifying the telecom web application to reflect the editing commands entered by the user.
Controlling user interfaces with contextual voice commands
Tue, 22 Jul 2014 08:00:00 EDT
One or more voice-enabled user interfaces include a user interface, and a voice extension module associated with the user interface. The voice extension module is configured to voice-enable the user interface and includes a speech recognition engine, a preprocessor, and an input handler. The preprocessor registers with the speech recognition engine one or more voice commands for signaling for execution of one or more semantic operations that may be performed using a first user interface. The input handler receives a first voice command and communicates with the preprocessor to execute a semantic operation that is indicated by the first voice command. The first voice command is one of the voice commands registered with the speech recognition engine by the preprocessor.
Apparatus and method for determining an emotion state of a speaker
Tue, 22 Jul 2014 08:00:00 EDT
A method and apparatus for analyzing speech are provided. A method and apparatus for determining an emotion state of a speaker are provided, including providing an acoustic space having one or more dimensions, where each dimension corresponds to at least one baseline acoustic characteristic; receiving an utterance of speech by the speaker; measuring one or more acoustic characteristics of the utterance; comparing each of the measured acoustic characteristics to a corresponding baseline acoustic characteristic; and determining an emotion state of the speaker based on the comparison. An embodiment involves determining the emotion state of the speaker within one day of receiving the subject utterance of speech. An embodiment involves determining the emotion state of the speaker, where the emotion state of the speaker includes at least one magnitude along a corresponding at least one of the one or more dimensions within the acoustic space.
Satisfying specified intent(s) based on multimodal request(s)
Tue, 22 Jul 2014 08:00:00 EDT
Techniques are described herein that are capable of satisfying specified intent(s) based on multimodal request(s). A multimodal request is a request that includes at least one request of a first type and at least one request of a second type that is different from the first type. Example types of request include but are not limited to a speech request, a text command, a tactile command, and a visual command. A determination is made that one or more entities in visual content are selected in accordance with an explicit scoping command from a user. In response, speech understanding functionality is automatically activated, and audio signals are automatically monitored for speech requests from the user to be processed using the speech understanding functionality.
Speech synthesis from acoustic units with default values of concatenation cost
Tue, 22 Jul 2014 08:00:00 EDT
A speech synthesis system can select recorded speech fragments, or acoustic units, from a very large database of acoustic units to produce artificial speech. When a pair of acoustic units in the database does not have an associated concatenation cost, the system assigns a default concatenation cost. The system then synthesizes speech, identifies the acoustic unit sequential pairs generated and their respective concatenation costs, and stores those concatenation costs likely to occur.
Multi-purpose contextual control
Tue, 22 Jul 2014 08:00:00 EDT
A method and a system for activating functions including a first function and a second function, wherein the system is embedded in an apparatus, are disclosed. The system includes a control configured to be activated by a plurality of activation styles, wherein the control generates a signal indicative of a particular activation style from multiple activation styles; and controller configured to activate either the first function or the second function based on the particular activation style, wherein the first function is configured to be executed based only on the activation style, and wherein the second function is further configured to be executed based on a speech input.
Language model creation device, language model creation method, and computer-readable storage medium
Tue, 22 Jul 2014 08:00:00 EDT
The present invention uses a language model creation device 200 that creates a new language model using a standard language model created from standard language text. The language model creation device 200 includes a transformation rule storage section 201 that stores transformation rules used for transforming dialect-containing word strings into standard language word strings, and a dialect language model creation section 203 that creates dialect-containing n-grams by applying the transformation rules to word n-grams in the standard language model and, furthermore, creates the new language model (dialect language model) by adding the created dialect-containing n-grams to the word n-grams.
System and method for babble noise detection
Tue, 22 Jul 2014 08:00:00 EDT
A method, device, system, and computer program product calculate a gradient index as a sum of magnitudes of gradients of speech signals from a received frame at each change of direction; and provide an indication that the frame contains babble noise if the gradient index, energy information, and background noise level exceed pre-determined thresholds or a voice activity detector algorithm and sound level indicate babble noise.
Audio encoding method, audio decoding method, audio encoding device, audio decoding device, program, and audio encoding/decoding system
Tue, 22 Jul 2014 08:00:00 EDT
An audio encoding device (1A) corrects initial gain information calculated for an arbitrary frame, based on gain information of a stored past frame, thereby calculating gain information to be used in the frame. The audio encoding device (1A) encodes the calculated gain information as a difference from the gain information of the past frame. An audio decoding device (3A) receives the differential gain, and calculates the gain of the arbitrary frame based on the gain used in the past frame, thereby generating a decoded audio signal.
Natural language processing for analyzing internet content and finding solutions to needs expressed in text
Tue, 22 Jul 2014 08:00:00 EDT
A natural language processing methodology to automatically transform push advertising into pull advertising. Text found in forum, blog, and social media postings throughout the Internet is grammatically analyzed to identify potential customers who have expressed a clear problem. Only parsing rules with the least likely elements present are evaluated. In response, personalized replies are produced that contain pertinent and useful information about a potential product or service. Those replies appear to come from other Internet users, thus converting expressed needs of user/prospects into discussions with sales technicians.
Incorporation of variables into textual content
Tue, 22 Jul 2014 08:00:00 EDT
Embodiments of the invention provide techniques for incorporating variable values into textual content. In one embodiment, an abstract phrase including a text phrase and a variable at a particular position in the text phrase is received. The abstract phrase may include multiple variables. A text value for the variable is received. The text phrase of the abstract phrase is combined with the text value according to the particular position of the variable. An integration rule is applied at a boundary of the text phrase of the abstract phrase and the text value, where the integration rule is based on a language rule. The integration rule modifies a portion of the text phrase of the abstract phrase or a portion of the text value to produce an integrated phrase.
Method and system for analyzing text
Tue, 22 Jul 2014 08:00:00 EDT
An apparatus for providing a control input signal for an industrial process or technical system having one or more controllable elements includes elements for generating a semantic space for a text corpus, and elements for generating a norm from one or more reference words or texts, the or each reference word or text being associated with a defined respective value on a scale, and the norm being calculated as a reference point or set of reference points in the semantic space for the or each reference word or text with its associated respective scale value. Elements for reading at least one target word included in the text corpus, elements for predicting a value of a variable associated with the target word based on the semantic space and the norm, and elements for providing the predicted value in a control input signal to the industrial process or technical system. A method for predicting a value of a variable associated with a target word is also disclosed together with an associated system and computer readable medium.
Generating snippets based on content features
Tue, 22 Jul 2014 08:00:00 EDT
Systems, methods, and computer storage media having computer-executable instructions embodied thereon that facilitate generation of snippets. In embodiments, text features within a keyword-sentence window are identified. The text features are utilized to determine break features that indicate favorability of breaking at a particular location of the keyword-sentence window. The break features are used to recognize features of partial snippets such that a snippet score to indicate the strength of the partial snippet can be calculated. Snippet scores associated with partial snippets are compared to select an optimal snippet, that is, the snippet having the highest snippet score.
Rules-based language detection
Tue, 22 Jul 2014 08:00:00 EDT
Techniques for determining one or more preferred languages for a user are provided. The preferred languages may be determined based upon a set of language indicators. The language indicators are analyzed using, for example, rules-based techniques, clustering, language classifiers, and the like, or combinations thereof. Language indicators can include or be derived from information about the user's behavior, location, preferences, social connections, or other data related to the user.
Machine translation using global lexical selection and sentence reconstruction
Tue, 22 Jul 2014 08:00:00 EDT
Disclosed are systems, methods, and computer-readable media for performing translations from a source language to a target language. The method comprises receiving a source phrase, generating a target bag of words based on a global lexical selection of words that loosely couples the source words/phrases and target words/phrases, and reconstructing a target phrase or sentence by considering all permutations of words with a conditional probability greater than a threshold.
Unified cross platform input method framework
Tue, 22 Jul 2014 08:00:00 EDT
The disclosed subject matter provides a system and method for responding to messages related to multilingual input from different operating systems. A central hub component executing on a processor manages communication between connected components, such as computer applications and different processes. An input message may be analyzed by the processor to determine a type of input device. A combination of rule sets may be applied to the input message sets. The first and second rule sets may be selected based on the analysis. The first and second rule sets may be applied to the input message. A probable output message may be generated from the application of each of the rule sets in the combination of rule sets, and forwarded for delivery to an output device.
Multiple language voice recognition
Tue, 22 Jul 2014 08:00:00 EDT
Computer implemented speech processing generates one or more pronunciations of an input word in a first language by a non-native speaker of the first language who is a native speaker of a second language. The input word is converted into one or more pronunciations. Each pronunciation includes one or more phonemes selected from a set of phonemes associated with the second language. Each pronunciation is associated with the input word in an entry in a computer database. Each pronunciation in the database is associated with information identifying a pronunciation language and/or a phoneme language.
Method of controlling dialing modes in a vehicle
Tue, 22 Jul 2014 08:00:00 EDT
A dialing mode of a telematics unit in a vehicle is controlled by monitoring for dialing digits from a vehicle occupant, determining whether the type of dialing digits are continuous dialing digits or discrete dialing digits, establishing a continuous mode for receiving continuous dialing digits or a discrete mode for receiving discrete dialing digits based on the determination, and if the type of dialing digits changes, switching the established mode.
Method and system for interference suppression using blind source separation
Tue, 22 Jul 2014 08:00:00 EDT
A method of interference suppression is provided that includes receiving a first audio signal from a first audio capture device and a second audio signal from a second audio capture device wherein the first audio signal includes a first combination of desired audio content and interference and the second audio signal includes a second combination of the desired audio content and the interference, performing blind source separation using the first audio signal and the second audio signal to generate an output interference signal and an output audio signal including the desired audio content with the interference suppressed, estimating interference remaining in the output audio signal using the output interference signal, and subtracting the estimated interference from the output audio signal to generate a final output audio signal with the interference further suppressed.

Language Selection

linkedin        twitter

Company Search

Advanced Search   


logo inttranet

logo inttrastats

logo Inttranews

logo Inttrasearch

logo linguists of the year