...A Planet of Partners™

  • Increase font size
  • Default font size
  • Decrease font size

Patents

 
Error concealment method and apparatus for audio signal and decoding method and apparatus for audio signal using the same
Tue, 21 Jun 2016 08:00:00 EDT
An error concealment method and apparatus for an audio signal and a decoding method and apparatus for an audio signal using the error concealment method and apparatus. The error concealment method includes selecting one of an error concealment in a frequency domain and an error concealment in a time domain as an error concealment scheme for a current frame based on a predetermined criteria when an error occurs in the current frame, selecting one of a repetition scheme and an interpolation scheme in the frequency domain as the error concealment scheme for the current frame based on a predetermined criteria when the error concealment in the frequency domain is selected, and concealing the error of the current frame using the selected scheme.
Time warp contour calculator, audio signal encoder, encoded audio signal representation, methods and computer program
Tue, 29 Mar 2016 08:00:00 EDT
A time warp contour calculator for use in an audio signal decoder receives an encoded warp ratio information, derives a sequence of warp ratio values from the encoded warp ratio information, and obtains warp contour node values starting from a time warp contour start value. Ratios between the time warp contour node values and the time warp contour starting value are determined by the warp ratio values. The time warp contour calculator computes a time warp contour node value of a given time warp contour node, on the basis of a product-formation having a ratio between the time warp contour node values of the intermediate time warp contour node and the time warp contour starting value and a ratio between the time warp contour node values of the given time warp contour node and of the intermediate time warp contour node as factors.
Soft-sending chat messages
Tue, 26 Jan 2016 08:00:00 EST
Techniques are disclosed for supplying users in an online environment with a safe and effective chat facility. The chat facility is “safe” in the sense that the ability of users to compose inappropriate messages is greatly restricted, while “effective” in the sense that users are still allowed a broad range of expressivity in composing and exchanging chat messages.
Communicating physical layer wireless parameters over an application programming interface
Tue, 22 Sep 2015 08:00:00 EDT
Methods, systems, and devices are described for communicating physical layer wireless parameters over an application programming interface. A wireless modem of a wireless device may measure at least one physical layer wireless parameter. The wireless modem may report the at least one physical layer wireless parameter to an application running on the wireless device over an application programming interface between the wireless modem and the application. A behavior of the application may be adapted to control wireless communications between the application and a network based on the at least one physical layer wireless parameter reported by the wireless modem. Other aspects, embodiments, and features are also claimed and described.
Method, system, and computer readable medium for creating clusters of text in an electronic document
Tue, 21 Jul 2015 08:00:00 EDT
Disclosed herein are systems and methods for navigating electronic texts. According to an aspect, a method may include determining text subgroups within an electronic text. The method may also include selecting a text seed within one of the text subgroups. Further, the method may include determining a similarity relationship between the text seed and one or more adjacent text subgroups that do not include the selected text seed. The method may also include associating the text seed with the one or more adjacent text subgroups based on the similarity relationship to create a text cluster.
Handheld electronic device and method for performing spell checking during text entry and for providing a spell-check learning feature
Tue, 16 Jun 2015 08:00:00 EDT
A handheld electronic device includes a reduced QWERTY keyboard and is enabled with a disambiguation routine that is operable to disambiguate text input. In addition to identifying and outputting representations of language objects that are stored in the memory and that correspond with a text input, the device is able to perform a spell check routine during input of a text entry and to learn and automatically correct mistakes typically made by the particular user.
Integrated psychoacoustic bass enhancement (PBE) for improved audio
Tue, 09 Jun 2015 08:00:00 EDT
Psychoacoustic Bass Enhancement (PBE) is integrated with one or more other audio processing techniques, such as active noise cancellation (ANC), and/or receive voice enhancement (RVE), leveraging each technique to achieve improved audio output. This approach can be advantageous for improving the performance of headset speakers, which often lack adequate low-frequency response to effectively support ANC.
Apparatus and method for encoding and decoding an audio signal using an aligned look-ahead portion
Tue, 02 Jun 2015 08:00:00 EDT
An apparatus for encoding an audio signal having a stream of audio samples has: a windower for applying a prediction coding analysis window to the stream of audio samples to obtain windowed data for a prediction analysis and for applying a transform coding analysis window to the stream of audio samples to obtain windowed data for a transform analysis, wherein the transform coding analysis window is associated with audio samples within a current frame of audio samples and with audio samples of a predefined portion of a future frame of audio samples being a transform-coding look-ahead portion, wherein the prediction coding analysis window is associated with at least the portion of the audio samples of the current frame and with audio samples of a predefined portion of the future frame being a prediction coding look-ahead portion, wherein the transform coding look-ahead portion and the prediction coding look-ahead portion are identically to each other or are different from each other by less than 20%; and an encoding processor for generating prediction coded data or for generating transform coded data.
Multilingual electronic transfer dictionary containing topical codes and method of use
Tue, 26 May 2015 08:00:00 EDT
A multilingual electronic transfer dictionary provides for automatic topic disambiguation by including one or more topic codes in definitions contained the dictionary. Automatic topic disambiguation is accomplished by determining the frequencies of topic codes within a block of text. Dictionary entries having more frequently occurring topic codes are preferentially selected over those having less frequently occurring topic codes. When the topic codes are members of a hierarchical topical coding system, such as the International Patent Classification system, an iterative method can be used with starts with a coarser level of the coding system and is repeated at finer levels until an ambiguity is resolved. The dictionary is advantageously used for machine translation, e.g. between Japanese and English.
Inmate information center for correctional facility processing
Tue, 26 May 2015 08:00:00 EDT
A platform application and methods of operation that integrate both native and third-party modules into an integrated environment on an inmate computing device is disclosed. Third-party modules or systems are applications meant to operate independent from the platform application. Information is communicated between the platform application and third-party module or system to add audit, alarm and other functions across all modules or systems controlled by the platform software. The third-party module or system is audited to allow triggering of rules that cause remedial action to be taken. Triggers can be on actions not monitored by a particular third-party module or system.
Weighted N-finger scaling and scrolling
Tue, 26 May 2015 08:00:00 EDT
In one example, a method includes receiving an indication of an input gesture detected at a presence-sensitive input device, where the input gesture includes one or more input points and each input point is detected at a respective location of the presence-sensitive input device. The method may also include determining a focal point of the input gesture, and determining a radius length. The method may also include determining a shape centered at the focal point and having a size determined based on the radius length. The method may also include responding to a change in a geometric property of the shape by scaling information included in a graphical user interface, where the scaling of the information being centered at the focal point.
Image-based character recognition
Tue, 26 May 2015 08:00:00 EDT
Various embodiments enable a device to perform tasks such as processing an image to recognize and locate text in the image, and providing the recognized text an application executing on the device for performing a function (e.g., calling a number, opening an internet browser, etc.) associated with the recognized text. In at least one embodiment, processing the image includes substantially simultaneously or concurrently processing the image with at least two recognition engines, such as at least two optical character recognition (OCR) engines, running in a multithreaded mode. In at least one embodiment, the recognition engines can be tuned so that their respective processing speeds are roughly the same. Utilizing multiple recognition engines enables processing latency to be close to that of using only one recognition engine.
Audio signal decoder, time warp contour data provider, method and computer program
Tue, 26 May 2015 08:00:00 EDT
An audio signal decoder has a time warp contour calculator, a time warp contour data rescaler and a warp decoder. The time warp contour calculator is configured to generate time warp contour data repeatedly restarting from a predetermined time warp contour start value, based on time warp contour evolution information describing a temporal evolution of the time warp contour. The time warp contour data rescaler is configured to rescale at least a portion of the time warp contour data such that a discontinuity at a restart is avoided, reduced or eliminated in a rescaled version of the time warp contour. The warp decoder is configured to provide the decoded audio signal representation, based on an encoded audio signal representation and using the rescaled version of the time warp contour.
Multi-resolution switched audio encoding/decoding scheme
Tue, 26 May 2015 08:00:00 EDT
An audio encoder for encoding an audio signal has a first coding branch, the first coding branch comprising a first converter for converting a signal from a time domain into a frequency domain. Furthermore, the audio encoder has a second coding branch comprising a second time/frequency converter. Additionally, a signal analyzer for analyzing the audio signal is provided. The signal analyzer, on the hand, determines whether an audio portion is effective in the encoder output signal as a first encoded signal from the first encoding branch or as a second encoded signal from a second encoding branch. On the other hand, the signal analyzer determines a time/frequency resolution to be applied by the converters when generating the encoded signals. An output interface includes, in addition to the first encoded signal and the second encoded signal, a resolution information identifying the resolution used by the first time/frequency converter and used by the second time/frequency converter.
Systems, methods, and apparatus for gain factor attenuation
Tue, 26 May 2015 08:00:00 EDT
A method of signal processing according to one embodiment includes calculating an envelope of a first signal that is based on a low-frequency portion of a speech signal, calculating an envelope of a second signal that is based on a high-frequency portion of the speech signal, and calculating a plurality of gain factor values according to a time-varying relation between the envelopes of the first and second signal. The method includes attenuating, based on a variation over time of a relation between the envelopes of the first and second signals, at least one of the plurality of gain factor values. In one example, the variation over time of a relation between the envelopes is indicated by at least one distance among the plurality of gain factor values.
Speech recognition and synthesis utilizing context dependent acoustic models containing decision trees
Tue, 26 May 2015 08:00:00 EDT
A speech recognition method including the steps of receiving a speech input from a known speaker of a sequence of observations and determining the likelihood of a sequence of words arising from the sequence of observations using an acoustic model. The acoustic model has a plurality of model parameters describing probability distributions which relate a word or part thereof to an observation and has been trained using first training data and adapted using second training data to said speaker. The speech recognition method also determines the likelihood of a sequence of observations occurring in a given language using a language model and combines the likelihoods determined by the acoustic model and the language model and outputs a sequence of words identified from said speech input signal. The acoustic model is context based for the speaker, the context based information being contained in the model using a plurality of decision trees and the structure of the decision trees is based on second training data.
Messaging response system providing translation and conversion written language into different spoken language
Tue, 26 May 2015 08:00:00 EDT
A messaging response system is disclosed wherein a service providing system provides services to users via messaging communications. In accordance with an exemplary embodiment of the present invention, multiple respondents servicing users through messaging communications may appear to simultaneously use a common “screen name” identifier.
Low power activation of a voice activated device
Tue, 26 May 2015 08:00:00 EDT
In a mobile device, a bone conduction or vibration sensor is used to detect the user's speech and the resulting output is used as the source for a low power Voice Trigger (VT) circuit that can activate the Automatic Speech Recognition (ASR) of the host device. This invention is applicable to mobile devices such as wearable computers with head mounted displays, mobile phones and wireless headsets and headphones which use speech recognition for the entering of input commands and control. The speech sensor can be a bone conduction microphone used to detect sound vibrations in the skull, or a vibration sensor, used to detect sound pressure vibrations from the user's speech. This VT circuit can be independent of any audio components of the host device and can therefore be designed to consume ultra-low power. Hence, this VT circuit can be active when the host device is in a sleeping state and can be used to wake the host device on detection of speech from the user. This VT circuit will be resistant to outside noise and react solely to the user's voice.
Biometric voice command and control switching device and method of use
Tue, 26 May 2015 08:00:00 EDT
A biometric voice command and control switching device has a microphone assembly for receiving a currently spoken challenge utterance and a reference utterance, and a voice processing circuit for creating electronic signals indicative thereof. The device further includes a memory for storing the electronic signals, and a processor for comparing the electronic signals to determine if there is a match. If there is a match, an interface circuit enables the operable control of the controlled device.
Language model creation device
Tue, 26 May 2015 08:00:00 EDT
This device 301 stores a first content-specific language model representing a probability that a specific word appears in a word sequence representing a first content, and a second content-specific language model representing a probability that the specific word appears in a word sequence representing a second content. Based on a first probability parameter representing a probability that a content represented by a target word sequence included in a speech recognition hypothesis generated by a speech recognition process of recognizing a word sequence corresponding to a speech, a second probability parameter representing a probability that the content represented by the target word sequence is a second content, the first content-specific language model and the second content-specific language model, the device creates a language model representing a probability that the specific word appears in a word sequence corresponding to a part corresponding to the target word sequence of the speech.
System, method and program product for providing automatic speech recognition (ASR) in a shared resource environment
Tue, 26 May 2015 08:00:00 EDT
A speech recognition system, method of recognizing speech and a computer program product therefor. A client device identified with a context for an associated user selectively streams audio to a provider computer, e.g., a cloud computer. Speech recognition receives streaming audio, maps utterances to specific textual candidates and determines a likelihood of a correct match for each mapped textual candidate. A context model selectively winnows candidate to resolve recognition ambiguity according to context whenever multiple textual candidates are recognized as potential matches for the same mapped utterance. Matches are used to update the context model, which may be used for multiple users in the same context.
Speaker recognition from telephone calls
Tue, 26 May 2015 08:00:00 EDT
The present invention relates to a method for speaker recognition, comprising the steps of obtaining and storing speaker information for at least one target speaker; obtaining a plurality of speech samples from a plurality of telephone calls from at least one unknown speaker; classifying the speech samples according to the at least one unknown speaker thereby providing speaker-dependent classes of speech samples; extracting speaker information for the speech samples of each of the speaker-dependent classes of speech samples; combining the extracted speaker information for each of the speaker-dependent classes of speech samples; comparing the combined extracted speaker information for each of the speaker-dependent classes of speech samples with the stored speaker information for the at least one target speaker to obtain at least one comparison result; and determining whether one of the at least one unknown speakers is identical with the at least one target speaker based on the at least one comparison result.
System and methods for matching an utterance to a template hierarchy
Tue, 26 May 2015 08:00:00 EDT
A system and methods for matching at least one word of an utterance against a set of template hierarchies to select the best matching template or set of templates corresponding to the utterance. Certain embodiments of the system and methods determines at least one exact, inexact, and partial match between the at least one word of the utterance and at least one term within the template hierarchy to select and populate a template or set of templates corresponding to the utterance. The populated template or set of templates may then be used to generate a narrative template or a report template.
Dynamic language model
Tue, 26 May 2015 08:00:00 EDT
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for speech recognition. One of the methods includes receiving a base language model for speech recognition including a first word sequence having a base probability value; receiving a voice search query associated with a query context; determining that a customized language model is to be used when the query context satisfies one or more criteria associated with the customized language model; obtaining the customized language model, the customized language model including the first word sequence having an adjusted probability value being the base probability value adjusted according to the query context; and converting the voice search query to a text search query based on one or more probabilities, each of the probabilities corresponding to a word sequence in a group of one or more word sequences, the group including the first word sequence having the adjusted probability value.
Thought recollection and speech assistance device
Tue, 26 May 2015 08:00:00 EDT
Some embodiments of the inventive subject matter include a method for detecting speech loss and supplying appropriate recollection data to the user. Such embodiments include detecting a speech stream from a user, converting the speech stream to text, storing the text, detecting an interruption to the speech stream, wherein the interruption to the speech stream indicates speech loss by the user, searching a catalog using the text as a search parameter to find relevant catalog data and, presenting the relevant catalog data to remind the user about the speech stream.
Audio encoder, audio decoder, methods for encoding and decoding an audio signal, and a computer program
Tue, 26 May 2015 08:00:00 EDT
An encoder for providing an audio stream on the basis of a transform-domain representation of an input audio signal includes a quantization error calculator configured to determine a multi-band quantization error over a plurality of frequency bands of the input audio signal for which separate band gain information is available. The encoder also includes an audio stream provider for providing the audio stream such that the audio stream includes information describing an audio content of the frequency bands and information describing the multi-band quantization error. A decoder for providing a decoded representation of an audio signal on the basis of an encoded audio stream representing spectral components of frequency bands of the audio signal includes a noise filler for introducing noise into spectral components of a plurality of frequency bands to which separate frequency band gain information is associated on the basis of a common multi-band noise intensity value.
Encoder, decoder and methods for encoding and decoding data segments representing a time-domain data stream
Tue, 26 May 2015 08:00:00 EDT
An apparatus for decoding data segments representing a time-domain data stream, a data segment being encoded in the time domain or in the frequency domain, a data segment being encoded in the frequency domain having successive blocks of data representing successive and overlapping blocks of time-domain data samples. The apparatus includes a time-domain decoder for decoding a data segment being encoded in the time domain and a processor for processing the data segment being encoded in the frequency domain and output data of the time-domain decoder to obtain overlapping time-domain data blocks. The apparatus further includes an overlap/add-combiner for combining the overlapping time-domain data blocks to obtain a decoded data segment of the time-domain data stream.
Method and apparatus for processing audio frames to transition between different codecs
Tue, 26 May 2015 08:00:00 EDT
A method (700, 800) and apparatus (100, 200) processes audio frames to transition between different codecs. The method can include producing (720), using a first coding method, a first frame of coded output audio samples by coding a first audio frame in a sequence of frames. The method can include forming (730) an overlap-add portion of the first frame using the first coding method. The method can include generating (740) a combination first frame of coded audio samples based on combining the first frame of coded output audio samples with the overlap-add portion of the first frame. The method can include initializing (760) a state of a second coding method based on the combination first frame of coded audio samples. The method can include constructing (770) an output signal based on the initialized state of the second coding method.
Adaptive grouping of parameters for enhanced coding efficiency
Tue, 26 May 2015 08:00:00 EDT
The present invention is based on the finding that parameters including: a first set of parameters of a representation of a first portion of an original signal and a second set of parameters of a representation of a second portion of the original signal can be efficiently encoded when the parameters are arranged in a first sequence of tuples and a second sequence of tuples. The first sequence of tuples includes tuples of parameters having two parameters from a single portion of the original signal and the second sequence of tuples includes tuples of parameters having one parameter from the first portion and one parameter from the second portion of the original signal. A bit estimator estimates the number of necessary bits to encode the first and the second sequence of tuples. Only the sequence of tuples, which results in the lower number of bits, is encoded.
Manner of pronunciation-influenced search results
Tue, 26 May 2015 08:00:00 EDT
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating search results. In one aspect, a method includes obtaining a transcription of a voice query, and data that identifies an accent of the voice query, submitting the transcription and the data that identifies the accent of the voice query to a search engine to generate one or more accent-influenced results of the voice query, and providing the accent-influenced results to a client device for display.

Language Selection

linkedin        twitter

Company Search


Advanced Search   


Services


logo inttranet


logo inttrastats


logo Inttranews


logo Inttrasearch


logo linguists of the year