Which device encodes




















Year of fee payment : 6. Ref country code : SK. Ref country code : RO. Ref country code : CZ. Ref country code : SM. Ref country code : BE. Ref country code : BG. Ref country code : SI. Ref legal event code : PL. Ref country code : MC. Ref legal event code : MM4A.

Ref country code : LI. Ref country code : LU. Year of fee payment : 7. Ref country code : FI. Payment date : Ref country code : IT.

Ref country code : CY. Ref country code : IS. Ref country code : MK. Ref country code : MT. Ref country code : TR. Ref country code : AL. Ref legal event code : FD2A. Year of fee payment : Provisional Patent Application Serial No. The present disclosure relates generally to electronic devices. More specifically, the present disclosure relates to devices for adaptively encoding and decoding a watermarked signal.

In the last several decades, the use of electronic devices has become common. In particular, advances in electronic technology have reduced the cost of increasingly complex and useful electronic devices. Cost reduction and consumer demand have proliferated the use of electronic devices such that they are practically ubiquitous in modern society.

As the use of electronic devices has expanded, so has the demand for new and improved features of electronic devices. More specifically, electronic devices that perform functions faster, more efficiently or with higher quality are often sought after. Some electronic devices e. These electronic devices may encode speech signals for storage or transmission.

For example, a cellular phone captures a user's voice or speech using a microphone. For instance, the cellular phone converts an acoustic signal into an electronic signal using the microphone. This electronic signal may then be formatted for transmission to another device e. Improved quality or additional capacity in a communicated signal is often sought for. For example, cellular phone users may desire greater quality in a communicated speech signal.

As can be observed from this discussion, systems and methods that allow efficient signal communication may be beneficial. An embedding judgement determines whether or not data should be embedded in a predetermined parameter code using control parameters.

In particular, if a frame is judged to be non-speech, then data is embedded. The publication B. Geiser and P. The invention is defined by the electronic devices according to claims 1,6, the methods according to claims 12, 13 and a computer-program product according to claim The systems and methods disclosed herein may be applied to a variety of electronic devices. Examples of electronic devices include voice recorders, video cameras, audio players e.

One kind of electronic device is a communication device, which may communicate with another device. Examples of communication devices include telephones, laptop computers, desktop computers, cellular phones, smartphones, wireless or wired modems, e-readers, tablet devices, gaming systems, cellular telephone base stations or nodes, access points, wireless gateways and wireless routers. Other examples of standards that a communication device may comply with include IEEE For example, some communication devices may communicate with other devices using an Ethernet protocol.

In one configuration, the systems and methods disclosed herein may be applied to a communication device that communicates with another device using a satellite. As used herein, the term "couple" may denote a direct connection or an indirect connection. For example, if a first component is coupled to a second component, the first component may be directly connected to the second component or may be indirectly connected to the second component through a third component, for example.

The systems and methods disclosed herein describe adaptive watermarking. For example, the systems and methods disclosed herein may be used for adaptive watermarking for algebraic code excited linear prediction ACELP codecs. Watermarking or data hiding in speech codec bitstreams allows the transmission of extra data in-band with no changes to the network infrastructure.

This can be used for a range of applications e. One possible application of the systems and methods disclosed herein is bandwidth extension, in which one codec's bitstream e. Decoding the carrier bitstream and the hidden bits allows synthesis of a bandwidth that is greater than the bandwidth of the carrier codec e.

For example, a standard narrowband codec can be used to encode a kilohertz kHz low-band part of speech, while a kHz high-band part is encoded separately. The bits for the high band may be hidden within the narrowband speech bitstream. In this case, wideband may be decoded at the receiver despite using the legacy narrowband bitstream. In another example, a standard wideband codec may be used to encode a kHz low-band part of speech, while a kHz high-band part is encoded separately and hidden in the wideband bitstream.

In this case, super-wideband may be decoded at the receiver despite using the legacy wideband bitstream. The bits are hidden by restricting the number of allowed pulse combinations. In the case of AMR-NB, where there are two pulses per track, one approach includes constraining the pulse positions so that an exclusive OR XOR of the two pulse positions on a given track are equal to the watermark to transmit.

One or two bits per track may be transmitted this way. In practice, this can add significant distortion as it may significantly alter the main pitch pulses. This may be especially detrimental for bandwidth extension applications where the low band excitation is used to generate the high band excitation, as the low band degradation may also cause degradation in the high band.

In the systems and methods disclosed herein, the watermark is made adaptive. Instead of embedding a fixed number of bits per pulse track e. This may be done, for example, using information already present at both an encoder and decoder, such that information indicating which tracks are perceptually most important does not need to be additionally or separately transmitted.

In one configuration, a long term prediction LTP contribution may be used to protect the most important tracks from the watermark. For instance, the LTP contribution normally exhibits clear peaks at the main pitch pulse, and may be available already at both encoder and decoder.

In one example, two tracks corresponding to the highest absolute values of the LTP contribution may be deemed important or designated "high priority" tracks and are not watermarked. The other three tracks are likely to be less important and may be designated or referred to as "low priority" tracks, for example , and may receive a watermark. Thus, if the three remaining tracks are watermarked with two bits each, this leads to six bits of watermark per five millisecond ms subframe, for a total of 1.

One refinement provided by the systems and methods disclosed herein may include replacing the LTP contribution by a memory-limited LTP contribution because the LTP signal is sensitive to errors and packet losses and errors may propagate indefinitely. This may lead to the encoder and decoder being out of sync for long periods after an erasure or bit errors. Instead, a memory-limited version of the LTP may be constructed based only on the quantized pitch values and codebook contributions of the last N frames plus the current frame.

Gains may be set to unity. It should be noted that the original LTP may be used for the low band coding. In some configurations, the memory-limited LTP may be used solely for determining the priority of the tracks for watermarking purposes. Adapting the watermark to the speech characteristics may allow better speech quality by hiding the watermark where it is perceptually less important. In particular, preserving the pitch pulse may have a positive impact on speech quality.

When the systems and methods described herein are not used, for instance, the quality impact of a watermark at the same bit rate may be more severe.

In some configurations, the systems and methods disclosed herein may be used to provide a codec that is a backward interoperable version of narrowband AMR For convenience, this codec may be referred to as "eAMR" herein, though the codec could be referred to using a different term.

This may provide true wideband encoding and not blind bandwidth extension. The watermark used may have a negligible impact on narrowband quality for legacy interoperation. With the watermark, narrowband quality may be slightly degraded in comparison with AMR In some configurations, an encoder may detect a legacy remote through not detecting a watermark on the return channel, for example and stop adding watermark, returning to legacy AMR In some configurations, eAMR may require new handsets with wideband acoustics, for example.

However, this may or may not affect eAMR operation. AMR-WB may offer true wideband quality. AMR-WB may use a bit rate of AMR-WB may require new handsets with wideband acoustics, for example and infrastructure modifications. More detail on one example of an AMR The codebook excitation is made of pulses and allows efficient computations. In Enhanced Full Rate EFR , each 20 millisecond ms frame of samples, for example is split into 4x5 ms frames of 40 samples.

Each subframe of 40 samples is split into five interleaved tracks with eight positions per track. Two pulses and one sign bit may be used per track, where the order of pulses determines the second sign. Stacking may be allowed. One example of tracks, pulses, amplitudes and positions that may be used according to an ACELP fixed codebook is given in Table 1.

One example of a watermarking scheme is given as follows. A watermark may be added to a fixed codebook FCB by limiting the pulse combinations allowed. Watermarking in an AMR Basically, the XOR of the last bit of the two indexes pos0 and pos1 may be constrained to be equal to the chosen bit of information to be transmitted e. This leads to one bit per track e. For instance, the XOR of the two least significant bits LSBs of the indexes may be constrained to be the two bits of information to be transmitted.

For example, a search may be performed over pulse positions that will decode into the correct watermark. This approach may provide low complexity.

In this approach, however, the main pitch pulse may be significantly affected e. In accordance with the systems and methods disclosed herein, tracks with the most impact may be identified and not watermarked.

In one approach, a long term prediction LTP contribution may be used to identify two important e. However, this approach may require an identical LTP contribution at an encoder and a decoder.

In another approach, limited-memory LTP may be used. In this approach, an LTP contribution may be recomputed using only M past frames of excitation and pitch lags. This may eliminate error propagation beyond M frames. It should be noted that a single frame loss may imply that potentially three frames are lost for a high band when a bad frame indication from the low band is provided to the high band.

More specifically, a bad frame indication BFI is a flag that a channel decoder provides to a speech decoder, indicating when it has failed to properly decode a frame. The decoder may then ignore the received data and perform error concealment. Error concealment may then be performed on the high band e.

It should be noted that although a For example, one operating point of eAMR is In one configuration of the systems and methods disclosed herein, lower rates may be used e.

Thus, bandwidth switching between narrowband and wideband, for example may be a challenge. Wideband speech, for example, may be maintained with lower rates of eAMR. Each rate may use a watermarking scheme. For example, the watermarking scheme used for a A limited-memory LTP scheme could be used for other rates.

Table 2 illustrates examples of bit allocations per frame for differing rates. Table 2 Rate kbps One configuration of the systems and methods disclosed herein may be used for the extension of code-excited linear prediction CELP speech coders using watermarking techniques to embed data.

Wideband e. However, the majority of existing mobile communication networks support narrowband coding only e. Deploying wideband coders e. Furthermore, the next generation of services may support wideband coders e. Again, operators may eventually face the costs of deploying yet another codec to move customers to super-wideband. One configuration of the systems and methods disclosed herein may use an advanced model that can encode additional bandwidth very efficiently and hide this information in a bitstream already supported by existing network infrastructure.

The information hiding may be performed by watermarking the bitstream. One example of this technique watermarks the fixed codebook of a CELP coder. For example, the upper band of a wideband input e. In another example, the upper band of a super-wideband input e.

Other secondary bitstreams, perhaps unrelated to bandwidth extension, may be carried as well. This technique allows the encoder to produce a bitstream compatible with existing infrastructures. A legacy decoder may produce a narrowband output with a quality similar to standard encoded speech without the watermark, for example , while a decoder that is aware of the watermark may produce wideband speech. Various configurations are now described with reference to the Figures, where like element names may indicate functionally similar elements.

Do you want to get pizza tonight? Figure 1. The communication process. Encoding, media, and decoding Hawkins, Encoded messages are sent through a channel, or a sensory route, on which a message travels to the receiver for decoding.

If your roommate has headphones on and is engrossed in a video game, you may need to get their attention by waving your hands before you can ask them about dinner. This model focuses on the sender and message within a communication encounter. Although the receiver is included in the model, this role is viewed as more of a target or end point rather than part of an ongoing process.

Moreover, temporal-spatial image processing is performed so that the difficult-to-compress noise components are removed in advance from the 4K images. That enables achieving control on the decline in the feel of resolution at a low bit rate. For that reason, even in the case in which the additional bit rate is set to 10 Mbps, it becomes possible to maintain the feel of resolution of the 4K images.

In order to distribute videos of 4K resolution, it is necessary to have the bit rate of, for example, about 30 Mbps. However, as a result of implementing the HEVC encoding technology, the additional transmission bandwidth that is required can be reduced to about 10 Mbps that is one-third of 30 Mbps. In a narrow-bandwidth video transmission system in which SHVC is used, reference images are shared in advance between the transmission side and the reception side.

The transmission side transmits encoding data representing only the difference with the reference images. The reception side decodes the received encoding data using the reference images shared in advance. Since the encoding data represents only the difference with the reference images, it becomes possible to reduce the volume of data and to perform video streaming having excellent bandwidth usage efficiency.

According to an embodiment, an encoding device includes a storage controller, a difference generator, and a communication controller. The storage controller is configured to store a base stream in a storage. The base stream serves as basis for encoding and decoding and is shared with a decoding device. The difference generator is configured to generate a difference bit stream that represents difference between an input bit stream that has been input and the base stream. The communication controller is configured to control a communicating unit to transmit the difference bit stream.

Exemplary embodiments of a streaming system are described below in detail with reference to the accompanying drawings. The streaming system according to the embodiments is, for example, a system that transmits and receives videos for surveillance purposes. An encoder on the transmission side generates in advance, for each predetermined screen pattern, a base stream of, for example, intra-picture NAL units serving as the basis for encoding and decoding, and stores the base stream.

Each generated base stream is also stored in the decoder on the reception side. Thus, the base streams are shared between the encoder side and the decoder side. The timing of storing the base streams in the decoder on the reception side can be at the time of the factory shipment or the initial setup of the decoder on the reception side. Alternatively, the base streams are dynamically updated in an asynchronous or synchronous manner during the actual stream transmission.

In the case of encoding input videos, the encoder makes use of the base streams shared with the decoder side and encodes the input videos; and, based on the encoding state, generates difference bit streams of P picture NAL units, for example.

Thus, only the difference bit streams are transmitted to the decoder. That enables achieving reduction in the volume of data to be transmitted.

Moreover, the difference bit streams to be transmitted to the decoder can have information included therein to indicate whether or not base streams are used. The decoder on the receiver side makes use of the base streams shared in advance and the received difference bit streams, and reconfigures bit streams equivalent to having the identical image quality and the identical resolution to the bit streams that were originally supposed to be generated by the encoder.

In other words, the decoder reconfigures the bit streams that are reproduced using the base streams shared in advance and the received difference bit streams and that represent the state in which the screens corresponding to the difference bit streams are encoded.

Then, the decoder decodes the reconfigured streams and displays them. Hence, during screen transmission, it becomes possible to perform video streaming having excellent bandwidth usage efficiency.

Moreover, it becomes possible to make a dedicated decoder redundant. Furthermore, the amount of processing for reconfiguration can be reduced as compared to the transcode too. Thus, the receiving device decoder can be implemented using a general-purpose decoder chip and an inexpensive central processing unit CPU , thereby enabling achieving reduction in the cost of the streaming system. As illustrated in FIG. The server device 1 and the client device 2 are connected to each other via a network 3 such as the Internet.

The server device 1 performs scalable expansion of predetermined video streams using the SHVC encoding technology; lowers the bitrate of high-resolution video signals, such as 4K video signals or 8K video signals, while maintaining the feel of high resolution thereof; and distributes the video signals via the network 3.

The client device 2 represents a general-purpose HEVC decoder that, as described later, decodes the video signals that have been subjected to SHVC encoding and reproduces the decoded video signals without any special configuration. The decoded video signals are either displayed on a monitor device or recorded in a recorder device. The CPU 11 to the communicating unit 15 are connected to each other via a bus line The HDD 14 is used to store an encoding program for encoding video streams.

The CPU 11 performs operations according to the encoding program and functions as an encoder 17 for performing SHVC encoding of video streams and distributing encoded video streams. The CPU 21 to the communicating unit 25 are connected to each other via a bus line The HDD 24 is used to store a decoding program for decoding the SHVC-encoded video streams that are transmitted from the server device 1.

The CPU 21 performs operations according to the decoding program and functions as a decoder 27 for decoding the video streams received from the server device 1.

The following explanation is given for an example in which the encoder 17 and the decoder 27 are implemented using software.

This layer varies, however. Insulating the upper layers from complexities of the other layers is accomplished by providing functions necessary to guarantee a reliable network link. Examples include error recovery and flow control between the two end points of the network connection. The session layer oversees user connections and manages the interaction between different stations.

Services include session establishment, maintenance and termination, as well as session support. The presentation layer formats the data that is presented to the application layer. These transformations provide a common interface for user applications. The layer provides character code translation, data conversion, data compression and data encryption. The final layer serves as the window for users and application process to access network services.

The application layer provides a wealth of services due to the potentially wide variety of applications involved. These can include remote file and printer access, resource sharing and device redirection, directory services and electronic messaging. Data encoding is a basic concept within computer science. With an online computer science degree from Concordia University Texas, you can learn the knowledge and skills needed to pursue a rewarding career in this field. These graduates also have a high full-time employment rate You understand that these contacts may be generated using automated technology and that you are not required to give this consent to enroll in programs with the school.

Call CTX. Blog Posts Start the conversation.



0コメント

  • 1000 / 1000