14 Oct 2024, 10:16 AM
23 Feb 2025, 10:22 PM

5. Watermarking embedding and detection

Watermarking Embedding

In the context of watermarking, the embedding process is where a watermark is inserted into a host asset (such as an image, audio, or video) while maintaining the quality and integrity of the original content. Depending on the strategy, the watermark embedding process can either be independent of the host asset or can adapt to it. The latter, known as informed watermarking, adjusts the embedding to the specific characteristics of the host, optimizing invisibility and robustness.

Additive Watermarking

A simple and widely used embedding strategy is additive watermarking, where the watermark signal is directly added to the host signal. The formula used for additive watermarking is:

fw,i=fi+γwi

Where:

This approach is simple and flexible, allowing the watermark strength to be adapted to the local characteristics of the host. For instance, the watermarking strength can be locally adjusted as:

fw,i=fi+γiwi

Or, in cases where the watermark strength depends on the host feature fi, it can be expressed as:

fw,i=fi+γiwifi

This technique assumes that the host features follow a Gaussian distribution and that the primary form of attack is the addition of white Gaussian noise (AWGN). Under these conditions, correlation-based decoding becomes optimal, minimizing the overall error probability for a given false detection rate.

Example of Additive Watermarking in DWT Domain

A well-known application of additive watermarking can be found in the work of Barni et al. (IEEE Transactions on Image Processing, 2001). The authors propose a watermarking method based on the Discrete Wavelet Transform (DWT). The watermark is embedded in the wavelet coefficients of the host image, and the modification is limited to the amount that can be tolerated without degrading the visual quality. This is achieved by exploiting perceptual rules, such as:

The watermarking process is further optimized using a weighting function that takes into account the local sensitivity of different regions of the image to noise. This function ensures that the watermark is embedded more strongly in areas where it is less likely to be noticed, thus improving the invisibility of the watermark while maintaining robustness. The method effectively places the watermark in regions where the human eye is less perceptive to changes, allowing it to be hidden more securely.

The watermark itself is a binary pseudorandom sequence, which is embedded in the DWT coefficients of the three largest detail subbands of the image. These subbands correspond to the areas where modifications are less perceptible, ensuring that the watermark is well hidden. The DWT is particularly effective in this context, as it allows the identification of image regions where disturbances can be masked more easily, leveraging the multi-scale analysis inherent to the wavelet domain.

For the detection phase, Barni et al. adopt a strategy based on the Newman-Pearson theorem, which ensures that the probability of detecting the watermark is maximized while minimizing the risk of false positives. This approach provides a reliable detection mechanism, even in scenarios where the watermarked image may have been subjected to various forms of attack or degradation.

By combining these perceptual insights with the DWT's ability to localize modifications in less noticeable regions, Barni et al. achieve an effective balance between robustness and invisibility, making their watermarking technique both practical and resilient.

Multiplicative Watermarking

In contrast to additive watermarking, multiplicative watermarking scales the watermark based on the energy of the corresponding host feature. The watermark signal is embedded as:

fw,i=fi+γwifior alternativelyfw,i=fi+γwi|fi|

This technique is more commonly used in the frequency domain, as it leverages the fact that disturbances at certain frequencies are less perceptible if the host asset already contains similar frequency components. It also makes it harder for attackers to estimate the watermark by averaging multiple watermarked versions of the host asset.

Multiplicative watermarking tends to insert the watermark into the most important parts of the host, making it more challenging to analyze and remove. Furthermore, traditional results from digital communication and information theory may not directly apply because they often assume additive noise, while multiplicative watermarking introduces different types of distortions.

Watermarking Robustness and Optimization

One of the key goals in watermarking is to achieve both invisibility and robustness. These two requirements often conflict, as increasing the strength of the watermark to improve robustness may make it more visible. Therefore, finding an optimal balance is crucial. In many cases, the watermark is embedded in important areas of the host asset to simultaneously meet invisibility and robustness constraints.

However, achieving this balance requires solving complex optimization problems, as traditional communication theory does not fully apply to the type of distortions introduced by watermarking. Moreover, multiplicative watermarking, in particular, requires sophisticated detection and decoding mechanisms due to its non-additive nature.

Watermarking Detection

In watermarking, once the watermark has been embedded in a host asset, the next crucial step is detection. The ability to detect the watermark reliably, even after the host signal has been attacked or modified, is essential for ensuring the watermark’s robustness and effectiveness.

In most cases, the structure of the detector (or decoder) is derived using a simplified channel model. This simplification is necessary due to the wide range of possible attacks and the difficulties in developing accurate statistical models of the host's features. In general, watermark detection can be treated as a signal detection problem in a noisy environment, where the noise may include both the unknown characteristics of the host signal and the effects of possible attacks.

A or Aw to indicate that it may not coincide with either the original or the watermarked version of the asset. The decision is based on the set of observed variables, f, which are the features extracted from A.

Reformulating Detection as Hypothesis Testing

We can reformulate the detection problem as a hypothesis testing problem, a classic statistical approach where the goal is to choose between two hypotheses based on observed data. In this context, the detection problem can be formulated as follows:

H0:A does not contain the watermark wH1:A contains the watermark w

Where:

The goal of the detection process is to define a test that distinguishes between H1 and the composite alternative hypothesis H0. The test must be optimum with respect to a predefined criterion, such as minimizing error rates or maximizing detection robustness.

Bayes Risk

In Bayesian hypothesis testing, decisions are made by minimizing the Bayes risk. The Bayes risk is defined as the expected value of a loss function that penalizes incorrect decisions. For our case, the loss function Lij is defined as follows:

The observation variables correspond to the vector of extracted features f, and the decision rule maps f to a binary decision:

The decision regions are:

The decision rule can be expressed as:

ϕ(f)={1if fR1(we accept H1)0if fR0(we accept H0)

We said we want to minimize the bayes risk, and this leads to a decision criterion based on the likelihood ratio.

Minimization of Bayes Risk and Likelihood Ratio Test

The likelihood ratio is a fundamental concept in hypothesis testing, where the goal is to compare the probability of the observed data under two competing hypotheses: the presence or absence of the watermark.

The likelihood ratio, λ(f), is defined as:

λ(f)=p(f|H1)p(f|H0)

Where:

Let p0 and p1 be the a priori probabilities of hypotheses H0 and H1, respectively. The decision rule for minimizing the Bayes risk can be defined as:

ϕ(f)={1if p(f|H1)p(f|H0)>p0L01p1L100otherwise

Where:

This decision rule implies that the detector operates by comparing the likelihood ratio to a detection threshold. The threshold, T, is derived based on the probabilities and the associated losses.

Detection Threshold

The detection threshold T is given by:

T=p0L01p1L10

Setting the threshold is critical to minimize the overall error probability Pe. The error probability is a weighted sum of the probabilities of false alarms and missed detections, defined as:

Pe=p0Pf+p1(1Pd)

Where:

To minimize Pe, the optimal detection threshold T should be set to 1 when:

In this case, the false alarm probability Pf and the detection probability Pd are balanced, meaning:

Pf=1Pd

Thus, the minimum error probability is achieved when the probability of missing the watermark and the probability of falsely detecting the watermark are equal.

Challenges in Practical Applications

While the Bayes decision theory provides an optimal solution in theory, in practical applications, several challenges arise:

  1. Unknown A Priori Probabilities:

    • In many cases, the a priori probabilities p0 and p1 are unknown, making it difficult to calculate the exact threshold.
  2. Non-Gaussian Attacks:

    • When the attack model deviates from additive white Gaussian noise (AWGN), the probability of missing the watermark increases, complicating detection.
  3. False Detection Constraints:

    • In many real-world scenarios, it is often not acceptable for the false detection probability Pf to exceed a certain level. For example, false positives in copyright protection can lead to unnecessary legal disputes or penalties.

Alternative Optimization Criteria

Given the limitations in practical watermark detection systems, an alternative optimization criterion is often adopted. Instead of minimizing the overall error probability Pe, the goal is to minimize the probability of missing the watermark** Pm (where Pm=1Pd) subject to a constraint on the maximum allowable false alarm probability Pfmax.

This criterion is more practical for applications where false alarms have a higher cost. The steps for optimizing under this criterion are:

  1. Set the Maximum Allowed False Alarm Probability:

    • Define a maximum allowable value for Pf, based on the specific application.
  2. Use the Likelihood Ratio as the Detection Statistic:

    • Continue using the likelihood ratio λ(f)=p(f|H1)p(f|H0) as the detection statistic.
  3. Determine the Threshold According to the False Alarm Probability:

    • Adjust the threshold T to ensure that the false alarm probability does not exceed the predefined maximum value Pfmax.

In practical terms, this means that the detection system is fine-tuned to minimize missed detections, ensuring that the watermark is reliably detected when present, while keeping the false alarm rate within acceptable limits.

Ecco la versione finale dei tuoi appunti trasformata in un capitolo coerente e adatto per il tuo libro. Ho organizzato il testo in modo fluido, mantenendo la formalità e la chiarezza necessarie per un libro tecnico, e includendo le formule in formato LaTeX.

Neyman-Pearson Detection Criterion

In many practical applications of watermark detection, it is often preferable to adopt a detection criterion that focuses on minimizing the probability of missing the watermark while maintaining control over the false alarm probability. The most widely used approach in such cases is the Neyman-Pearson detection criterion, which provides a framework for optimizing the detector’s performance under these constraints.

Minimizing Missed Detection Subject to a False Alarm Constraint

The Neyman-Pearson criterion aims to maximize the probability of correctly detecting the watermark (i.e., minimizing the missed detection probability Pm while ensuring that the false alarm probability Pf does not exceed a prescribed limit. The optimization problem can be described as:

P{λ(f)>TH0}=Pf

Where:

Once the detection threshold T is set based on the desired false alarm rate, the probability of missing the watermark can be computed as:

1Pd=Tp(λH1)dλ

Where:

Performance Evaluation Using ROC Curves

The performance of a detector based on the Neyman-Pearson criterion is typically evaluated using Receiver Operating Characteristic (ROC) curves. These curves plot:

The ROC curve provides a visual representation of the trade-off between detection accuracy and false alarms. A steeper ROC curve indicates better detector performance, as it reflects higher detection rates for lower false alarm probabilities.

The decision rule for the Neyman-Pearson criterion is given by:

ϕ(f)={1if λ(f)>T0otherwise

AWGN Channel Model

To model a specific watermark embedding and detection strategy, we need to consider both the host signal features and the attacks on the watermarked signal. A common and simplified model used in watermarking is the Additive White Gaussian Noise (AWGN) channel. In this context, we assume:

In this scenario, the watermarked signal fw,i is described by the equation:

fw,i=fi+γwi+ni

Where:

Additionally, the model assumes that:

Optimum Detection in the AWGN Case

In the case of an AWGN channel, the optimum detection strategy is to use correlation detection, where the decision statistic is based on the correlation between the received signal and the watermark.

The decision rule for correlation detection is given by:

ϕ=1ni=1nfiwi

Where fi represents the received (potentially noisy) feature, and wi is the original watermark signal. The decision rule compares the correlation ϕ to a threshold Tϕ:

{ϕ>Tϕ(accept H1)ϕ<Tϕ(accept H0)

Computing False Alarm and Missed Detection Probabilities

In the AWGN case, the false alarm and missed detection probabilities can be computed, provided that the variance and mean of the host features are known. Let μρ and σρ represent the mean and variance of the correlation statistic under hypotheses H0 and H1, respectively.

The false alarm probability Pf and detection probability Pd are determined by the overlap between the distributions of the correlation statistic under the two hypotheses. By fixing the threshold Tϕ, we can control Pf and Pd.

AWGN Channel and Detector Performance

In the case of an Additive White Gaussian Noise (AWGN) channel, we can derive a formula that completely characterizes the performance of the watermark detector. This allows us to evaluate the detection performance in terms of false alarm probability Pf, missed detection probability Pd, and the signal-to-noise ratio (SNR).

The detection threshold Tϕ for the AWGN channel is given by:

Tϕ=2σx2σw2nerfc1(2Pf)

Where:

The overall performance of the detector, in terms of the missed detection probability 1Pd, is:

(1Pd)(Pf)=12erfc(nSNR2erfc1(2Pf))

Where the signal-to-noise ratio (SNR) is defined as:

SNR=γ2σw2σx2

Key Points About AWGN Channel Performance

  1. Detector Independence: The detector does not need to know the exact strength γ of the watermark used during embedding. This allows flexibility in the detection process, as the watermark strength can vary.

  2. Embedding Flexibility: The embedder can adjust the watermark strength γ based on the specific needs of the application, balancing the trade-off between imperceptibility (ensuring the watermark remains undetected by the human eye) and robustness (ensuring the watermark is detectable by the system). Importantly, the detector need not be informed about the specific strength chosen for embedding.