Source code for compressai.ops.ops

# Copyright (c) 2021-2024, InterDigital Communications, Inc
# All rights reserved.

# Redistribution and use in source and binary forms, with or without
# modification, are permitted (subject to the limitations in the disclaimer
# below) provided that the following conditions are met:

# * Redistributions of source code must retain the above copyright notice,
#   this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above copyright notice,
#   this list of conditions and the following disclaimer in the documentation
#   and/or other materials provided with the distribution.
# * Neither the name of InterDigital Communications, Inc nor the names of its
#   contributors may be used to endorse or promote products derived from this
#   software without specific prior written permission.

# NO EXPRESS OR IMPLIED LICENSES TO ANY PARTY'S PATENT RIGHTS ARE GRANTED BY
# THIS LICENSE. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
# CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT
# NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
# PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR
# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
# EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
# OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
# WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
# OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
# ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

import torch

from torch import Tensor


[docs] def compute_padding(in_h: int, in_w: int, *, out_h=None, out_w=None, min_div=1): """Returns tuples for padding and unpadding. Args: in_h: Input height. in_w: Input width. out_h: Output height. out_w: Output width. min_div: Length that output dimensions should be divisible by. """ if out_h is None: out_h = (in_h + min_div - 1) // min_div * min_div if out_w is None: out_w = (in_w + min_div - 1) // min_div * min_div if out_h % min_div != 0 or out_w % min_div != 0: raise ValueError( f"Padded output height and width are not divisible by min_div={min_div}." ) left = (out_w - in_w) // 2 right = out_w - in_w - left top = (out_h - in_h) // 2 bottom = out_h - in_h - top pad = (left, right, top, bottom) unpad = (-left, -right, -top, -bottom) return pad, unpad
[docs] def quantize_ste(x: Tensor) -> Tensor: """ Rounding with non-zero gradients. Gradients are approximated by replacing the derivative by the identity function. Used in `"Lossy Image Compression with Compressive Autoencoders" <https://arxiv.org/abs/1703.00395>`_ .. note:: Implemented with the pytorch `detach()` reparametrization trick: `x_round = x_round - x.detach() + x` """ return (torch.round(x) - x).detach() + x