SIGN IN SIGN UP
PaddlePaddle / Paddle UNCLAIMED

PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)

# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import annotations
import os
import numpy as np
import paddle
from paddle.base.data_feeder import check_type, convert_dtype
from ..framework import core
__all__ = []
class PrintOptions:
precision = 8
threshold = 1000
edgeitems = 3
linewidth = 80
sci_mode = False
DEFAULT_PRINT_OPTIONS = PrintOptions()
def set_printoptions(
precision: int | None = None,
threshold: int | None = None,
edgeitems: int | None = None,
sci_mode: bool | None = None,
linewidth: int | None = None,
) -> None:
"""Set the printing options for Tensor.
Args:
precision (int|None, optional): Number of digits of the floating number, default 8.
threshold (int|None, optional): Total number of elements printed, default 1000.
edgeitems (int|None, optional): Number of elements in summary at the beginning and ending of each dimension, default 3.
sci_mode (bool|None, optional): Format the floating number with scientific notation or not, default False.
linewidth (int|None, optional): Number of characters each line, default 80.
Returns:
None.
Examples:
.. code-block:: pycon
>>> import paddle
>>> paddle.seed(10)
>>> a = paddle.rand([10, 20])
>>> paddle.set_printoptions(4, 100, 3)
>>> print(a)
Tensor(shape=[10, 20], dtype=float32, place=Place(cpu), stop_gradient=True,
[[0.2727, 0.5489, 0.8655, ..., 0.2916, 0.8525, 0.9000],
[0.3806, 0.8996, 0.0928, ..., 0.9535, 0.8378, 0.6409],
[0.1484, 0.4038, 0.8294, ..., 0.0148, 0.6520, 0.4250],
...,
[0.3426, 0.1909, 0.7240, ..., 0.4218, 0.2676, 0.5679],
[0.5561, 0.2081, 0.0676, ..., 0.9778, 0.3302, 0.9559],
[0.2665, 0.8483, 0.5389, ..., 0.4956, 0.6862, 0.9178]])
"""
kwargs = {}
if precision is not None:
check_type(precision, 'precision', (int), 'set_printoptions')
DEFAULT_PRINT_OPTIONS.precision = precision
kwargs['precision'] = precision
if threshold is not None:
check_type(threshold, 'threshold', (int), 'set_printoptions')
DEFAULT_PRINT_OPTIONS.threshold = threshold
kwargs['threshold'] = threshold
if edgeitems is not None:
check_type(edgeitems, 'edgeitems', (int), 'set_printoptions')
DEFAULT_PRINT_OPTIONS.edgeitems = edgeitems
kwargs['edgeitems'] = edgeitems
if linewidth is not None:
check_type(linewidth, 'linewidth', (int), 'set_printoptions')
DEFAULT_PRINT_OPTIONS.linewidth = linewidth
kwargs['linewidth'] = linewidth
if sci_mode is not None:
check_type(sci_mode, 'sci_mode', (bool), 'set_printoptions')
DEFAULT_PRINT_OPTIONS.sci_mode = sci_mode
kwargs['sci_mode'] = sci_mode
core.set_printoptions(**kwargs)
def _to_summary(var):
edgeitems = DEFAULT_PRINT_OPTIONS.edgeitems
# Handle tensor of shape contains 0, like [0, 2], [3, 0, 3]
if np.prod(var.shape) == 0:
return np.array([])
if len(var.shape) == 0:
return var
elif len(var.shape) == 1:
if var.shape[0] > 2 * edgeitems:
return np.concatenate([var[:edgeitems], var[(-1 * edgeitems) :]])
else:
return var
else:
# recursively handle all dimensions
if var.shape[0] > 2 * edgeitems:
begin = list(var[:edgeitems])
end = list(var[(-1 * edgeitems) :])
return np.stack([_to_summary(x) for x in (begin + end)])
else:
return np.stack([_to_summary(x) for x in var])
def _format_item(np_var, max_width=0, signed=False):
if (
np_var.dtype == np.float32
or np_var.dtype == np.float64
or np_var.dtype == np.float16
):
if DEFAULT_PRINT_OPTIONS.sci_mode:
item_str = f'{np_var:.{DEFAULT_PRINT_OPTIONS.precision}e}'
elif np.ceil(np_var) == np_var:
item_str = f'{np_var:.0f}.'
else:
item_str = f'{np_var:.{DEFAULT_PRINT_OPTIONS.precision}f}'
elif np_var.dtype == np.complex64 or np_var.dtype == np.complex128:
re = np.real(np_var)
im = np.imag(np_var)
prec = DEFAULT_PRINT_OPTIONS.precision
if DEFAULT_PRINT_OPTIONS.sci_mode:
if im >= 0:
item_str = f'({re:.{prec}e}+{im:.{prec}e}j)'
else:
item_str = f'({re:.{prec}e}{im:.{prec}e}j)'
else:
if im >= 0:
item_str = f'({re:.{prec}f}+{im:.{prec}f}j)'
else:
item_str = f'({re:.{prec}f}{im:.{prec}f}j)'
else:
item_str = f'{np_var}'
if max_width > len(item_str):
2023-12-18 15:39:14 +08:00
if signed: # handle sign character for tensor with negative item
if np_var < 0:
return item_str.ljust(max_width)
else:
return ' ' + item_str.ljust(max_width - 1)
else:
return item_str.ljust(max_width)
else: # used for _get_max_width
return item_str
def _get_max_width(var):
# return max_width for a scalar
max_width = 0
signed = False
for item in list(var.flatten()):
if (not signed) and (item < 0):
signed = True
item_str = _format_item(item)
max_width = max(max_width, len(item_str))
return max_width, signed
def _format_tensor(var, summary, indent=0, max_width=0, signed=False):
"""
Format a tensor
Args:
var(Tensor): The tensor to be formatted.
summary(bool): Do summary or not. If true, some elements will not be printed, and be replaced with "...".
indent(int): The indent of each line.
max_width(int): The max width of each elements in var.
signed(bool): Print +/- or not.
"""
edgeitems = DEFAULT_PRINT_OPTIONS.edgeitems
linewidth = DEFAULT_PRINT_OPTIONS.linewidth
if len(var.shape) == 0:
# 0-D Tensor, whose shape = [], should be formatted like this.
return _format_item(var, max_width, signed)
elif len(var.shape) == 1:
item_length = max_width + 2
items_per_line = max(1, (linewidth - indent) // item_length)
if summary and var.shape[0] > 2 * edgeitems:
items = (
[
_format_item(var[i], max_width, signed)
for i in range(edgeitems)
]
+ ['...']
+ [
_format_item(var[i], max_width, signed)
for i in range(var.shape[0] - edgeitems, var.shape[0])
]
)
else:
items = [
_format_item(var[i], max_width, signed)
for i in range(var.shape[0])
]
lines = [
items[i : i + items_per_line]
for i in range(0, len(items), items_per_line)
]
s = (',\n' + ' ' * (indent + 1)).join(
[', '.join(line) for line in lines]
)
return '[' + s + ']'
else:
# recursively handle all dimensions
if summary and var.shape[0] > 2 * edgeitems:
vars = (
[
_format_tensor(
var[i], summary, indent + 1, max_width, signed
)
for i in range(edgeitems)
]
+ ['...']
+ [
_format_tensor(
var[i], summary, indent + 1, max_width, signed
)
for i in range(var.shape[0] - edgeitems, var.shape[0])
]
)
else:
vars = [
_format_tensor(var[i], summary, indent + 1, max_width, signed)
for i in range(var.shape[0])
]
s = (',' + '\n' * (len(var.shape) - 1) + ' ' * (indent + 1)).join(vars)
return '[' + s + ']'
def to_string(var, prefix='Tensor'):
indent = len(prefix) + 1
dtype = convert_dtype(var.dtype)
if var.dtype == paddle.bfloat16:
dtype = 'bfloat16'
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient},\n{indent}{data})"
tensor = var.value().get_tensor()
if not tensor._is_initialized():
return "Tensor(Not initialized)"
if var.dtype == paddle.bfloat16:
if not var.place.is_cpu_place():
paddle.device.synchronize()
var = var.astype('float32')
np_var = var.numpy(False)
if len(var.shape) == 0:
size = 0
else:
size = 1
for dim in var.shape:
size *= dim
summary = False
if size > DEFAULT_PRINT_OPTIONS.threshold:
summary = True
max_width, signed = _get_max_width(_to_summary(np_var))
data = _format_tensor(
np_var, summary, indent=indent, max_width=max_width, signed=signed
)
return _template.format(
prefix=prefix,
shape=var.shape,
dtype=dtype,
place=var._place_str,
stop_gradient=var.stop_gradient,
indent=' ' * indent,
data=data,
)
def mask_xpu_bf16_tensor(np_tensor):
# For XPU, we mask out the 0x8000 added to the tail when converting bf16 to fp32.
mask = np.array(0xFFFF0000, dtype='uint32')
return (np_tensor.view('uint32') & mask).view('float32')
def _format_dense_tensor(tensor, indent):
dtype = tensor.dtype
2025-09-09 14:33:58 +08:00
if dtype in {
paddle.bfloat16,
paddle.float8_e4m3fn,
paddle.float8_e5m2,
}:
if not tensor.place.is_cpu_place():
paddle.device.synchronize()
tensor = tensor.astype('float32')
# TODO(zhouwei): will remove 0-D Tensor.numpy() hack
np_tensor = tensor.numpy(False)
if (
paddle.is_compiled_with_xpu()
and os.getenv("XPU_PADDLE_MASK_BF16_PRINT") is not None
and (dtype == paddle.bfloat16 or dtype == core.VarDesc.VarType.BF16)
):
np_tensor = mask_xpu_bf16_tensor(np_tensor)
summary = (
np.prod(tensor.shape, dtype="int64") > DEFAULT_PRINT_OPTIONS.threshold
)
max_width, signed = _get_max_width(_to_summary(np_tensor))
data = _format_tensor(
np_tensor, summary, indent=indent, max_width=max_width, signed=signed
)
return data
2024-12-09 17:35:04 +08:00
def selected_rows_tensor_to_string(tensor, dtype, prefix='Tensor'):
indent = len(prefix) + 1
if tensor.is_selected_rows():
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient}, rows={rows},\n{indent}{data})"
data = _format_dense_tensor(tensor, indent)
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
2024-12-09 17:35:04 +08:00
dtype=dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
indent=' ' * indent,
data=data,
rows=tensor.rows(),
)
def sparse_tensor_to_string(tensor, prefix='Tensor'):
indent = len(prefix) + 1
if tensor.is_sparse_coo():
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient}, \n{indent}{indices}, \n{indent}{values})"
indices_tensor = tensor.indices()
values_tensor = tensor.values()
indices_data = 'indices=' + _format_dense_tensor(
indices_tensor, indent + len('indices=')
)
values_data = 'values=' + _format_dense_tensor(
values_tensor, indent + len('values=')
)
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
dtype=tensor.dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
indent=' ' * indent,
indices=indices_data,
values=values_data,
)
else:
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient}, \n{indent}{crows}, \n{indent}{cols}, \n{indent}{values})"
crows_tensor = tensor.crows()
cols_tensor = tensor.cols()
elements_tensor = tensor.values()
crows_data = 'crows=' + _format_dense_tensor(
crows_tensor, indent + len('crows=')
)
cols_data = 'cols=' + _format_dense_tensor(
cols_tensor, indent + len('cols=')
)
values_data = 'values=' + _format_dense_tensor(
elements_tensor, indent + len('values=')
)
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
dtype=tensor.dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
indent=' ' * indent,
crows=crows_data,
cols=cols_data,
values=values_data,
)
def dist_tensor_to_string(tensor, prefix='Tensor'):
# TODO(dev): Complete tensor will be printed after reshard
# is ready.
indent = len(prefix) + 1
dtype = convert_dtype(tensor.dtype)
if tensor.dtype == paddle.bfloat16:
dtype = 'bfloat16'
if not tensor._is_dense_tensor_hold_allocation():
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient}, process_mesh={process_mesh}, placements={placements}, GlobalDenseTensor Not initialized)"
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
dtype=dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
process_mesh=tensor.process_mesh,
placements=tensor._placements_str,
)
else:
indent = len(prefix) + 1
# If we print a dist_tensor with bf16 dtype and Partial placement, it is essential to ensure that the AllReduce communication
# is performed in bf16. After completing the communication, convert it to fp32, and then convert it into a numpy array.
from paddle.distributed import Replicate, reshard
placements = [Replicate() for _ in range(tensor.process_mesh.ndim)]
global_tensor = reshard(tensor, tensor.process_mesh, placements)
data = _format_dense_tensor(global_tensor, indent)
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient}, process_mesh={process_mesh}, placements={placements}, GlobalDenseTensor=\n{indent}{data})"
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
dtype=dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
process_mesh=tensor.process_mesh,
placements=tensor._placements_str,
indent=' ' * indent,
data=data,
)
def tensor_to_string(tensor, prefix='Tensor'):
indent = len(prefix) + 1
dtype = convert_dtype(tensor.dtype)
if tensor.dtype == paddle.bfloat16:
dtype = 'bfloat16'
_template = "{prefix}(shape={shape}, dtype={dtype}, place={place}, stop_gradient={stop_gradient},\n{indent}{data})"
if tensor.is_sparse():
return sparse_tensor_to_string(tensor, prefix)
2024-12-09 17:35:04 +08:00
if tensor.is_selected_rows():
return selected_rows_tensor_to_string(tensor, dtype, prefix)
if tensor.is_dist():
return dist_tensor_to_string(tensor, prefix)
if not tensor._is_dense_tensor_hold_allocation():
return "Tensor(Not initialized)"
else:
data = _format_dense_tensor(tensor, indent)
return _template.format(
prefix=prefix,
shape=list(tensor.shape),
dtype=dtype,
place=tensor._place_str,
stop_gradient=tensor.stop_gradient,
indent=' ' * indent,
data=data,
)