Browse Source

modify doc

pull/431/head
xumengjuan1 2 years ago
parent
commit
fef203141e
2 changed files with 2 additions and 2 deletions
  1. +1
    -1
      examples/natural_robustness/ocr_evaluate/cnn_ctc/README.md
  2. +1
    -1
      examples/natural_robustness/ocr_evaluate/cnn_ctc/README_CN.md

+ 1
- 1
examples/natural_robustness/ocr_evaluate/cnn_ctc/README.md View File

@@ -94,7 +94,7 @@ This takes around 75 minutes.

## Mixed Precision

The [mixed precision](https://www.mindspore.cn/tutorials/experts/en/master/others/mixed_precision.html) training method accelerates the deep learning neural network training process by using both the single-precision and half-precision data formats, and maintains the network precision achieved by the single-precision training at the same time. Mixed precision training can accelerate the computation process, reduce memory usage, and enable a larger model or batch size to be trained on specific hardware.
The [mixed precision](https://mindspore.cn/tutorials/en/master/advanced/mixed_precision.html) training method accelerates the deep learning neural network training process by using both the single-precision and half-precision data formats, and maintains the network precision achieved by the single-precision training at the same time. Mixed precision training can accelerate the computation process, reduce memory usage, and enable a larger model or batch size to be trained on specific hardware.
For FP16 operators, if the input data type is FP32, the backend of MindSpore will automatically handle it with reduced precision. Users could check the reduced-precision operators by enabling INFO log and then searching ‘reduce precision’.

# [Environment Requirements](#contents)


+ 1
- 1
examples/natural_robustness/ocr_evaluate/cnn_ctc/README_CN.md View File

@@ -95,7 +95,7 @@ python src/preprocess_dataset.py

## 混合精度

采用[混合精度](https://www.mindspore.cn/tutorials/experts/zh-CN/master/others/mixed_precision.html)的训练方法使用支持单精度和半精度数据来提高深度学习神经网络的训练速度,同时保持单精度训练所能达到的网络精度。混合精度训练提高计算速度、减少内存使用的同时,支持在特定硬件上训练更大的模型或实现更大批次的训练。
采用[混合精度](https://mindspore.cn/tutorials/zh-CN/master/advanced/mixed_precision.html)的训练方法使用支持单精度和半精度数据来提高深度学习神经网络的训练速度,同时保持单精度训练所能达到的网络精度。混合精度训练提高计算速度、减少内存使用的同时,支持在特定硬件上训练更大的模型或实现更大批次的训练。
以FP16算子为例,如果输入数据类型为FP32,MindSpore后台会自动降低精度来处理数据。用户可打开INFO日志,搜索“reduce precision”查看精度降低的算子。

# 环境要求


Loading…
Cancel
Save