From c3cd9482c3121c6d9197eb7fdf9d09533495b1fe Mon Sep 17 00:00:00 2001 From: yourName Date: Mon, 7 Sep 2020 14:06:09 +0800 Subject: [PATCH] add high level api doc --- .../high_level_api/high_level_api.ipynb | 623 ++++++++++++++++++ 1 file changed, 623 insertions(+) create mode 100644 paddle2.0_docs/high_level_api/high_level_api.ipynb diff --git a/paddle2.0_docs/high_level_api/high_level_api.ipynb b/paddle2.0_docs/high_level_api/high_level_api.ipynb new file mode 100644 index 00000000..6474cbc0 --- /dev/null +++ b/paddle2.0_docs/high_level_api/high_level_api.ipynb @@ -0,0 +1,623 @@ +{ + "metadata": { + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.7.4-final" + }, + "orig_nbformat": 2, + "kernelspec": { + "name": "python37464bitc4da1ac836094043840bff631bedbf7f", + "display_name": "Python 3.7.4 64-bit" + } + }, + "nbformat": 4, + "nbformat_minor": 2, + "cells": [ + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "# 飞桨高层API使用指南\n", + "\n", + "## 1. 简介\n", + "\n", + "飞桨2.0全新推出高层API,是对飞桨API的进一步封装与升级,提供了更加简洁易用的API,进一步提升了飞桨的易学易用性,并增强飞桨的功能。\n", + "\n", + "飞桨高层API面向从深度学习小白到资深开发者的所有人群,对于AI初学者来说,使用高层API可以简单快速的构建深度学习项目,对于资深开发者来说,可以快速完成算法迭代。\n", + "\n", + "飞桨高层API具有以下特点:\n", + "\n", + "* 易学易用: 高层API是对普通动态图API的进一步封装和优化,同时保持与普通API的兼容性,高层API使用更加易学易用,同样的实现使用高层API可以节省大量的代码。\n", + "* 低代码开发: 使用飞桨高层API的一个明显特点是,用户可编程代码量大大缩减。\n", + "* 动静转换: 高层API支持动静转换,用户只需要改一行代码即可实现将动态图代码在静态图模式下训练,既方便用户使用动态图调试模型,又提升了模型训练效率。\n", + "\n", + "在功能增强与使用方式上,高层API有以下升级:\n", + "\n", + "* 模型训练方式升级: 高层API中封装了Model类,继承了Model类的神经网络可以仅用几行代码完成模型的训练。\n", + "* 新增图像处理模块transform: 飞桨新增了图像预处理模块,其中包含数十种数据处理函数,基本涵盖了常用的数据处理、数据增强方法。\n", + "* 提供常用的神经网络模型可供调用: 高层API中集成了计算机视觉领域和自然语言处理领域常用模型,包括但不限于mobilenet、resnet、yolov3、cyclegan、bert、transformer、seq2seq等等。同时发布了对应模型的预训练模型,用户可以直接使用这些模型或者在此基础上完成二次开发。\n" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 2. 安装并使用飞桨高层API\n", + "\n", + "飞桨高层API无需独立安装,只需要安装好paddlepaddle即可,安装完成后import paddle即可使用相关高层API,如:paddle.Model、视觉领域paddle.vision、NLP领域paddle.text。" + ] + }, + { + "cell_type": "code", + "execution_count": 4, + "metadata": {}, + "outputs": [ + { + "output_type": "execute_result", + "data": { + "text/plain": "'0.0.0'" + }, + "metadata": {}, + "execution_count": 4 + } + ], + "source": [ + "import paddle\n", + "import paddle.vision as vision\n", + "import paddle.text as text\n", + "\n", + "paddle.__version__" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 2. 目录\n", + "\n", + "本指南教学内容覆盖\n", + "\n", + "* 使用高层API提供的自带数据集进行相关深度学习任务训练。\n", + "* 使用自定义数据进行数据集的定义、数据预处理和训练。\n", + "* 如何在数据集定义和加载中应用数据增强相关接口。\n", + "* 如何进行模型的组网。\n", + "* 高层API进行模型训练的相关API使用。\n", + "* 如何在fit接口满足需求的时候进行自定义,使用基础API来完成训练。\n", + "* 如何使用多卡来加速训练。\n", + "\n", + "其他端到端的示例教程:\n", + "* TBD" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 3. 数据集定义、加载和数据预处理\n", + "\n", + "对于深度学习任务,均是框架针对各种类型数字的计算,是无法直接使用原始图片和文本等文件来完成。那么就是涉及到了一项动作,就是将原始的各种数据文件进行处理加工,转换成深度学习任务可以使用的数据。\n", + "\n", + "### 3.1 框架自带数据集使用\n", + "\n", + "高层API将一些我们常用到的数据集作为领域API对用户进行开放,对应API所在目录为`paddle.vision.datasets`,那么我们先看下提供了哪些数据集。" + ] + }, + { + "cell_type": "code", + "execution_count": 17, + "metadata": { + "tags": [] + }, + "outputs": [ + { + "output_type": "execute_result", + "data": { + "text/plain": "['DatasetFolder',\n 'ImageFolder',\n 'MNIST',\n 'Flowers',\n 'Cifar10',\n 'Cifar100',\n 'VOC2012']" + }, + "metadata": {}, + "execution_count": 17 + } + ], + "source": [ + "paddle.vision.datasets.__all__" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "这里我们是加载一个手写数字识别的数据集,用`mode`来标识是训练数据还是测试数据集。数据集接口会自动从远端下载数据集到本机缓存目录`~/.cache/paddle/dataset`。" + ] + }, + { + "cell_type": "code", + "execution_count": 24, + "metadata": {}, + "outputs": [], + "source": [ + "# 测试数据集\n", + "train_dataset = vision.datasets.MNIST(mode='train')\n", + "\n", + "# 验证数据集\n", + "val_dataset = vision.datasets.MNIST(mode='test')" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 3.2 自定义数据集\n", + "\n", + "更多的时候我们是需要自己使用已有的相关数据来定义数据集,那么这里我们通过一个案例来了解如何进行数据集的定义,飞桨为用户提供了`paddle.io.Dataset`基类,让用户通过类的集成来快速实现数据集定义。" + ] + }, + { + "cell_type": "code", + "execution_count": 23, + "metadata": { + "tags": [] + }, + "outputs": [ + { + "output_type": "stream", + "name": "stdout", + "text": "=============train dataset=============\ntraindata1 label1\ntraindata2 label2\ntraindata3 label3\ntraindata4 label4\n=============evaluation dataset=============\ntestdata1 label1\ntestdata2 label2\ntestdata3 label3\ntestdata4 label4\n" + } + ], + "source": [ + "from paddle.io import Dataset\n", + "\n", + "\n", + "class MyDataset(Dataset):\n", + " \"\"\"\n", + " 步骤一:继承paddle.io.Dataset类\n", + " \"\"\"\n", + " def __init__(self, mode='train'):\n", + " \"\"\"\n", + " 步骤二:实现构造函数,定义数据读取方式,划分训练和测试数据集\n", + " \"\"\"\n", + " super(MyDataset, self).__init__()\n", + "\n", + " if mode == 'train':\n", + " self.data = [\n", + " ['traindata1', 'label1'],\n", + " ['traindata2', 'label2'],\n", + " ['traindata3', 'label3'],\n", + " ['traindata4', 'label4'],\n", + " ]\n", + " else:\n", + " self.data = [\n", + " ['testdata1', 'label1'],\n", + " ['testdata2', 'label2'],\n", + " ['testdata3', 'label3'],\n", + " ['testdata4', 'label4'],\n", + " ]\n", + " \n", + " def __getitem__(self, index):\n", + " \"\"\"\n", + " 步骤三:实现__getitem__方法,定义指定index时如何获取数据,并返回单条数据(训练数据,对应的标签)\n", + " \"\"\"\n", + " data = self.data[index][0]\n", + " label = self.data[index][1]\n", + "\n", + " return data, label\n", + "\n", + " def __len__(self):\n", + " \"\"\"\n", + " 步骤四:实现__len__方法,返回数据集总数目\n", + " \"\"\"\n", + " return len(self.data)\n", + "\n", + "# 测试定义的数据集\n", + "train_dataset = MyDataset(mode='train')\n", + "val_dataset = MyDataset(mode='test')\n", + "\n", + "print('=============train dataset=============')\n", + "for data, label in train_dataset:\n", + " print(data, label)\n", + "\n", + "print('=============evaluation dataset=============')\n", + "for data, label in val_dataset:\n", + " print(data, label)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 3.3 数据增强\n", + "\n", + "训练过程中有时会遇到过拟合的问题,其中一个解决方法就是对训练数据做增强,对数据进行处理得到不同的图像,从而泛化数据集。数据增强API是定义在领域目录的transofrms下,这里我们介绍两种使用方式,一种是基于框架自带数据集,一种是基于自己定义的数据集。\n", + "\n", + "#### 3.3.1 框架自带数据集" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "from paddle.vision.transforms import Compose, Resize, ColorJitter\n", + "\n", + "\n", + "# 定义想要使用那些数据增强方式,这里用到了随机调整亮度、对比度和饱和度,改变图片大小\n", + "transform = Compose([ColorJitter(), Resize(size=100)])\n", + "\n", + "# 通过transform参数传递定义好的数据增项方法即可完成对自带数据集的应用\n", + "train_dataset = vision.datasets.MNIST(mode='train', transform=transform)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "#### 3.3.2 自定义数据集\n", + "\n", + "针对自定义数据集使用数据增强有两种方式,一种是在数据集的构造函数中进行数据增强方法的定义,之后对__getitem__中返回的数据进行应用。另外一种方式也可以给自定义的数据集类暴漏一个构造参数,在实例化类的时候将数据增强方法传递进去。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "from paddle.io import Dataset\n", + "\n", + "\n", + "class MyDataset(Dataset):\n", + " def __init__(self, mode='train'):\n", + " super(MyDataset, self).__init__()\n", + "\n", + " if mode == 'train':\n", + " self.data = [\n", + " ['traindata1', 'label1'],\n", + " ['traindata2', 'label2'],\n", + " ['traindata3', 'label3'],\n", + " ['traindata4', 'label4'],\n", + " ]\n", + " else:\n", + " self.data = [\n", + " ['testdata1', 'label1'],\n", + " ['testdata2', 'label2'],\n", + " ['testdata3', 'label3'],\n", + " ['testdata4', 'label4'],\n", + " ]\n", + "\n", + " # 定义要使用的数据预处理方法,针对图片的操作\n", + " self.transform = Compose([ColorJitter(), Resize(size=100)])\n", + " \n", + " def __getitem__(self, index):\n", + " data = self.data[index][0]\n", + "\n", + " # 在这里对训练数据进行应用\n", + " # 这里只是一个示例,测试时需要将数据集更换为图片数据进行测试\n", + " data = self.transform(data)\n", + "\n", + " label = self.data[index][1]\n", + "\n", + " return data, label\n", + "\n", + " def __len__(self):\n", + " return len(self.data)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 4. 模型组网\n", + "\n", + "针对高层API在模型组网上和基础API是统一的一套,无需投入额外的学习使用成本。那么这里我举几个简单的例子来做示例。\n", + "\n", + "### 4.1 Sequential组网\n", + "\n", + "针对顺序的线性网络结构我们可以直接使用Sequential来快速完成组网,可以减少类的定义等代码编写。" + ] + }, + { + "cell_type": "code", + "execution_count": 28, + "metadata": {}, + "outputs": [], + "source": [ + "# Sequential形式组网\n", + "mnist = paddle.nn.Sequential(\n", + " paddle.nn.Flatten(),\n", + " paddle.nn.Linear(784, 512),\n", + " paddle.nn.ReLU(),\n", + " paddle.nn.Dropout(0.2),\n", + " paddle.nn.Linear(512, 10)\n", + ")" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 4.2 SubClass组网\n", + "针对一些比较复杂的网络结构,就可以使用Layer子类定义的方式来进行模型代码编写,在`__init__`构造函数中进行组网Layer的声明,在`forward`中使用声明的Layer变量进行前向计算。子类组网方式也可以实现sublayer的复用,针对相同的layer可以在构造函数中一次性定义,在forward中多次调用。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# Layer类继承方式组网\n", + "class Mnist(paddle.nn.Layer):\n", + " def __init__(self):\n", + " super(Mnist, self).__init__()\n", + "\n", + " self.flatten = paddle.nn.Flatten()\n", + " self.linear_1 = paddle.nn.Linear(784, 512)\n", + " self.linear_2 = paddle.nn.Linear(512, 10)\n", + " self.relu = paddle.nn.ReLU()\n", + " self.dropout = paddle.nn.Dropout(0.2)\n", + "\n", + " def forward(self, inputs):\n", + " y = self.flatten(inputs)\n", + " y = self.linear_1(y)\n", + " y = self.relu(y)\n", + " y = self.dropout(y)\n", + " y = self.linear_2(y)\n", + "\n", + " return y\n", + "\n", + "mnist = Mnist()" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 4.3 模型封装\n", + "\n", + "定义好网络结构之后我们来使用`paddle.Model`完成模型的封装,将网络结构组合成一个可快速使用高层API进行训练、评估和预测的类。\n", + "\n", + "在封装的时候我们有两种场景,动态图训练模式和静态图训练模式。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# 场景1:动态图模式\n", + "\n", + "# 启动动态图训练模式\n", + "paddle.disable_static()\n", + "# 使用GPU训练\n", + "paddle.set_device('gpu')\n", + "# 模型封装\n", + "model = paddle.Model(mnist)\n", + "\n", + "\n", + "# 场景2:静态图模式\n", + "\n", + "# input = paddle.static.InputSpec([None, 1, 28, 28], dtype='float32')\n", + "# label = paddle.static.InputSpec([None, 1], dtype='int8')\n", + "# model = paddle.Model(mnist, input, label)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 4.4 模型可视化\n", + "\n", + "在组建好我们的网络结构后,一般我们会想去对我们的网络结构进行一下可视化,逐层的去对齐一下我们的网络结构参数,看看是否符合我们的预期。这里可以通过`Model.summary`接口进行可视化展示。\n" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "model.summary((1, 28, 28))" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "另外,summary接口有两种使用方式,下面我们通过两个示例来做展示,除了`Model.summary`这种配套`paddle.Model`封装使用的接口外,还有一套配合没有经过`paddle.Model`封装的方式来使用。可以直接将实例化好的Layer子类放到`paddle.summary`接口中进行可视化呈现。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "paddle.summary(mnist, (1, 28, 28))" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "这里面有一个注意的点,有的用户可能会疑惑为什么要传递`(1, 28, 28)`这个input_size参数,因为在动态图中,网络定义阶段是还没有得到输入数据的形状信息,我们想要做网络结构的呈现就无从下手,那么我们通过告知接口网络结构的输入数据形状,这样网络可以通过逐层的计算推导得到完整的网络结构信息进行呈现。如果是动态图运行模式,那么就不需要给summary接口传递输入数据形状这个值了,因为在Model封装的时候我们已经定义好了InputSpec,其中包含了输入数据的形状格式。" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 5. 模型训练\n", + "\n", + "使用`paddle.Model`封装成模型类后进行训练非常的简洁方便,我们可以直接通过调用`Model.fit`就可以完成训练过程。\n", + "\n", + "在使用`Model.fit`接口启动训练前,我们先通过`Model.prepare`接口来对训练进行提前的配置准备工作,包括设置模型优化器,Loss计算方法,精度计算方法等。\n", + "\n" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# 为模型训练做准备,设置优化器,损失函数和精度计算方式\n", + "model.prepare(paddle.optimizer.Adam(parameters=model.parameters()), \n", + " paddle.nn.CrossEntropyLoss(),\n", + " paddle.metric.Accuracy())" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "做好模型训练的前期准备工作后,我们正式调用`fit()`接口来启动训练过程,需要指定一下至少3个关键参数:训练数据集,训练轮次和单次训练数据批次大小。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# 启动模型训练,指定训练数据集,设置训练轮次,设置每次数据集计算的批次大小,设置日志格式\n", + "model.fit(train_dataset, \n", + " epochs=10, \n", + " batch_size=32,\n", + " verbose=1)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 5.1 单机单卡\n", + "\n", + "我们把刚才单步教学的训练代码做一个整合,这个完整的代码示例就是我们的单机单卡训练程序。" + ] + }, + { + "cell_type": "code", + "execution_count": 30, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "# 启动动态图训练模式\n", + "paddle.disable_static()\n", + "\n", + "# 使用GPU训练\n", + "paddle.set_device('gpu')\n", + "\n", + "# 构建模型训练用的Model,告知需要训练哪个模型\n", + "model = paddle.Model(mnist)\n", + "\n", + "# 为模型训练做准备,设置优化器,损失函数和精度计算方式\n", + "model.prepare(paddle.optimizer.Adam(parameters=model.parameters()), \n", + " paddle.nn.CrossEntropyLoss(),\n", + " paddle.metric.Accuracy())\n", + "\n", + "# 启动模型训练,指定训练数据集,设置训练轮次,设置每次数据集计算的批次大小,设置日志格式\n", + "model.fit(train_dataset, \n", + " epochs=10, \n", + " batch_size=32,\n", + " verbose=1)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 5.2 单机多卡\n", + "\n", + "对于高层API来实现单机多卡非常简单,整个训练代码和单机单卡没有差异。直接使用`paddle.distributed.launch`启动单机单卡的程序即可。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# train.py里面包含的就是单机单卡代码\n", + "python -m paddle.distributed.launch train.py" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 6. 模型评估\n", + "\n", + "对于训练好的模型进行评估操作可以使用`evaluate`接口来实现。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "result = model.evaluate(val_dataset, verbose=1)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 7. 模型预测\n", + "\n", + "高层API中提供`predict`接口,支持用户使用测试数据来完成模型的预测。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "pred_result = model.predict(val_dataset)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## 8. 模型部署\n", + "\n", + "### 8.1 模型存储\n", + "\n", + "模型训练和验证达到我们的预期后,可以使用`save`接口来将我们的模型保存下来,用于后续模型的Fine-tuning或推理部署。" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# 保存用于推理部署的模型(training=False)\n", + "model.save('~/model/mnist', training=False)" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### 8.2 预测部署\n", + "\n", + "有了用于推理部署的模型,就可以使用推理部署框架来完成预测服务部署,具体可以参见:[预测部署](https://www.paddlepaddle.org.cn/documentation/docs/zh/advanced_guide/inference_deployment/index_cn.html), 包括服务端部署、移动端部署和模型压缩。" + ] + } + ] +} \ No newline at end of file