René's URL Explorer Experiment


Title: 一文归纳Ai数据增强之法 · Issue #13 · aialgorithm/Blog · GitHub

Open Graph Title: 一文归纳Ai数据增强之法 · Issue #13 · aialgorithm/Blog

X Title: 一文归纳Ai数据增强之法 · Issue #13 · aialgorithm/Blog

Description: 数据、算法、算力是人工智能发展的三要素。数据决定了Ai模型学习的上限,数据规模越大、质量越高,模型就能够拥有更好的泛化能力。然而在实际工程中,训练的数据相对模型而言数据量太少,或者很难覆盖全部的场景等问题,解决这问题的一个有效途径是通过数据增强(Data Augmentation),使模型获得较好的泛化性能。 1 数据增强介绍 数据增强(Data Augmentation)是在不实质性的增加数据的情况下,从原始数据加工出更多的表示,提高原数据的数量及质量,以接近于更多数...

Open Graph Description: 数据、算法、算力是人工智能发展的三要素。数据决定了Ai模型学习的上限,数据规模越大、质量越高,模型就能够拥有更好的泛化能力。然而在实际工程中,训练的数据相对模型而言数据量太少,或者很难覆盖全部的场景等问题,解决这问题的一个有效途径是通过数据增强(Data Augmentation),使模型获得较好的泛化性能。 1 数据增强介绍 数据增强(Data Augmentation)是在不实质性的增加...

X Description: 数据、算法、算力是人工智能发展的三要素。数据决定了Ai模型学习的上限,数据规模越大、质量越高,模型就能够拥有更好的泛化能力。然而在实际工程中,训练的数据相对模型而言数据量太少,或者很难覆盖全部的场景等问题,解决这问题的一个有效途径是通过数据增强(Data Augmentation),使模型获得较好的泛化性能。 1 数据增强介绍 数据增强(Data Augmentation)是在不实质性的增加...

Opengraph URL: https://github.com/aialgorithm/Blog/issues/13

X: @github

direct link

Domain: github.com


Hey, it has json ld scripts:
{"@context":"https://schema.org","@type":"DiscussionForumPosting","headline":"一文归纳Ai数据增强之法","articleBody":"数据、算法、算力是人工智能发展的三要素。数据决定了Ai模型学习的上限,数据规模越大、质量越高,模型就能够拥有更好的泛化能力。然而在实际工程中,训练的数据相对模型而言数据量太少,或者很难覆盖全部的场景等问题,解决这问题的一个有效途径是通过数据增强(Data Augmentation),使模型获得较好的泛化性能。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-f5a8a09a31f82304.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n## 1 数据增强介绍\r\n数据增强(Data Augmentation)是在不实质性的增加数据的情况下,从原始数据加工出更多的表示,提高原数据的数量及质量,以接近于更多数据量产生的价值。其原理是,通过对原始数据融入先验知识,加工出更多数据的表示,有助于模型判别数据中统计噪声,减少模型过拟合。\r\n\r\n\u003e如经典的机器学习例子--哈士奇误分类为狼:\r\n通过可解释性方法,可发现错误分类是由于图像上的雪造成的。通常狗对比狼的图像里面雪地背景比较少,分类器学会使用雪作为一个特征来将图像分类为狼还是狗,而忽略了动物本体的特征。此时,可以通过数据增强的方法,增加变换后的数据(如背景换色、加入噪声等方式)来训练模型,帮助模型学习到本体的特征,提高泛化能力。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-4c57d961120b2afb.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n需要关注的是,数据增强样本也有可能是引入片面噪声,导致过拟合。此时需要考虑的是调整数据增强方法,或者通过算法(可借鉴Pu-Learning思路)选择增强数据的最佳子集,以提高模型的泛化能力。\r\n\r\n常用数据增强方法可分为:基于样本变换的数据增强及基于深度学习的数据增强。\r\n\r\n\r\n## 2  基于样本变换的数据增强\r\n样本变换数据增强即采用预设的数据变换规则进行已有数据的扩增,包含单样本数据增强和多样本数据增强。\r\n### 2.1 单样本增强\r\n单(图像)样本增强主要有几何操作、颜色变换、随机擦除、添加噪声等方法,可参见imgaug开源库。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-d672060606dccc79.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n\r\n\r\n### 2.2 多样本数据增强方法   \r\n\r\n多样本增强是通过先验知识组合及转换多个样本,主要有Smote、SamplePairing、Mixup等方法在特征空间内构造已知样本的邻域值。\r\n\r\n- Smote(Synthetic Minority Over-sampling Technique)\r\n\r\nSmote方法较常用于样本均衡学习,核心思想是从训练集随机同类的两近邻样本合成一个新的样本,其方法可以分为三步:\r\n\r\n1、 对于各样本X_i,计算与同类样本的欧式距离,确定其同类的K个(如图3个)近邻样本;\r\n\r\n2、从该样本k近邻中随机选择一个样本如近邻X_ik,生成新的样本 \r\n```\r\nXsmote_ik =  Xi  +  rand(0,1) ∗ ∣X_i − X_ik∣\r\n```\r\n3、重复2步骤迭代N次,可以合成N个新的样本。\r\n\r\n![](https://upload-images.jianshu.io/upload_images/11682271-291fa2abd69481c8.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n```\r\n# SMOTE\r\nfrom imblearn.over_sampling import SMOTE\r\n\r\nprint(\"Before OverSampling, counts of label\\n{}\".format(y_train.value_counts()))\r\nsmote = SMOTE()\r\nx_train_res, y_train_res = smote.fit_resample(x_train, y_train)\r\nprint(\"After OverSampling, counts of label\\n{}\".format(y_train_res.value_counts()))\r\n```\r\n- SamplePairing\r\nSamplePairing算法的核心思想是从训练集随机抽取的两幅图像叠加合成一个新的样本(像素取平均值),使用第一幅图像的label作为合成图像的正确label。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-16a224115a4332c5.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n- Mixup\r\n\r\nMixup算法的核心思想是按一定的比例随机混合两个训练样本及其标签,这种混合方式不仅能够增加样本的多样性,且能够使决策边界更加平滑,增强了难例样本的识别,模型的鲁棒性得到提升。其方法可以分为两步:\r\n\r\n1、从原始训练数据中随机选取的两个样本(xi, yi) and (xj, yj)。其中y(原始label)用one-hot 编码。\r\n\r\n2、对两个样本按比例组合,形成新的样本和带权重的标签\r\n```\r\nx˜ = λxi + (1 − λ)xj\r\ny˜ = λyi + (1 − λ)yj\r\n```\r\n最终的loss为各标签上分别计算cross-entropy loss,加权求和。其中 λ ∈ [0, 1], λ是mixup的超参数,控制两个样本插值的强度。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-f508b2a041f7cc59.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n```python\r\n# Mixup\r\ndef mixup_batch(x, y, step, batch_size, alpha=0.2):\r\n    \"\"\"\r\n    get batch data\r\n    :param x: training data\r\n    :param y: one-hot label\r\n    :param step: step\r\n    :param batch_size: batch size\r\n    :param alpha: hyper-parameter α, default as 0.2\r\n    :return:  x y \r\n    \"\"\"\r\n    candidates_data, candidates_label = x, y\r\n    offset = (step * batch_size) % (candidates_data.shape[0] - batch_size)\r\n \r\n    # get batch data\r\n    train_features_batch = candidates_data[offset:(offset + batch_size)]\r\n    train_labels_batch = candidates_label[offset:(offset + batch_size)]\r\n\r\n    if alpha == 0:\r\n        return train_features_batch, train_labels_batch\r\n\r\n    if alpha \u003e 0:\r\n        weight = np.random.beta(alpha, alpha, batch_size)\r\n        x_weight = weight.reshape(batch_size, 1)\r\n        y_weight = weight.reshape(batch_size, 1)\r\n        index = np.random.permutation(batch_size)\r\n        x1, x2 = train_features_batch, train_features_batch[index]\r\n        x = x1 * x_weight + x2 * (1 - x_weight)\r\n        y1, y2 = train_labels_batch, train_labels_batch[index]\r\n        y = y1 * y_weight + y2 * (1 - y_weight)\r\n        return x, y\r\n```\r\n\r\n\r\n## 3  基于深度学习的数据增强\r\n### 3.1 特征空间的数据增强\r\n不同于传统在输入空间变换的数据增强方法,神经网络可将输入样本映射为网络层的低维向量(表征学习),从而直接在学习的特征空间进行组合变换等进行数据增强,如MoEx方法等。![](https://upload-images.jianshu.io/upload_images/11682271-f8ce094003492646.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n\r\n### 3.2 基于生成模型的数据增强\r\n\r\n生成模型如变分自编码网络(Variational Auto-Encoding network, VAE)和生成对抗网络(Generative Adversarial Network, GAN),其生成样本的方法也可以用于数据增强。这种基于网络合成的方法相比于传统的数据增强技术虽然过程更加复杂, 但是生成的样本更加多样。\r\n\r\n- 变分自编码器VAE\r\n变分自编码器(Variational Autoencoder,VAE)其基本思路是:将真实样本通过编码器网络变换成一个理想的数据分布,然后把数据分布再传递给解码器网络,构造出生成样本,模型训练学习的过程是使生成样本与真实样本足够接近。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-5622b54abf20887a.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n```\r\n# VAE模型\r\nclass VAE(keras.Model):\r\n    ...\r\n    def train_step(self, data):\r\n        with tf.GradientTape() as tape:\r\n            z_mean, z_log_var, z = self.encoder(data)\r\n            reconstruction = self.decoder(z)\r\n            reconstruction_loss = tf.reduce_mean(\r\n                tf.reduce_sum(\r\n                    keras.losses.binary_crossentropy(data, reconstruction), axis=(1, 2)\r\n                )\r\n            )\r\n            kl_loss = -0.5 * (1 + z_log_var - tf.square(z_mean) - tf.exp(z_log_var))\r\n            kl_loss = tf.reduce_mean(tf.reduce_sum(kl_loss, axis=1))\r\n            total_loss = reconstruction_loss + kl_loss\r\n        grads = tape.gradient(total_loss, self.trainable_weights)\r\n        self.optimizer.apply_gradients(zip(grads, self.trainable_weights))\r\n        self.total_loss_tracker.update_state(total_loss)\r\n        self.reconstruction_loss_tracker.update_state(reconstruction_loss)\r\n        self.kl_loss_tracker.update_state(kl_loss)\r\n        return {\r\n            \"loss\": self.total_loss_tracker.result(),\r\n            \"reconstruction_loss\": self.reconstruction_loss_tracker.result(),\r\n            \"kl_loss\": self.kl_loss_tracker.result(),\r\n        }\r\n```\r\n- 生成对抗网络GAN\r\n生成对抗网络GAN(Generative Adversarial Network) 由生成网络(Generator, *G*)和判别网络(Discriminator, *D*)两部分组成, 生成网络构成一个映射函数*G*: *Z*→*X*(输入噪声*z*, 输出生成的图像数据*x*), 判别网络判别输入是来自真实数据还是生成网络生成的数据。 \r\n![](https://upload-images.jianshu.io/upload_images/11682271-4268490d8556b5f0.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n```\r\n# DCGAN模型\r\n\r\nclass GAN(keras.Model):\r\n    ...\r\n    def train_step(self, real_images):\r\n        batch_size = tf.shape(real_images)[0]\r\n        random_latent_vectors = tf.random.normal(shape=(batch_size, self.latent_dim))\r\n        # G: Z→X(输入噪声z, 输出生成的图像数据x)\r\n        generated_images = self.generator(random_latent_vectors)\r\n        # 合并生成及真实的样本并赋判定的标签\r\n        combined_images = tf.concat([generated_images, real_images], axis=0)\r\n        labels = tf.concat(\r\n            [tf.ones((batch_size, 1)), tf.zeros((batch_size, 1))], axis=0\r\n        )\r\n        # 标签加入随机噪声\r\n        labels += 0.05 * tf.random.uniform(tf.shape(labels))\r\n        # 训练判定网络\r\n        with tf.GradientTape() as tape:\r\n            predictions = self.discriminator(combined_images)\r\n            d_loss = self.loss_fn(labels, predictions)\r\n        grads = tape.gradient(d_loss, self.discriminator.trainable_weights)\r\n        self.d_optimizer.apply_gradients(\r\n            zip(grads, self.discriminator.trainable_weights)\r\n        )\r\n        \r\n        random_latent_vectors = tf.random.normal(shape=(batch_size, self.latent_dim))\r\n        # 赋生成网络样本的标签(都赋为真实样本)\r\n        misleading_labels = tf.zeros((batch_size, 1))\r\n        # 训练生成网络\r\n        with tf.GradientTape() as tape:\r\n            predictions = self.discriminator(self.generator(random_latent_vectors))\r\n            g_loss = self.loss_fn(misleading_labels, predictions)\r\n        grads = tape.gradient(g_loss, self.generator.trainable_weights)\r\n        self.g_optimizer.apply_gradients(zip(grads, self.generator.trainable_weights))\r\n        # 更新损失\r\n        self.d_loss_metric.update_state(d_loss)\r\n        self.g_loss_metric.update_state(g_loss)\r\n        return {\r\n            \"d_loss\": self.d_loss_metric.result(),\r\n            \"g_loss\": self.g_loss_metric.result(),\r\n        }\r\n```\r\n\r\n\r\n### 3.3 基于元学习的数据增强\r\n\r\n深度学习研究中的元学习(Meta learning)通常是指使用神经网络优化神经网络,元学习的数据增强有神经增强(Neural augmentation)等方法。\r\n- 神经增强\r\n\r\n神经增强(Neural augmentation)是通过神经网络组的学习以获得较优的数据增强并改善分类效果的一种方法。 其方法步骤如下:\r\n\r\n1、获取与target图像同一类别的一对随机图像,前置的增强网络通过CNN将它们映射为合成图像,合成图像与target图像对比计算损失;\r\n\r\n2、将合成图像与target图像神经风格转换后输入到分类网络中,并输出该图像分类损失;\r\n\r\n3、将增强与分类的loss加权平均后,反向传播以更新分类网络及增强网络权重。使得其输出图像的同类内差距减小且分类准确。\r\n![](https://upload-images.jianshu.io/upload_images/11682271-9e82226797f2017a.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n\r\n\r\n### 3.4 基于神经风格迁移的数据增强\r\n神经风格迁移(Neural Style Transfer)可以在保留原始内容的同时,将一个图像的样式转移到另一个图像上。除了实现类似色彩空间照明转换,还可以生成不同的纹理和艺术风格。![](https://upload-images.jianshu.io/upload_images/11682271-fea6f0de1e7f14a1.png?imageMogr2/auto-orient/strip%7CimageView2/2/w/1240)\r\n\r\n神经风格迁移是通过优化三类的损失来实现的:\r\n\r\nstyle_loss:使生成的图像接近样式参考图像的局部纹理;\r\n\r\ncontent_loss:使生成的图像的内容表示接近于基本图像的表示;\r\n\r\ntotal_variation_loss:是一个正则化损失,它使生成的图像保持局部一致。\r\n```\r\n# 样式损失\r\ndef style_loss(style, combination):\r\n    S = gram_matrix(style)\r\n    C = gram_matrix(combination)\r\n    channels = 3\r\n    size = img_nrows * img_ncols\r\n    return tf.reduce_sum(tf.square(S - C)) / (4.0 * (channels ** 2) * (size ** 2))\r\n\r\n# 内容损失\r\ndef content_loss(base, combination):\r\n    return tf.reduce_sum(tf.square(combination - base))\r\n\r\n# 正则损失\r\ndef total_variation_loss(x):\r\n    a = tf.square(\r\n        x[:, : img_nrows - 1, : img_ncols - 1, :] - x[:, 1:, : img_ncols - 1, :]\r\n    )\r\n    b = tf.square(\r\n        x[:, : img_nrows - 1, : img_ncols - 1, :] - x[:, : img_nrows - 1, 1:, :]\r\n    )\r\n    return tf.reduce_sum(tf.pow(a + b, 1.25))\r\n```\r\n\r\n\r\n\r\n\r\n---\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","author":{"url":"https://github.com/aialgorithm","@type":"Person","name":"aialgorithm"},"datePublished":"2021-03-17T11:35:57.000Z","interactionStatistic":{"@type":"InteractionCounter","interactionType":"https://schema.org/CommentAction","userInteractionCount":0},"url":"https://github.com/13/Blog/issues/13"}

route-pattern/_view_fragments/issues/show/:user_id/:repository/:id/issue_layout(.:format)
route-controllervoltron_issues_fragments
route-actionissue_layout
fetch-noncev2:fee46279-7ce6-6f4c-a797-912a59b5d62a
current-catalog-service-hash81bb79d38c15960b92d99bca9288a9108c7a47b18f2423d0f6438c5b7bcd2114
request-idB06C:281234:821F6D:B598EB:696A162C
html-safe-noncee7e08b468348490a989274eeaad51bd53de005e3b5abb96bfa394cbd727452a5
visitor-payloadeyJyZWZlcnJlciI6IiIsInJlcXVlc3RfaWQiOiJCMDZDOjI4MTIzNDo4MjFGNkQ6QjU5OEVCOjY5NkExNjJDIiwidmlzaXRvcl9pZCI6IjgxNzI1MTUzMDc5MDk5NDQ4NzYiLCJyZWdpb25fZWRnZSI6ImlhZCIsInJlZ2lvbl9yZW5kZXIiOiJpYWQifQ==
visitor-hmac6f80472a58b0f2d5ae0c4871ccdab871e40ed5c06bc64b6866d288a67b4f9563
hovercard-subject-tagissue:833678345
github-keyboard-shortcutsrepository,issues,copilot
google-site-verificationApib7-x98H0j5cPqHWwSMm6dNU4GmODRoqxLiDzdx9I
octolytics-urlhttps://collector.github.com/github/collect
analytics-location///voltron/issues_fragments/issue_layout
fb:app_id1401488693436528
apple-itunes-appapp-id=1477376905, app-argument=https://github.com/_view_fragments/issues/show/aialgorithm/Blog/13/issue_layout
twitter:imagehttps://opengraph.githubassets.com/11efae0c33d8e69060c72ce4db424b07dc74490c9e99c1499602e31eb2e9eb1e/aialgorithm/Blog/issues/13
twitter:cardsummary_large_image
og:imagehttps://opengraph.githubassets.com/11efae0c33d8e69060c72ce4db424b07dc74490c9e99c1499602e31eb2e9eb1e/aialgorithm/Blog/issues/13
og:image:alt数据、算法、算力是人工智能发展的三要素。数据决定了Ai模型学习的上限,数据规模越大、质量越高,模型就能够拥有更好的泛化能力。然而在实际工程中,训练的数据相对模型而言数据量太少,或者很难覆盖全部的场景等问题,解决这问题的一个有效途径是通过数据增强(Data Augmentation),使模型获得较好的泛化性能。 1 数据增强介绍 数据增强(Data Augmentation)是在不实质性的增加...
og:image:width1200
og:image:height600
og:site_nameGitHub
og:typeobject
og:author:usernameaialgorithm
hostnamegithub.com
expected-hostnamegithub.com
None34a52bd10bd674f68e5c1b6b74413b79bf2ca20c551055ace3f7cdd112803923
turbo-cache-controlno-preview
go-importgithub.com/aialgorithm/Blog git https://github.com/aialgorithm/Blog.git
octolytics-dimension-user_id33707637
octolytics-dimension-user_loginaialgorithm
octolytics-dimension-repository_id147093233
octolytics-dimension-repository_nwoaialgorithm/Blog
octolytics-dimension-repository_publictrue
octolytics-dimension-repository_is_forkfalse
octolytics-dimension-repository_network_root_id147093233
octolytics-dimension-repository_network_root_nwoaialgorithm/Blog
turbo-body-classeslogged-out env-production page-responsive
disable-turbofalse
browser-stats-urlhttps://api.github.com/_private/browser/stats
browser-errors-urlhttps://api.github.com/_private/browser/errors
releasee8bd37502700f365b18a4d39acf7cb7947e11b1a
ui-targetfull
theme-color#1e2327
color-schemelight dark

Links:

Skip to contenthttps://github.com/aialgorithm/Blog/issues/13#start-of-content
https://github.com/
Sign in https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Faialgorithm%2FBlog%2Fissues%2F13
GitHub CopilotWrite better code with AIhttps://github.com/features/copilot
GitHub SparkBuild and deploy intelligent appshttps://github.com/features/spark
GitHub ModelsManage and compare promptshttps://github.com/features/models
MCP RegistryNewIntegrate external toolshttps://github.com/mcp
ActionsAutomate any workflowhttps://github.com/features/actions
CodespacesInstant dev environmentshttps://github.com/features/codespaces
IssuesPlan and track workhttps://github.com/features/issues
Code ReviewManage code changeshttps://github.com/features/code-review
GitHub Advanced SecurityFind and fix vulnerabilitieshttps://github.com/security/advanced-security
Code securitySecure your code as you buildhttps://github.com/security/advanced-security/code-security
Secret protectionStop leaks before they starthttps://github.com/security/advanced-security/secret-protection
Why GitHubhttps://github.com/why-github
Documentationhttps://docs.github.com
Bloghttps://github.blog
Changeloghttps://github.blog/changelog
Marketplacehttps://github.com/marketplace
View all featureshttps://github.com/features
Enterpriseshttps://github.com/enterprise
Small and medium teamshttps://github.com/team
Startupshttps://github.com/enterprise/startups
Nonprofitshttps://github.com/solutions/industry/nonprofits
App Modernizationhttps://github.com/solutions/use-case/app-modernization
DevSecOpshttps://github.com/solutions/use-case/devsecops
DevOpshttps://github.com/solutions/use-case/devops
CI/CDhttps://github.com/solutions/use-case/ci-cd
View all use caseshttps://github.com/solutions/use-case
Healthcarehttps://github.com/solutions/industry/healthcare
Financial serviceshttps://github.com/solutions/industry/financial-services
Manufacturinghttps://github.com/solutions/industry/manufacturing
Governmenthttps://github.com/solutions/industry/government
View all industrieshttps://github.com/solutions/industry
View all solutionshttps://github.com/solutions
AIhttps://github.com/resources/articles?topic=ai
Software Developmenthttps://github.com/resources/articles?topic=software-development
DevOpshttps://github.com/resources/articles?topic=devops
Securityhttps://github.com/resources/articles?topic=security
View all topicshttps://github.com/resources/articles
Customer storieshttps://github.com/customer-stories
Events & webinarshttps://github.com/resources/events
Ebooks & reportshttps://github.com/resources/whitepapers
Business insightshttps://github.com/solutions/executive-insights
GitHub Skillshttps://skills.github.com
Documentationhttps://docs.github.com
Customer supporthttps://support.github.com
Community forumhttps://github.com/orgs/community/discussions
Trust centerhttps://github.com/trust-center
Partnershttps://github.com/partners
GitHub SponsorsFund open source developershttps://github.com/sponsors
Security Labhttps://securitylab.github.com
Maintainer Communityhttps://maintainers.github.com
Acceleratorhttps://github.com/accelerator
Archive Programhttps://archiveprogram.github.com
Topicshttps://github.com/topics
Trendinghttps://github.com/trending
Collectionshttps://github.com/collections
Enterprise platformAI-powered developer platformhttps://github.com/enterprise
GitHub Advanced SecurityEnterprise-grade security featureshttps://github.com/security/advanced-security
Copilot for BusinessEnterprise-grade AI featureshttps://github.com/features/copilot/copilot-business
Premium SupportEnterprise-grade 24/7 supporthttps://github.com/premium-support
Pricinghttps://github.com/pricing
Search syntax tipshttps://docs.github.com/search-github/github-code-search/understanding-github-code-search-syntax
documentationhttps://docs.github.com/search-github/github-code-search/understanding-github-code-search-syntax
Sign in https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Faialgorithm%2FBlog%2Fissues%2F13
Sign up https://github.com/signup?ref_cta=Sign+up&ref_loc=header+logged+out&ref_page=%2F%3Cuser-name%3E%2F%3Crepo-name%3E%2Fvoltron%2Fissues_fragments%2Fissue_layout&source=header-repo&source_repo=aialgorithm%2FBlog
Reloadhttps://github.com/aialgorithm/Blog/issues/13
Reloadhttps://github.com/aialgorithm/Blog/issues/13
Reloadhttps://github.com/aialgorithm/Blog/issues/13
aialgorithm https://github.com/aialgorithm
Bloghttps://github.com/aialgorithm/Blog
Notifications https://github.com/login?return_to=%2Faialgorithm%2FBlog
Fork 259 https://github.com/login?return_to=%2Faialgorithm%2FBlog
Star 942 https://github.com/login?return_to=%2Faialgorithm%2FBlog
Code https://github.com/aialgorithm/Blog
Issues 66 https://github.com/aialgorithm/Blog/issues
Pull requests 0 https://github.com/aialgorithm/Blog/pulls
Actions https://github.com/aialgorithm/Blog/actions
Projects 0 https://github.com/aialgorithm/Blog/projects
Security Uh oh! There was an error while loading. Please reload this page. https://github.com/aialgorithm/Blog/security
Please reload this pagehttps://github.com/aialgorithm/Blog/issues/13
Insights https://github.com/aialgorithm/Blog/pulse
Code https://github.com/aialgorithm/Blog
Issues https://github.com/aialgorithm/Blog/issues
Pull requests https://github.com/aialgorithm/Blog/pulls
Actions https://github.com/aialgorithm/Blog/actions
Projects https://github.com/aialgorithm/Blog/projects
Security https://github.com/aialgorithm/Blog/security
Insights https://github.com/aialgorithm/Blog/pulse
New issuehttps://github.com/login?return_to=https://github.com/aialgorithm/Blog/issues/13
New issuehttps://github.com/login?return_to=https://github.com/aialgorithm/Blog/issues/13
一文归纳Ai数据增强之法https://github.com/aialgorithm/Blog/issues/13#top
https://github.com/aialgorithm
https://github.com/aialgorithm
aialgorithmhttps://github.com/aialgorithm
on Mar 17, 2021https://github.com/aialgorithm/Blog/issues/13#issue-833678345
https://camo.githubusercontent.com/9f65b919662dd16e8bb012ede2097d9398f6efc23f132d840b580c67c5905c2d/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d663561386130396133316638323330342e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/4da724652354f52497c5bbc3940446843acfea9986cf92902d8e77a691960d7d/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d346335376439363131323062326166622e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/4148929375358e2a332c4a1c113a8fb2edf79459beb4add565aa1375ebd926c1/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d643637323036303630366463636337392e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/ffabacd36bd24892e8e9963a865abe079f8f2030655b65be15aa93136feaf480/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d323931666132616264363934383163382e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/34fdddf0615398f80f81f22d493b6657e102ef6f1b0913f5dd17f7e23ef86a7d/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d313661323234313135613433333263352e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/eaa05d454ddd0bb9ca86bf3571b80505a19978f84f6e22e39553c20de9471771/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d663530386232613034316637636335392e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/0b2e80f4ac49f076a9b4b6ffd9861af5971807676b72d2c8539dd1817130fc6b/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d663863653039343030333439323634362e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/643b943ef6aa1b435acecde16d77d10dcac80c2db632e75947ec07aefe4d7d31/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d353632326235346162663230383837612e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/6bb666cd8e0886e47f5c40574b6336ea8601ac61c029b3878013335d8fc15eb5/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d343236383439306438353536623566302e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/569cbe31434167473c631a547930a1e9057e891cf7a93a5485d1e644e4e58e25/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d396538323232363739376632303137612e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://camo.githubusercontent.com/95fa503ba32f8fb0bb3830df593030efb459e6e5adfbe5e8b04e9fc67c447042/68747470733a2f2f75706c6f61642d696d616765732e6a69616e7368752e696f2f75706c6f61645f696d616765732f31313638323237312d666561366630646531653766313461312e706e673f696d6167654d6f6772322f6175746f2d6f7269656e742f7374726970253743696d61676556696577322f322f772f31323430
https://github.com
Termshttps://docs.github.com/site-policy/github-terms/github-terms-of-service
Privacyhttps://docs.github.com/site-policy/privacy-policies/github-privacy-statement
Securityhttps://github.com/security
Statushttps://www.githubstatus.com/
Communityhttps://github.community/
Docshttps://docs.github.com/
Contacthttps://support.github.com?tags=dotcom-footer

Viewport: width=device-width


URLs of crawlers that visited me.