tensorflow 批次读取文件内的数据,并将顺序随机化处理. --[python]

Posted Gxjun

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了tensorflow 批次读取文件内的数据,并将顺序随机化处理. --[python]相关的知识,希望对你有一定的参考价值。

使用tensorflow批次的读取预处理之后的文本数据,并将其分为一个迭代器批次:

比如此刻,我有一个处理之后的数据包: data.csv  shape =(8,10),其中这个结构中,前五个列为feature , 后五列为label

1,2,3,4,5,6,7,8,9,10
11,12,13,14,15,16,17,18,19,20
21,22,23,24,25,26,27,28,29,30
31,32,33,34,35,36,37,38,39,40
41,42,43,44,45,46,47,48,49,50
51,52,53,54,55,56,57,58,59,60
1,1,1,1,1,2,2,2,2,2
3,3,3,3,3,4,4,4,4,4

现在我需要将其分为4个批次: 也就是每个批次batch的大小为2

然后我可能需要将其顺序打乱,所以这里提供了两种方式,顺序和随机

#!/usr/bin/env python
# -*- coding: utf-8 -*-
__author__ = xijun1
import tensorflow as tf
import numpy as np

# data = np.arange(1, 100 + 1)
# print ",".join( [str(i) for i in data])
# data_input = tf.constant(data)
filename_queue = tf.train.string_input_producer(["data.csv"])
reader = tf.TextLineReader(skip_header_lines=0)
key, value = reader.read(filename_queue)
# decode_csv will convert a Tensor from type string (the text line) in
# a tuple of tensor columns with the specified defaults, which also
# sets the data type for each column
words_size = 5  # 每一行数据的长度
decoded = tf.decode_csv(
    value,
    field_delim=,,
    record_defaults=[[0] for i in range(words_size * 2)])

batch_size = 2 # 每一个批次的大小
# 随机
batch_shuffle = tf.train.shuffle_batch(decoded, batch_size=batch_size,
                                       capacity=batch_size * words_size,
                                       min_after_dequeue=batch_size)
#顺序
batch_no_shuffle = tf.train.batch(decoded, batch_size=batch_size, capacity=batch_size * words_size,
                                  allow_smaller_final_batch=batch_size)
shuffle_features = tf.transpose(tf.stack(batch_shuffle[0:words_size]))
shuffle_label = tf.transpose(tf.stack(batch_shuffle[words_size:]))
features = tf.transpose(tf.stack(batch_no_shuffle[0:words_size]))
label = tf.transpose(tf.stack(batch_no_shuffle[words_size:]))

with tf.Session() as sess:
    coord = tf.train.Coordinator()
    threads = tf.train.start_queue_runners(coord=coord)
    for i in range(8/batch_size):
        print (i+10, sess.run([shuffle_features, shuffle_label]))
        print (i, sess.run([features, label]))
    coord.request_stop()
    coord.join(threads)

当我们运行的时候,我们可以得到这个结果:

(10, [array([[ 1,  2,  3,  4,  5],
       [31, 32, 33, 34, 35]], dtype=int32), array([[ 6,  7,  8,  9, 10],
       [36, 37, 38, 39, 40]], dtype=int32)])
(0, [array([[11, 12, 13, 14, 15],
       [21, 22, 23, 24, 25]], dtype=int32), array([[16, 17, 18, 19, 20],
       [26, 27, 28, 29, 30]], dtype=int32)])
(11, [array([[51, 52, 53, 54, 55],
       [ 3,  3,  3,  3,  3]], dtype=int32), array([[56, 57, 58, 59, 60],
       [ 4,  4,  4,  4,  4]], dtype=int32)])
(1, [array([[41, 42, 43, 44, 45],
       [ 1,  1,  1,  1,  1]], dtype=int32), array([[46, 47, 48, 49, 50],
       [ 2,  2,  2,  2,  2]], dtype=int32)])
(12, [array([[ 3,  3,  3,  3,  3],
       [11, 12, 13, 14, 15]], dtype=int32), array([[ 4,  4,  4,  4,  4],
       [16, 17, 18, 19, 20]], dtype=int32)])
(2, [array([[ 1,  2,  3,  4,  5],
       [21, 22, 23, 24, 25]], dtype=int32), array([[ 6,  7,  8,  9, 10],
       [26, 27, 28, 29, 30]], dtype=int32)])
(13, [array([[31, 32, 33, 34, 35],
       [ 1,  1,  1,  1,  1]], dtype=int32), array([[36, 37, 38, 39, 40],
       [ 2,  2,  2,  2,  2]], dtype=int32)])
(3, [array([[41, 42, 43, 44, 45],
       [ 1,  1,  1,  1,  1]], dtype=int32), array([[46, 47, 48, 49, 50],
       [ 2,  2,  2,  2,  2]], dtype=int32)])

 

以上是关于tensorflow 批次读取文件内的数据,并将顺序随机化处理. --[python]的主要内容,如果未能解决你的问题,请参考以下文章

Tensorflow 2.0:从回调中访问批次的张量

tensorflow训练自己的数据集实现CNN图像分类

tensorflow数据读取机制tf.train.slice_input_producer 和 tf.train.batch 函数

合并 tensorflow 数据集批次

在 Tensorflow 中将数据拆分为批次进行分类

分批次读取csv文件,并保存到数据库