Skip to main content

GAN generator loss is 0 from the start

I need data augmentation for network traffic and I'm following an article in which the structure of the discriminator and generator are both specified. My input data is a collection of pcap files, each having 10 packets with 250 bytes. They are then transformed into a (10, 250) array and all the bytes are cast into float64.

def Byte_p_to_float(byte_p):
    float_b = []
    for byte in byte_p:
        float_b.append(float(str(byte)))
    return float_b

Xtrain = []
a = []
for i in range(len(Dataset)):
    for p in range(10):
        a.append(Byte_p_to_float(raw(Dataset[i][p]))) # packets transform to byte, and the float64
    Xtrain.append(a)
    a = []
Xtrain = np.asarray(Xtrain)
Xtrain = Xtrain / 127.5 - 1.0 # normalizing.

I then go on training the model but the generator loss is always 0 from start!

batch_size = 128
interval = 5
iterations = 2000

real = [] # a list of all 1s for true data label
for i in range (batch_size):
    real.append(1)
    
fake = [] # a list of all 1s for fake data label
for i in range (batch_size):
    fake.append(0)


for iteration in range(iterations):

    ids = np.random.randint(0,Xtrain.shape[0],batch_size)
    flows = Xtrain[ids]
    
    z = np.random.normal(0, 1, (batch_size, 100)) # generating gaussian noise vector!
    gen_flows = generator_v.predict(z) 
    gen_flows = ((gen_flows - np.amin(gen_flows))/(np.amax(gen_flows) - np.amin(gen_flows))) * 2 - 1 # normalizing. (-1,+1)
    
    # gen_flows returns float32 and here i transform to float 64. not sure if its necessary
    t = np.array([])
    for i in range(batch_size):
        t = np.append(t ,[np.float64(gen_flows[i])])
    t = t.reshape(batch_size, 2500)
    gen_flows = []
    gen_flows = t
    
    nreal = np.asarray(real)
    nfake = np.asarray(fake)
    nflows = flows.reshape(batch_size, 2500) # this way we match the article.
    dloss_real = discriminator_v.train_on_batch(nflows, nreal) # training the discriminator on real data
        
    dloss_fake = discriminator_v.train_on_batch(gen_flows, nfake) # training the discriminator on fake data

    dloss, accuracy = 0.5 * np.add(dloss_real,dloss_fake)

    z = np.random.normal(0, 1, (batch_size, 100)) # generating gaussian noise vector for GAN
    gloss = gan_v.train_on_batch(z, nreal)

    if (iteration + 1) % interval == 0:
        losses.append((dloss, gloss))
        accuracies.append(100.0 * accuracy)
        iteration_checks.append(iteration + 1)
        print("%d [D loss: %f , acc: %.2f] [G loss: %f]" % (iteration+1,dloss,100.0*accuracy,gloss))
        

[the model description in the article is here][1]

and finally here is my model:

losses=[]
accuracies=[]
iteration_checks=[]
zdim = np.random.normal(0,1,100) # 100 dimentional gaussian noise vector

def build_generator(gause_len):
    model = Sequential()

    model.add(Input(shape=(gause_len,)))
    model.add(Dense(256))
    model.add(LeakyReLU())
    model.add(BatchNormalization())

    model.add(Dense(512))
    model.add(LeakyReLU())
    model.add(BatchNormalization())

    model.add(Dense(1024))
    model.add(LeakyReLU())
    model.add(BatchNormalization())

    model.add(Dense(2500))
    model.add(LeakyReLU(2500))
    #model.add(reshape(img_shape))
    return model
def build_discriminator():
    model = Sequential()
    
    
    model.add(Input(shape=(2500))) #input shape
    model.add(Dense(2500))
    #model.add( Dense(2500, input_shape=img_shape) )
    model.add(LeakyReLU())
    model.add(Dropout(0.5))
    

    model.add(Dense(1024, ))
    model.add(LeakyReLU())
    model.add(Dropout(0.5))

    model.add(Dense(512, ))
    model.add(LeakyReLU())
    model.add(Dropout(0.5))

    model.add(Dense(256, ))
    model.add(LeakyReLU())
    model.add(Dropout(0.5))

    model.add(Dense(1, ))
    return model

def build_gan(generator, discriminator):
    model = Sequential()
    model.add(generator)
    model.add(discriminator)
    return model

# used for training the discriminator netowrk
discriminator_v = build_discriminator()
discriminator_v.compile(loss='binary_crossentropy', optimizer=Adam(), metrics=['accuracy'])

# used for training the Generator netowrk
generator_v = build_generator(len(zdim))
discriminator_v.trainable = False

# used for training the GAN netowrk
gan_v = build_gan(generator_v, discriminator_v)
gan_v.compile(loss='binary_crossentropy', optimizer=Adam())

AI isn't my area of expertise and all this is part of a much larger project, so the error may be obvious. any help will be much appreciated. [1]: https://i.stack.imgur.com/DqcjR.png



source https://stackoverflow.com/questions/72028701/gan-generator-loss-is-0-from-the-start

Comments

Popular posts from this blog

ValueError: X has 10 features, but LinearRegression is expecting 1 features as input

So, I am trying to predict the model but its throwing error like it has 10 features but it expacts only 1. So I am confused can anyone help me with it? more importantly its not working for me when my friend runs it. It works perfectly fine dose anyone know the reason about it? cv = KFold(n_splits = 10) all_loss = [] for i in range(9): # 1st for loop over polynomial orders poly_order = i X_train = make_polynomial(x, poly_order) loss_at_order = [] # initiate a set to collect loss for CV for train_index, test_index in cv.split(X_train): print('TRAIN:', train_index, 'TEST:', test_index) X_train_cv, X_test_cv = X_train[train_index], X_test[test_index] t_train_cv, t_test_cv = t[train_index], t[test_index] reg.fit(X_train_cv, t_train_cv) loss_at_order.append(np.mean((t_test_cv - reg.predict(X_test_cv))**2)) # collect loss at fold all_loss.append(np.mean(loss_at_order)) # collect loss at order plt.plot(np.log(al...

Sorting large arrays of big numeric stings

I was solving bigSorting() problem from hackerrank: Consider an array of numeric strings where each string is a positive number with anywhere from to digits. Sort the array's elements in non-decreasing, or ascending order of their integer values and return the sorted array. I know it works as follows: def bigSorting(unsorted): return sorted(unsorted, key=int) But I didnt guess this approach earlier. Initially I tried below: def bigSorting(unsorted): int_unsorted = [int(i) for i in unsorted] int_sorted = sorted(int_unsorted) return [str(i) for i in int_sorted] However, for some of the test cases, it was showing time limit exceeded. Why is it so? PS: I dont know exactly what those test cases were as hacker rank does not reveal all test cases. source https://stackoverflow.com/questions/73007397/sorting-large-arrays-of-big-numeric-stings

How to load Javascript with imported modules?

I am trying to import modules from tensorflowjs, and below is my code. test.html <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Document</title </head> <body> <script src="https://cdn.jsdelivr.net/npm/@tensorflow/tfjs@2.0.0/dist/tf.min.js"></script> <script type="module" src="./test.js"></script> </body> </html> test.js import * as tf from "./node_modules/@tensorflow/tfjs"; import {loadGraphModel} from "./node_modules/@tensorflow/tfjs-converter"; const MODEL_URL = './model.json'; const model = await loadGraphModel(MODEL_URL); const cat = document.getElementById('cat'); model.execute(tf.browser.fromPixels(cat)); Besides, I run the server using python -m http.server in my command prompt(Windows 10), and this is the error prompt in the console log of my browser: Failed to loa...