Skip to content

Hannah-Richert/iannwtf_hw7

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

iannwtf_hw7

How to run the code:

We made an input-pipeline, which saves and loads our datasets locally after creation.

  • tensorflow version >= 2.6 required (CPU works as well as GPU)
  • ensure that you are in the /iannwtf_hw7 directory
  • to run: python -m pipeline

Outstanding Questions

Can / should you use truncated BPTT here?

  • BPTT can be computationally expensive as the number of timesteps increases and can lead to gradient problems (vanishing,exploding).
  • TBPTT cuts down computation and memory requirements (but the truncation-length has to be chosen carefully in order to work well).
  • To use TBPTT we would need to implement backpropagation on a different level or way, because we would have to optimize our model for each bundle of timesteps, not at the end for all timesteps together.
  • We could theoretically use TBPTT to reduce computation and memory requirement while training our model.
  • The integrated forget-gate in the LSTM cells already helps with vanishing anf exploding gradients

Should you rather take this as a regression, or a classification problem?

  • In our problem: our input consists of n numbers (n = sequnece length) and our target is either 1 or 0, depending on the sum of all input numbers.
  • This is a function with the dimensions f: R^n -> {0,1}, and therefore a classification problem.
  • Nice to note: in the LSTM layers we perform regression tasks, for axample with the function f: R^n -> R (for the last hidden_output).

Model Parameter:

  • data_samples: 96000: 64000/16000/16000 (train,validation,test)
  • data_seq_length: 25
  • batch_size: 32
  • learning_rate = 0.001
  • optimizer: Adam
  • loss = BinaryCrossentropy
  • epochs = 3

Model Architechture

LSTM Cell

image image

Model Results

3epochs_hw7

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages