Code Monkey home page Code Monkey logo

fall2023project's Introduction

fall2023project

Project repository for Fall 2023.

Additional project updates and ideas are tracked here in this Google document.

Thurs., 31 Aug 2023

Need to debug/figure out why model(s') performance is/are drastically different from notebooks 4 and 6 & 7.

Mon., 28 Aug 2023

Results are located in sandbox4 notebook. At the moment, not all tasks assigned from Mon., 14 Aug 2023 have been completed. So far:

  • [] Add option to store gradient norm of each layer, stored separately
  • Change linear layers to: CNN + 1 linear layer
  • Make deep model (5 layers), and train it to perfection (up to 99% or higher train accuracy)
  • Save the model (we’ll call this the “ground model”) (if time, create 5 ground models)
  • [] Then, create 10 models per noise level (pick 10 noise levels, between totally destroyed and basically no impact) (also loop through which layer)→ turns into 500 models. Make them noisy, measure all the things above (robustness, generalization.1, try generalization.2)
  • [] Can experiment with gradcam (interesting but not most important)
  • Try training with 32 all the way (in conv layers) - see if model can still be 99% good
  • Use a smaller model, smallest non-trivial model
  • Reduce number of linear layers
  • Start profiling (draw on piece of paper)
  • [] Look for number of weights in each model
  • [] Get model training up to 100%

Before Thr.

  • Add option to store gradient norm of each layer, stored separately
  • Create table, row -> model, col -> specs (grad norm, layerwise norm, specify train/test accuracy), list number of tunable parameters for each model.
  • Add norms of total and/or per layer parameters to the table.
  • [] GradCam (wishlist or next step)

Mon., 21 Aug 2023

Results are located in sandbox4 notebook. At the moment, not all tasks assigned from last week have been completed. So far:

  • Add option to store gradient norm of each layer, stored separately
  • Change linear layers to: CNN + 1 linear layer
  • Make deep model (5 layers), and train it to perfection (up to 99% or higher train accuracy)
  • Save the model (we’ll call this the “ground model”) (if time, create 5 ground models)
  • Then, create 10 models per noise level (pick 10 noise levels, between totally destroyed and basically no impact) (also loop through which layer)→ turns into 500 models. Make them noisy, measure all the things above (robustness, generalization.1, try generalization.2)
  • Can experiment with gradcam (interesting but not most important)

Mon., 14 Aug 2023

Current results are located in the sandbox3 notebook. The results mainly include a function whose input is a model and it returns a dictionary of gradients for both the weights and the bias(es). The dataset used was the MNIST dataset. The model used was a simple 5-layer neural network.

fall2023project's People

Contributors

adebayobraimah avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.