Code Monkey home page Code Monkey logo

coursera-deep-learning's Introduction

Coursera

My notes / works from Coursera courses.

Contents

Introduction

This repository contains my solutions for labs and programming assignments on Coursera courses. Certain resources required by the codes may be lacking due to limitations on downloading course materials from Coursera and uploading them to GitHub. The lacking resources are mostly datasets, pre-trained models or certain weight matrices.

Courses

TensorFlow: Advanced Techniques (Specialization)

Details

Week 1 - Functional APIs

Week 2 - Custom Loss Functions

Week 3 - Custom Layers

Week 4 - Custom Models

Week 5 - Bonus Content - Callbacks

Details

Week 1 - Differentiation and Gradients

Week 2 - Custom Training

Week 3 - Graph Mode

Week 4 - Distributed Training

Details

Week 1 - Introduction to Computer Vision

Week 2 - Object Detection

Week 3 - Image Segmentation

Week 4 - Visualization and Interpretability


DeepLearning.AI TensorFlow Developer (Specialization)

Details

Week 1 - A New Programming Paradigm

Week 2 - Introduction to Computer Vision

Week 3 - Enchancing Vision with Convolutional Neural Networks

Week 4 - Using Real-world Images

Details

Week 1 - Exploring a Larger Dataset

Programming Assignment: Exercise 1 - Cats vs. Dogs

Week 2 - Augmentation: A Technique to Avoid Overfitting

Programming Assignment: Exercise 2 - Cats vs. Dogs using augmentation

Week 3 - Transfer Learning

Programming Assignment: Exercise 3 - Horses vs. humans using Transfer Learning

Week 4 - Multiclass Classifications

Programming Assignment: Exercise 4 - Multi-class classifier

Unable to download horse-or-human.zip

Details

Week 1 - Sentiment in Text

Week 2 - Word Embeddings

Week 3 - Sequence Models

Week 4 - Sequence Models and Literature

Details

Week 1 - Sequences and Prediction

Week 2 - Deep Neural Networks for Time Series

Week 3 - Recurrent Neural Networks for Time Series

Week 4 - Real-world Time Series Data


Generative Adversarial Networks (GANs) (Specialization)

Details

Week 1 - Intro to GANs

Week 2 - Deep Convolutional GANs

Week 3 - Wasserstein GANs with Gradient Penalty

Week 4 - Conditional GAN & Controllable Generation

Details

Week 1 - Evaluation of GANs

Unable to download inception_v3_google-1a9a5a14.pth, fid_images_tensor.npz

Week 2 - GAN Disadvantages and Bias

Week 3 - StyleGAN and Advancements

Details

Week 1 - GANs for Data Augmentation and Privacy

Week 2 - Image-to-Image Translation with Pix2Pix

Unable to download pix2pix_15000.pth, maps

Week 3 - Unpaired Translation with CycleGAN

Unable to download horse2zebra, cycleGAN_100000.pth


Natural Language Processing (Specialization)

Details

Week 1 - Sentiment Analysis with Logistic Regression

Week 2 - Sentiment Analysis with Naive Bayes

Week 3 - Vector Space Models

Week 4 - Machine Translation and Document Search

Details

Week 1 - Autocorrect

Week 2 - Part of Speech Tagging and Hidden Markov Models

Week 3 - Autocomplete and Language Models

Week 4 - Word Embeddings with Neural Networks

Details

Week 1 - Neural Netowrks for Sentiment Analysis

Week 2 - Recurrent Neural Networks for Language Modelling

Week 3 - LSTMs and Named Entity Recognition

Week 4 - Siamese Networks

Details

Week 1 - Neural Machine Translation

Week 2 - Text Summarization

Week 3 - Question Answering

Week 4 - Chatbot


Deep Learning (Specialization)

Details

Week 1 - Introduction to Deep Learning

  • No labs / programming assignments

Week 2 - Neural Network Basics

Week 3 - Shallow Neural Networks

Week 4 - Deep Neural Networks

Details

Week 1 - Practical Aspects of Deep Learning

Week 2 - Optimization Algorithms

Week 3 - Hyperparameter Tuning, Batch Normalization and Programming Frameworks

Details
  • No labs / programming assignments
Details

Week 1 - Foundations of Convolutional Neural Networks

Week 2 - Deep Convolutional Models: Case Studies

Week 3 - Object Detection

Week 4 - Special Applications: Face Recognition & Neural Style Transfer

Details

Week 1 - Recurrent Neural Networks

Week 2 - Natural Language Processing & Word Embeddings

Week 3 - Sequence Models & Attention Mechanism

Contributing

License

Coursera is licensed under the MIT license.

coursera-deep-learning's People

Contributors

stephenjohnmoore avatar y33-j3t avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

coursera-deep-learning's Issues

Siamese network for a sequence-to-sequence generation

Shall I use the siamese network for a sequence-to-sequence generation problem in machine learning?

Eg: Input 1: Sentence 1 (sequence) Input 2: Sentence 2 (sequence) Output: Newly Generated sentence (Generated sequence)

And is the siamese network accepting only fixed/same length sequences as inputs?

Please feel free to share your thoughts.

CountVectorizer returning an array with only zeros

Hi, everyone!

I am new in NLP and it's the first time I use sklearn vectorizer, following a tutorial with another corpus for sentiment analysis. For some reason the arrays are almost only zeros (a 1 here and there, but very few of them).

The following code is what I used to preprocess the corpus.

def get_part_of_speech(word):
  probable_part_of_speech = wordnet.synsets(word)
  pos_counts = Counter()
  pos_counts["n"] = len(  [ item for item in probable_part_of_speech if item.pos()=="n"]  )
  pos_counts["v"] = len(  [ item for item in probable_part_of_speech if item.pos()=="v"]  )
  pos_counts["a"] = len(  [ item for item in probable_part_of_speech if item.pos()=="a"]  )
  pos_counts["r"] = len(  [ item for item in probable_part_of_speech if item.pos()=="r"]  )
  most_likely_part_of_speech = pos_counts.most_common(1)[0][0]
  return most_likely_part_of_speech

def preprocess_text(text):
    
  cleaned = re.sub(r'\W+', ' ', text).lower()
  tokenized = word_tokenize(cleaned)
  lemmatized = [normalizer.lemmatize(token, get_part_of_speech(token)) for token in tokenized if token not in stopwords.words('english')]
  normalized = ' '.join(lemmatized)
  
  return normalized

And here is the code with the vectorizer (the vocab with the full vocabulary is one of the solutions I found in the threads, but still not working).

pos = open('NLTK/short_reviews/positive.txt', 'r', encoding='latin-1').read()
neg = open('NLTK/short_reviews/negative.txt', 'r', encoding='latin-1').read()

pos_clean = [preprocess_text(sen) for sen in pos.split('\n') if sen != '']
neg_clean = [preprocess_text(sen) for sen in neg.split('\n') if sen != '']

x_clean = pos_clean + neg_clean
labels = [1] * len(pos_clean) + [0] * len(neg_clean)

vocab = []

for sentence in x_clean:
  for word in sentence.split(' '):
    if word not in vocab:
      vocab.append(word)

x_train, x_test, y_train, y_test = train_test_split(
  x_clean, labels, test_size=0.2, random_state=42
)

vectorizer = CountVectorizer(vocabulary=vocab)

x_vec = vectorizer.fit_transform(x_train).toarray()

# xt_vec = vectorizer.transform(x_test).toarray()

with numpy.printoptions(threshold=numpy.inf):
    print(x_vec[0])

Thanks a lot in advance, and please do not hesitate if there is some lack of information!

I hope I can understand what is going on...

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.