Convolutional Neural Networks

share ›
‹ links

Below are the top discussions from Reddit that mention this online Coursera course from DeepLearning.AI.

Offered by DeepLearning.AI. In the fourth course of the Deep Learning Specialization, you will understand how computer vision has evolved ... Enroll for free.

Reddsera may receive an affiliate commission if you enroll in a paid course after using these buttons to visit Coursera. Thank you for using these buttons to support Reddsera.

Taught by
Andrew Ng
and 2 more instructors

Offered by

Reddit Posts and Comments

1 posts • 14 mentions • top 14 shown below

r/MachineLearning • post
95 points • curious_riddler
[N] Andrew Ng's Deep learning specialization's 4th course is up now on Coursera which is on CNNs
r/learnpython • comment
4 points • Nikota

You may be able to find some blog somewhere that does it, but that's likely a topic for a proper class on deep learning as it is very math heavy.

You can find one introductory-level online class on this topic on Coursera's Deep Learning by Andrew Ng. Here, you don't do most of the formal derivations for convolutional backprop, but you do fully implement a convnet in Python only using Numpy in a notebook:

r/programming • comment
3 points • torshi16

I also like the videos here:

r/MLQuestions • comment
3 points • knee2710

Yes there has been significant work in this area, most of the face recognition application like Apple's face ID system uses one shot learning. If you think about it Apple is not training the Face recognition algorithm on your face but even when you buy an IPhoneX you are able to achieve amazing accuracy (actually they claim the error to 1/1,000,000 which means there is a possibility of failing once in every 1 million tests) how are they doing it? They use something called Siamese network, these networks once trained can take two images as input and output how far are they from each other (basically means if they are same or not). So once you train one such model your model will learn the weights to find the similarity between a given pair of images and with enough data and training this model can accurately differentiate if two images are same or not. Going back to my IPhoneX example, when you login into your phone for the first time it asks you to register your face in different orientation and after that anytime anyone tries to log in it uses the pertained Siamese network to see if you are the same person or not. So technically the algorithm was able to find out the similarity with just one shot of learning. For some reference on the same please look at this video :

r/MLQuestions • comment
2 points • pram-ila

Recommending Andrew Ng courses is par for the course on this sub, but here's a specific video

r/learnmachinelearning • comment
1 points • Aprtnaak

This one:

r/deeplearning • comment
1 points • Alternative_Recipe39

Anything Andrew Ng produces. You could probably do this part of the deeplearning specialisation in a week. Or you could just watch the videos to get a conceptual understanding.

r/learnmachinelearning • comment
1 points • visarga

After Andrew's ML course try a course on neural nets, or a specialised course on convolutional neural nets. There's such a course from Andrew as well.

r/learnmachinelearning • comment
1 points • RamiNoob

For CNNs, Andrew Ng's course on Coursera is a good starting point.

r/learnmachinelearning • post
6 points • roar363
Yolo and yolo9000 class predictions [xpost r/MachineLearning]

I am trying to understand yolo ( and have some confusion over the final class predictions. The original paper and presentation by the author ( says "We only predict one set of class probabilities per grid cell, regardless of the number of boxes B." However Andrew Ng's CNN course ( and Siraj Raval's video ( say we predict a set of class probabilities for each bounding box. Has anyone come across this? Why the disparity? Which approach is actually used/which is better?

r/MachineLearning • comment
2 points • geaxart

>The only type of labeled training images that I've seen is when the entire image is labeled with an entity (ex: "dog", or "cat").

In the domain of machine learning, this is called "classification".

Detection could be seen as predicting bounding boxes + classification of the corresponding cropped areas. You can imagine an algorithm that takes every possible cropped areas of an image and feed them to a classifier. In theory, it would work, but practically would be way to long in processing time. Some of the first detection neural nets (like RCNN) had a module whose job was to propose areas/regions to the classifier, and thus limited the processing time.

YOLO is even faster by doing both tasks in a single convolution neural network. I think it is interesting for you to know how YOLO works, but the subject is too long to be described here. May I suggest to you to watch the videos from (week 3) ? I think you have to subscribe to access the videos, but it is free. And it is very well explained, I think.

r/deeplearning • comment
1 points • recsysdork

I've taken Andrew Ng's Coursera material on convolutional neural networks, which covers a lot more than just basic CNN theory. It's REALLY good:

r/computervision • comment
1 points • mikeyzhong

Yeah! What you want is the YOLO algorithm. It's state of the art for multiclass object detection.

r/compsci • comment
1 points • XG_West

It is a bit hard to recommend a course of study without knowing your language of choice or your background but I would recommend the deep learning course on Coursera as a starting point:

or specifically course 4 if you are just interested in CNNs and don't have much time:

For a very high level overview of different deep learning architectures and approaches, this is a nice summary:

It is a bit dated in that YOLOv3 (Darknet53) has been released and it doesn't mention anything on Retinanet.

A common trade-off you will hear is between accuracy and speed. I would urge you to consider the real implications of these results to your project as they tend to be competitive for mostly academic purposes.