Sep 27, 2020
Overall it was great a course. A little bit weak in theory. I think for practical purposes whatever was sufficient. The detection of Question duplication was a very much cool model. I enjoy it a lot.
Nov 11, 2021
This is the third course of NLP Specialization. This was a great course and the instructors was amazing. I really learned and understand everything they thought like LSTM, GRU, Siamese Networks etc.
By Julio W•
Jun 28, 2021
I learn to hate Trax in this course. In the assignements, Trax is used only for toy problems, and then, we use a precomputed model. Even the fastnp is used in a very slow mode. Why to learn NLP in a obscure and really bad documented framework if we will finish to use a precomputed models?
Moreover, I try to replicate the results in my own machine (or even in colab) it does not work, because Trax change a lot between versions. Again, why to use, in a course, a framework that is not stable?
In my opinion, using a new and obscure framework to tech new concepts, only because you love it, is (at least) antipedadogical.
By Jorge A C•
Oct 13, 2020
As other course reviewers noted, this course did not help much to build the intuition underlying the methods used. The video lectures were short and the explanations, though concise, were convoluted and not clear at all. For a real understanding of what sequence models are capable of I recommend watching the lecture videos of Stanford CS224N. I
By DANG M K•
Aug 29, 2020
This course material is not good compare to the Deeplearning Specialization. I hope the instructor will write down to explain detail, not just reading from slide
Apr 23, 2021
I was disapointed by this course:
I did not like at all the use of Trax. At our level (student), we need a well established and documented library like Keras or Pytorch to illustrate the concepts. Trax is badly documented. And since the installation of the Trax version used in the assignement fails in Google Colab (!!), I had hard time reproducing the assignements in google colab.
Week 3 is just a scam since it says "go and read this blog" or "watch this video in another specialization". At that moment I simply felt robbed.
By Dimitry I•
Apr 14, 2021
Very superficial course, just like the rest in the specialization. Quizzes and assignments are a joke. Didn't want to give negative feedback at first, but now that I am doing course #4 in the specialization, which covers material I don't know much about (Attention), I've realized how bad these courses are. Very sad.
By Baurjan S•
Sep 26, 2020
Great Course as usual. Tried siamese models but got a very different results. Will need to study more on the conceptual side and implementation behind them. But overall, I am glad I touched LSTMs.
Aug 16, 2020
Learning about the Trax library and solving practical problems with the library was really interesting. Siamese network architecture was great thing to learn.
By Zoltan S•
Aug 1, 2020
This is an excellent course with some cutting edge material, and also an introduction to a new learning framework trax.
By Jerry C•
Oct 11, 2020
Great course! However the assignments are handholding too much step by step... I'd prefer the assignments to allow students to think more for themselves when implementing functions etc. (and only unhide hints or seek help on Slack when struggling for a long time)
By Swakkhar S•
Sep 23, 2020
First two courses were much better. It introduces trax, which is great. However, the materials of this course is already covered in the 5th course in the deep learning specialization. On the whole, great course, great efforts by the team.
By JJ Y•
Sep 26, 2020
Sequence models are heavy subjects and it would be unrealistic to expect a 4-week course to go into all the depths of RNN, GRN, LSTM etc. This course does a great job covering important types of neural networks and showing their applications. However, the labs and assignments could have done more in (a) helping us look a little deeper into the implementations of different NN building components, and (b) aligning better with the lecture videos.
Really Good examples: Week1 labs and assignment illustrate the implementations of some of the basic layer classes, and outline the overall flow of NN training with Trax. Week4 labs and assignment illustrate the implementation of the loss layer based on the unique triple loss function.
Not so Good examples: Week1 uses a whole video explaining gradient calculation in Trax. Yet there is no illustration of how it's integrated in backward propagation in Trax. Week2 videos and the labs/assignment are more disjoint. There is a video explaining the scan() function, but it does not show up in the assignment at all.
By Kota M•
Aug 21, 2021
Sadly, the quality of the material is much lower than the previous two courses. Assignment repeatedly asks to implement data generators with a lot of for-loops. We should focus more on the network architecture rather than python programming. That being said, the implementation is not good either. Learners would have to learn to program anyways.
By Patrick C•
Dec 23, 2020
Assignments are very difficult to complete because of inaccurate information (off by one on indices and other sloppy mistakes). You also don't learn much from them because almost all the code is already provided. It would be much better if they built up your understand from first principles instead of rushing through fill in the blank problems.
By Mostafa E•
Dec 13, 2020
The course did well in explaining the concepts of RNNs... but it may in fact have provided less knowledge than the NLP course in Deep Learning specialization.
I was looking forward to see more details on how translation works using LSTMs, go over some famous LSTM networks such as GNMT, and explain some accuracy measures such as the BLEU score.
By Artem R•
Dec 1, 2020
Course could be completed without watching videos - just by using hints and comments in assignments, videos are short and shallow, choice of Deep Learning framework (TRAX) is questionable - I won't use it in production.
Despite the course is 4 weeks long it could be accomplished in 4 days - I don't feel that it was worth the time.
By George L•
Mar 21, 2021
Compared with the Deep Learning specialization, this specialization was designed in a way that nobody can understand. Although the assignment could be easy at times, the point is being missed when people cannot really understand and learn. Bad teacher. Andrew Ng, please!
By Youran W•
Dec 4, 2020
All the assignments are extremely similar.
By Xinlong L•
Aug 22, 2021
I did not enjoy the course at all. It looks like the instructor is just reading materials rather than really teaching. He just focused on reading and did not explain anything. I took Andrew's deep learning specialization, and that course was really great. But I am so disappointed at this course. deeplearning.ai please do strict quality control on the courses otherwise it harms your brand
By Yanting H•
Oct 13, 2020
Oversimplified illustration of all core definitions and it is not reasonable from any sense to use trax instead of a popular framework like Tensorflow or Pytorch for the assignment. Also, the design of assignment is weak, you can barely learn anything from filling the blanks.
Nov 27, 2020
1) The course videos just throw out various nouns and you need to goole them to understand what they mean.
2) The assignments try their best to explain concepts in a way that often seems redundant.
By Emanuel D•
Jan 26, 2021
For me, it is very dissapointing, time is spent on irrelevant things, like python syntax and generators in first week. There are missing video tutorials on how to use Trax.
By Siddharth S•
Sep 19, 2021
Hard to follow explanations and TRAX absolutely made it super hard to learn and follow.
By Alistair M•
Feb 19, 2022
Superficial descriptions of the topics; quality definitely lacking
By Alice M•
Nov 15, 2020
No mentors were available or contactable during this course
By Nicolás E C R•
Dec 23, 2020