Chevron Left
Back to Generative AI and LLMs: Architecture and Data Preparation

Learner Reviews & Feedback for Generative AI and LLMs: Architecture and Data Preparation by IBM

4.6
stars
281 ratings

About the Course

Ready to explore the exciting world of generative AI and large language models (LLMs)? This IBM course, part of the Generative AI Engineering Essentials with LLMs Professional Certificate, gives you practical skills to harness AI to transform industries. Designed for data scientists, ML engineers, and AI enthusiasts, you’ll learn to differentiate between various generative AI architectures and models, such as recurrent neural networks (RNNs), transformers, generative adversarial networks (GANs), variational autoencoders (VAEs), and diffusion models. You’ll also discover how LLMs, such as generative pretrained transformers (GPT) and bidirectional encoder representations from transformers (BERT), power real-world language tasks. Get hands-on with tokenization techniques using NLTK, spaCy, and Hugging Face, and build efficient data pipelines with PyTorch data loaders to prepare models for training. A basic understanding of Python, PyTorch, and familiarity with machine learning and neural networks are helpful but not mandatory. Enroll today and get ready to launch your journey into generative AI!...

Top reviews

VK

Oct 18, 2024

I am pretty much new to NLP data preparation. However this course made me comfortable with Date preparation activities.

JR

Mar 1, 2025

Was waiting for a course like this for a long time. Very happy with it. Library installation on labs seems a bit slow

Filter by:

51 - 62 of 62 Reviews for Generative AI and LLMs: Architecture and Data Preparation

By Justin R

Oct 27, 2024

The content in the lectures is complex but the slides are not made available to download. Also the Cheat Sheets and other similar materials are presented in weird "windows" that also do not make them available for download. This is a first for me in a Coursera course and I'm find it not very conducive to learning. These material should be easily available. Not certain I will complete the full Specialization if the materials are not made available.

By Yongchang L

Jul 14, 2024

I found the course on LLMs to be a solid introduction, particularly appreciating the cheatsheet and experiments included. However, the requirement to purchase a $49 certificate to complete the course felt excessive. The course producer should learn from many other courses on Coursera, completing the course should be free with the option to purchase the certificate as an add-on.

By Uday T

Aug 25, 2025

There are some lab works where training was consuming enormous time, due to which I was not able to complete that lab. This needs to be resolved so that other learner would get benefit out of this.

By fidel m

Feb 9, 2025

so much of reading material and so less of actual videos. the speaking voice in video is also in a rush

By Jimmy M

Mar 22, 2025

Content was decent for intermediate intro, but all labs were broken. Easy 4 stars if they worked.

By Sailesh M

Jan 17, 2025

Labs don't work as torchtext is deprecated and doesn't run on Python 3.12 kernel

By Jochen G

Mar 20, 2025

The course is not well maintained, and rather superficial

By AYA A

Jul 29, 2025

need live coding labs to test out the scripts

By Jonas K

Aug 26, 2025

It basically just consists of Jupyter Notebooks without much explanation. I tried to understand the content and implement it on my own because the exercises did not help me understand it at all. As a result, it took me much longer than expected to work through the content, which disrupted my schedule. I would not recommend this course. Unfortunately, I am not aware of any better alternatives for learning about this topic.

By Fan Y

Oct 15, 2024

Tokenizer & dataloader are quite important parts but I am surprised by how shallow they are touched and how easy are the quiz questions.

By Ethan K

Aug 27, 2025

I would not recommend this course. It is basically llm-slop being used to explain llms.

By Serhii S

Nov 8, 2024

very superficial