Welcome to this course on applied natural language processing in engineering. This course is designed to provide you with an in-depth understanding of NLP, a pivotal area of artificial intelligence that empowers computers to comprehend, interpret, and generate human language. Throughout this course, you will explore a wide range of topics, from fundamental NLP tasks like text classification and Named Entity Recognition (NER) to advanced techniques in neural machine translation and optimization methods critical for machine learning. We will delve into the complexities of teaching language to machines, addressing challenges like ambiguity, grammar, and cultural nuances. By the end of this part 1 course, you will have a foundational understanding of how modern NLP systems work - particularly those involving machine learning and deep learning. These topics will equip you to build, analyze and improve NLP systems across many applications.

Applied Natural Language Processing in Engineering Part 1
Sparen Sie mit 40% Rabatt auf 3 Monate Coursera Plus bei den Fähigkeiten, die Sie zum Strahlen bringen. Jetzt sparen
kurs ist nicht verfügbar in Deutsch (Deutschland)

Kompetenzen, die Sie erwerben
- Kategorie: Artificial Intelligence and Machine Learning (AI/ML)Artificial Intelligence and Machine Learning (AI/ML)
- Kategorie: Natural Language ProcessingNatural Language Processing
- Kategorie: Artificial Neural NetworksArtificial Neural Networks
- Kategorie: Machine LearningMachine Learning
- Kategorie: Deep LearningDeep Learning
- Kategorie: Machine Learning MethodsMachine Learning Methods
- Kategorie: Dependency AnalysisDependency Analysis
- Kategorie: EmbeddingsEmbeddings
- Kategorie: Supervised LearningSupervised Learning
- Kategorie: Model OptimizationModel Optimization
- Kategorie: Model EvaluationModel Evaluation
Wichtige Details

Zu Ihrem LinkedIn-Profil hinzufügen
22 Aufgaben
Erfahren Sie, wie Mitarbeiter führender Unternehmen gefragte Kompetenzen erwerben.

In diesem Kurs gibt es 7 Module
This module provides an in-depth exploration of Natural Language Processing (NLP), a crucial area of artificial intelligence that enables computers to understand, interpret, and generate human language. By combining computational linguistics with machine learning, NLP is applied in various technologies, from chatbots and sentiment analysis to machine translation and speech recognition. The module introduces fundamental NLP tasks such as text classification, Named Entity Recognition (NER), and neural machine translation, showcasing how these applications shape real-world interactions with AI. Additionally, it highlights the complexities of teaching language to machines, including handling ambiguity, grammar, and cultural nuances. Through the course, you will gain hands-on experience and knowledge about key techniques like word representation and distributional semantics, preparing them to solve language-related challenges in modern AI systems.
Das ist alles enthalten
4 Videos19 Lektüren2 Aufgaben1 App-Element
4 Videos•Insgesamt 6 Minuten
- Course Introduction•3 Minuten
- Meet Your Faculty•1 Minute
- Natural Language Processing (NLP)•1 Minute
- Representing the Meaning of a Word•1 Minute
19 Lektüren•Insgesamt 154 Minuten
- Course Introduction•2 Minuten
- Syllabus - Applied Natural Language Processing in Engineering Part 1•10 Minuten
- Academic Integrity•1 Minute
- Recommended Prior Knowledge•100 Minuten
- Week 1 Introduction•2 Minuten
- Introduction to NLP•5 Minuten
- Example: Chatbots•2 Minuten
- Example: Email Filtering•2 Minuten
- Example: Sentiment Analysis•3 Minuten
- Example: GPT - 3•3 Minuten
- Example: ChatGPT Capabilities•5 Minuten
- Natural Language Processing•1 Minute
- Funny Takes on Language Evolution•2 Minuten
- How Do We Represent the Meaning of a Word?•2 Minuten
- How Do We Have Usable Meaning in a Computer?•4 Minuten
- Words as Discrete Symbols•5 Minuten
- Representing Words by Their Context•2 Minuten
- Word Vectors•2 Minuten
- Final Thoughts on NLP•1 Minute
2 Aufgaben•Insgesamt 36 Minuten
- Assess Your Learning: What is NLP?•18 Minuten
- Assess Your Learning: Motivation•18 Minuten
1 App-Element•Insgesamt 15 Minuten
- Challenges of Teaching Language to AI•15 Minuten
This module focuses on optimization techniques critical for machine learning, particularly in natural language processing (NLP) tasks. It introduces Gradient Descent (GD), a fundamental algorithm used to minimize cost functions by iteratively adjusting model parameters. You’ll explore variants like Stochastic Gradient Descent (SGD) and Mini-Batch Gradient Descent to learn more about their efficiency in handling large datasets. Advanced methods such as Momentum and Adam are covered to give you insight on how to enhance convergence speed by smoothing updates and adapting learning rates. The module also covers second-order techniques like Newton’s Method and Quasi-Newton methods (e.g., BFGS), which leverage curvature information for more direct optimization, although they come with higher computational costs. Overall, this module emphasizes balancing efficiency, accuracy, and computational feasibility in optimizing machine learning models.
Das ist alles enthalten
2 Videos16 Lektüren3 Aufgaben
2 Videos•Insgesamt 8 Minuten
- Machine Learning and NLP•4 Minuten
- Optimization Techniques•4 Minuten
16 Lektüren•Insgesamt 82 Minuten
- Week 2 Overview•2 Minuten
- Machine Learning•2 Minuten
- Variations of Gradient Descent•2 Minuten
- Types of ML in NLP•6 Minuten
- What is a Model in NLP and How Does it Learn?•6 Minuten
- Understanding Cost Functions•2 Minuten
- Minimizing the Cost Function in NLP•10 Minuten
- Why Optimization Techniques Matter•1 Minute
- Why SGD Works•10 Minuten
- Jacobian Matrix & Hessian Matrix•5 Minuten
- Momentum•10 Minuten
- Newton's Methods•5 Minuten
- Quasi-Newton Methods•5 Minuten
- Root Mean Square Propagation (RMSProp)•5 Minuten
- Adaptive Moment Estimation (Adam)•10 Minuten
- Overall Challenges of Second-Order Optimization Techniques•1 Minute
3 Aufgaben•Insgesamt 81 Minuten
- Assess Your Learning: ML in NLP•18 Minuten
- Assess Your Learning: Optimization Techniques•18 Minuten
- Module 2 Quiz•45 Minuten
This module explores Named Entity Recognition (NER), a core task in Natural Language Processing (NLP) that identifies and classifies entities like people, locations, and organizations in text. We’ll begin by examining how logistic regression can be used to model NER as a binary classification problem, though this approach faces limitations with complexity and context capture. We’ll then transition to more advanced techniques, such as neural networks, which excel at handling the complex patterns and large-scale data that traditional models struggle with. Neural networks' ability to learn hierarchical features makes them ideal for NER tasks, as they can capture contextual information more effectively than simpler models. Throughout the module, we compare these methods and highlight how deep learning approaches such as Recurrent Neural Networks (RNNs) and transformers like BERT improve NER accuracy and scalability.
Das ist alles enthalten
2 Videos14 Lektüren3 Aufgaben1 App-Element
2 Videos•Insgesamt 4 Minuten
- Neural Networks Definitions•4 Minuten
- Network Propagation•0 Minuten
14 Lektüren•Insgesamt 89 Minuten
- Week 3 Overview•2 Minuten
- Neural Networks•2 Minuten
- Named Entity Recognition (NER)•5 Minuten
- NER as a Binary Regression Problem•5 Minuten
- Neural Network•5 Minuten
- Neural Network Structure•5 Minuten
- How Does a Neural Network Learn?•10 Minuten
- Mathematical Representation•20 Minuten
- Steps in Back Propagation Algorithm•5 Minuten
- Stochastic Gradient•5 Minuten
- Classification Tasks•5 Minuten
- Sequence-to-Sequence Tasks•5 Minuten
- Sequence Labeling Tasks•5 Minuten
- Regression Tasks & Divergence Measures•10 Minuten
3 Aufgaben•Insgesamt 81 Minuten
- Assess Your Learning: NER & Neural Networks•18 Minuten
- Assess Your Learning: Cost Functions•18 Minuten
- Module 3 Quiz•45 Minuten
1 App-Element•Insgesamt 10 Minuten
- Some Common Activation Functions•10 Minuten
The Word2Vec and GloVe models are popular word embedding techniques in Natural Language Processing (NLP), each offering unique advantages. Word2Vec, developed by Google, operates via two key models: Continuous Bag of Words (CBOW) and Skip-gram, focusing on predicting a word based on its context or vice versa (Word2Vec). The GloVe model, on the other hand, created by Stanford, combines count-based and predictive approaches by leveraging word co-occurrence matrices to learn word vectors (GloVe). Both models represent words in a high-dimensional vector space and capture semantic relationships. Word2Vec focuses on local contexts, learning efficiently from large datasets, while GloVe emphasizes global word co-occurrence patterns across the entire corpus, revealing deeper word associations. These embeddings enable tasks like analogy-solving, semantic similarity, and other linguistic computations, making them central to modern NLP applications.
Das ist alles enthalten
3 Videos29 Lektüren4 Aufgaben1 App-Element
3 Videos•Insgesamt 11 Minuten
- GLoVe Training Process•5 Minuten
- Word2Vec•4 Minuten
- Skip-Gram•2 Minuten
29 Lektüren•Insgesamt 267 Minuten
- Week 4 Overview•2 Minuten
- Introduction to GLoVe•5 Minuten
- Co-occurrence Matrix•5 Minuten
- Objective: Ratio of Co-occurrences•5 Minuten
- Calculating Probability Ratios•5 Minuten
- Symmetry and Linearity in GloVe•5 Minuten
- Minimizing the Cost Function and Optimizing Word Vectors•5 Minuten
- Optimization Process•10 Minuten
- Final Word Vectors•2 Minuten
- Implicit Properties in GloVe•5 Minuten
- GLoVe Introduction•2 Minuten
- What is Language Modeling?•5 Minuten
- Co-occurrence Matrix•5 Minuten
- Vector Representations for Word•3 Minuten
- Continuous Bag of Words (CBOW)•5 Minuten
- Mathematical Objectives•10 Minuten
- Mathematical Objectives 2•15 Minuten
- Limitations of CBOW•1 Minute
- Skip-Gram•15 Minuten
- Gradient Derivation•15 Minuten
- The Challenge of Training Skip-Gram•10 Minuten
- Binary Classification Perspective•10 Minuten
- Gradient of Negative Sampling Objective•10 Minuten
- Connecting Between Skip-Gram, Negative Sampling, and One Sampling•2 Minuten
- Skip-Gram with Negative Sampling Across All Words•10 Minuten
- Negative Sampling in Skip-Gram Model•10 Minuten
- Word2Vec Example•30 Minuten
- Word2Vec Worked Example •30 Minuten
- Word2Vec Example 2•30 Minuten
4 Aufgaben•Insgesamt 99 Minuten
- Assess Your Learning: GLoVe•18 Minuten
- Assess Your Learning: Word2Vec & CBOW•18 Minuten
- Assess Your Learning: Skip-Gram & Negative Sampling•18 Minuten
- Module 4 Quiz•45 Minuten
1 App-Element•Insgesamt 3 Minuten
- GloVe Training Process•3 Minuten
This module delves into the evaluation techniques for Natural Language Processing (NLP) models, focusing on both intrinsic and extrinsic evaluation methods. Intrinsic evaluation assesses the model's performance based on internal criteria, such as word embedding quality, parsing accuracy, and language model perplexity. In contrast, extrinsic evaluation measures the model's effectiveness in real-world applications, including tasks like machine translation, sentiment analysis, and named entity recognition. You’ll also learn more about key differences between these evaluation types, and the importance of context and application in determining a model's utility. Additionally, you’ll review specific metrics like cross-entropy loss, perplexity, BLEU, and ROUGE scores, providing a comprehensive understanding of how to evaluate and improve NLP models.
Das ist alles enthalten
9 Lektüren2 Aufgaben1 App-Element
9 Lektüren•Insgesamt 99 Minuten
- Week 5 Overview•2 Minuten
- General Concept of Evaluation (in NLP)•15 Minuten
- Key Differences Between Intrinsic and Extrinsic Evaluation•2 Minuten
- Cross-Entropy Loss - Intrinsic•10 Minuten
- Cross-Entropy and Learning from Incorrect Predictions•15 Minuten
- Perplexity - Intrinsic•15 Minuten
- Bilingual Evaluation Understudy Score (BLEU) - Extrinsic•15 Minuten
- Recall and Precision in Text Summarization or Translation•15 Minuten
- Recall-Oriented Understudy for Gisting Evaluation (ROUGE) - Extrinsic•10 Minuten
2 Aufgaben•Insgesamt 63 Minuten
- Assess Your Learning: NLP Model Evaluation•18 Minuten
- Module 5 Quiz•45 Minuten
1 App-Element•Insgesamt 10 Minuten
- Evaluation Techniques•10 Minuten
This module explores various techniques for topic modeling in natural language processing (NLP), focusing on methods like Latent Semantic Analysis (LSA), Non-Negative Matrix Factorization (NMF), and Latent Dirichlet Allocation (LDA). It begins with an introduction to matrix factorization and the importance of transforming textual data into numerical representations. You’ll delve into the mechanics of LSA and NMF, paying attention to their use of TF-IDF and Singular Value Decomposition (SVD) to uncover latent semantic structures. Additionally, you’ll review LDA's probabilistic approach to topic modeling, explaining its reliance on Dirichlet distributions and Bayesian inference to identify hidden topics within a corpus. Through detailed examples and mathematical explanations, the module provides a comprehensive understanding of how these techniques can be applied to extract meaningful topics from large text datasets.
Das ist alles enthalten
1 Video16 Lektüren4 Aufgaben1 App-Element
1 Video•Insgesamt 5 Minuten
- Topic Modeling•5 Minuten
16 Lektüren•Insgesamt 133 Minuten
- Week 6 Overview•2 Minuten
- Matrix Factorization•1 Minute
- Latent Semantic Analysis (LSA)•15 Minuten
- LSA Example•15 Minuten
- Topic Modeling Using Latent Semantic Analysis (LSA)•5 Minuten
- Dimensions and Applications•5 Minuten
- Non-Negative Matrix Factorization (NMF)•5 Minuten
- Operationalizing NMF•7 Minuten
- Numerical Example of NMF•15 Minuten
- Applications of NMF•2 Minuten
- Latent Dirichlet Allocation (LDA)•5 Minuten
- Defining the Problem and Key Assumptions•1 Minute
- Mathematical Model of LDA•10 Minuten
- Steps in LDA: Mathematical Explanation•15 Minuten
- Maximizing the Posterior Probability in LDA•15 Minuten
- Full Example•15 Minuten
4 Aufgaben•Insgesamt 99 Minuten
- Assess Your Learning: Latent Semantic Analysis•18 Minuten
- Assess Your Learning: Non-Negative Matrix Factorization•18 Minuten
- Assess Your Learning: Latent Dirichlet Allocation•18 Minuten
- Module 6 Quiz•45 Minuten
1 App-Element•Insgesamt 10 Minuten
- Recapping NMF & LDA•10 Minuten
This module delves into the essential techniques of syntactic and semantic parsing in natural language processing (NLP). You’ll begin with an exploration of linguistic structures, focusing on phrase structure and dependency structure, which are fundamental for understanding sentence syntax. Then you’ll review various parsing methods, including transition-based and graph-based dependency parsing, highlighting their respective advantages and challenges. Additionally, you’ll touch on neural transition-based parsing, which leverages neural networks for improved accuracy and efficiency. Finally, the module touches on semantic parsing, emphasizing its role in mapping sentences to formal representations of meaning, crucial for applications like dialogue systems and information extraction.
Das ist alles enthalten
2 Videos32 Lektüren4 Aufgaben
2 Videos•Insgesamt 5 Minuten
- Transition-Based and Graph Parsing Examples•2 Minuten
- Neural Advancements in Parsing: Dependency and Semantics•3 Minuten
32 Lektüren•Insgesamt 160 Minuten
- Week 7 Overview•2 Minuten
- Introduction to Dependency Parsing•5 Minuten
- What is a Constituent?•5 Minuten
- Pseudoclefting & Sentence Fragments•5 Minuten
- Substitution of Well-formed Sentences•5 Minuten
- Naming the Other Phases•5 Minuten
- Immediate Domination, Domination, & Feminine Kinship Relations•5 Minuten
- Syntactic Parsing - Phrase Structure•10 Minuten
- Phrase Structure•3 Minuten
- Phrase Structure Rules & Explanation•2 Minuten
- Building the Tree•2 Minuten
- Summary•5 Minuten
- Dependency Structure•2 Minuten
- Key Concepts & Example•10 Minuten
- Advantages of Dependency Structure•5 Minuten
- What is Transition-based Parsing?•2 Minuten
- The Greedy Transition-Based Method•10 Minuten
- Type 2: Shift-Reduce Parser•10 Minuten
- Introduction•5 Minuten
- Key Concepts & Configuration•10 Minuten
- Advantages, Challenges & Considerations•5 Minuten
- Introduction•2 Minuten
- Key Concepts of Graph-Based Dependency Parsing•5 Minuten
- Worked Example•10 Minuten
- Advantages and Disadvantages•2 Minuten
- Transition-Based vs Graph-Based Parsing•1 Minute
- Semantic Parsing•5 Minuten
- Example 1•5 Minuten
- Example 2•5 Minuten
- Example 3•5 Minuten
- Example 4•5 Minuten
- Congratulations! •2 Minuten
4 Aufgaben•Insgesamt 135 Minuten
- Assess Your Learning: Constituents & Detection of Sentence Parts•30 Minuten
- Assess Your Learning: Phrase and Dependency Structure•30 Minuten
- Assess Your Learning: Parsing•30 Minuten
- Module 7 Quiz•45 Minuten
Dozent

von

von

Founded in 1898, Northeastern is a global research university with a distinctive, experience-driven approach to education and discovery. The university is a leader in experiential learning, powered by the world’s most far-reaching cooperative education program. The spirit of collaboration guides a use-inspired research enterprise focused on solving global challenges in health, security, and sustainability.
Mehr von Machine Learning entdecken
Status: VorschauVorschauNNortheastern University
Kurs
Status: VorschauVorschauNNortheastern University
Kurs
Status: VorschauVorschauBBirla Institute of Technology & Science, Pilani
Kurs
Status: Kostenloser TestzeitraumKostenloser TestzeitraumSpezialisierung
Warum entscheiden sich Menschen für Coursera für ihre Karriere?

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.
Häufig gestellte Fragen
To access the course materials, assignments and to earn a Certificate, you will need to purchase the Certificate experience when you enroll in a course. You can try a Free Trial instead, or apply for Financial Aid. The course may offer 'Full Course, No Certificate' instead. This option lets you see all course materials, submit required assessments, and get a final grade. This also means that you will not be able to purchase a Certificate experience.
When you purchase a Certificate you get access to all course materials, including graded assignments. Upon completing the course, your electronic Certificate will be added to your Accomplishments page - from there, you can print your Certificate or add it to your LinkedIn profile.
Yes. In select learning programs, you can apply for financial aid or a scholarship if you can’t afford the enrollment fee. If fin aid or scholarship is available for your learning program selection, you’ll find a link to apply on the description page.
Weitere Fragen
Finanzielle Unterstützung verfügbar,


