Please Enable the Desktop mode for better view experience

SPEECH AND NATURAL LANGUAGE PROCESSING Previous Year 2024

Speech and NLP Previous Year – Part 1 – Digitaload Academy

SPEECH AND NATURAL LANGUAGE PROCESSING
B.Tech. CSE (AI & ML)

Exam May-2024

View All Categories 📂

Section-A (2 Marks Each)

(a) What is Natural Language Processing? 2 Marks

Definition

  • A field of computer science and AI.
  • Focuses on enabling computers to understand and process human language.
  • Includes analyzing text and speech.
  • Aims for communication between computers and humans using natural language.

Core Goal

  • Machines understand human language.
  • Bridge computer-human language gap.

(b) How NLP is used in machine translation system like Google Translate? 2 Marks

NLP in Machine Translation

  • Analyzes the source language text (tokenization, parsing, semantic understanding).
  • Transforms the meaning or structure into a representation.
  • Generates grammatically correct sentences in the target language.
  • Includes tasks like word sense disambiguation, parsing, text generation.

Role

  • Source text analysis.
  • Meaning transformation.
  • Target language generation.
  • Enables cross-language conversion.

(c) How do machine translation and database interfaces contribute to breaking down language barriers? 2 Marks

Breaking Language Barriers

  • Machine Translation: Translates text or speech across different human languages automatically.
  • Database Interfaces (via NLP): Allows users to query databases using natural language instead of formal query languages (like SQL).
  • Both enable information access and communication despite differences in human language or technical interfaces.

Contribution

  • Machine Translation: Between human languages.
  • Database Interfaces: Human language to database language.
  • Increase access to information/communication.

(d) What is the role of grammars in NLP? 2 Marks

Role of Grammars in NLP

  • Define the rules that govern the structure of sentences in a language.
  • Help in analyzing the syntax (grammatical structure) of sentences.
  • Represent permissible combinations of words and phrases.
  • Used in parsing to build structural representations of sentences.

Purpose

  • Model language structure.
  • Analyze sentence syntax.
  • Used by parsing algorithms.

(e) How are auxiliary verbs represented in context-free grammars? 2 Marks

Auxiliary Verbs in CFGs

  • Auxiliary verbs (like ‘be’, ‘have’, ‘do’, ‘can’, ‘will’) are typically treated as distinct categories in Context-Free Grammars (CFGs).
  • They are often represented by their own non-terminal symbol, for instance, ‘Aux’ or ‘Modal’ if it’s a modal auxiliary.
  • Production rules specify where these auxiliary verbs can appear in relation to other verb phrases (VPs) and sentence structures (S).
  • Example Rule: S → NP Aux VP (A Sentence can be a Noun Phrase followed by an Auxiliary Verb and a Verb Phrase).

Representation

  • Dedicated non-terminal symbol (e.g., ‘Aux’).
  • Used in production rules for sentence structure.

Section-A (Short Answer Questions – 4 Marks Each)

(f) What are some ethical consideration in the development and deployment of NLP systems? 4 Marks

Definition: Ethics in NLP

Ethical considerations in NLP involve the moral principles, societal impacts, and potential harms associated with the development and deployment of language processing technologies, addressing issues like bias, privacy, fairness, and transparency.

Ethical Considerations in NLP Systems

  • Bias: NLP models trained on biased data (reflecting societal prejudices like race, gender) can perpetuate or amplify these biases in their outputs (e.g., in translation, text generation, sentiment analysis).
  • Privacy: Processing sensitive personal text data (emails, messages, voice recordings) raises concerns about data security, consent, and potential misuse of information.
  • Fairness and Discrimination: Biased systems can lead to unfair treatment or discrimination (e.g., in hiring tools, loan applications, content moderation).
  • Transparency and Explainability: Understanding *why* an NLP system produces a certain output (especially from complex deep learning models) can be difficult, raising issues of accountability and trust.
  • Misinformation and Disinformation: NLP can be used to generate realistic-sounding fake news or propaganda at scale, making it hard to distinguish from truthful content.
  • Safety and Malicious Use: NLP systems could potentially be used for malicious purposes (e.g., targeted harassment, manipulation, phishing attacks).
  • Accessibility: Ensuring NLP systems are accessible and perform equally well across different languages, dialects, and communication styles is an ethical challenge.
  • Impact on Jobs: Automation of language-based tasks may lead to job displacement.

Key Issues

  • Bias in data/models.
  • Privacy concerns (data handling).
  • Fairness and discrimination risk.
  • Explainability/Transparency.
  • Misinformation/Misuse potential.

Impacts

  • Content moderation.
  • Recruitment tools.
  • Customer service bots.
  • Legal/medical text analysis.

(g) What role does NLP play in sentiment analysis of social media data? 4 Marks

Definition: Sentiment Analysis

Sentiment analysis (or opinion mining) is the computational technique of determining the emotional tone, attitude, or opinion expressed in text or speech. It often classifies sentiment as positive, negative, or neutral.

Role of NLP in Sentiment Analysis of Social Media Data

NLP is fundamental to sentiment analysis, especially for unstructured and informal text found on social media (Tweets, posts, comments):

  • Text Preprocessing:
    • Tokenization (splitting text into words/tokens).
    • Normalization (handling variations, slang, emojis, hashtags, removing noise like URLs/mentions).
    • Part-of-Speech tagging, Lemmatization/Stemming.
  • Feature Extraction:
    • Identifying words and phrases (unigrams, bigrams, trigrams) carrying sentiment (e.g., “amazing,” “terrible,” “not bad”).
    • Using techniques like TF-IDF or word embeddings to represent text numerically.
    • Identifying sentiment lexicon (lists of words with polarity).
  • Handling Nuances:
    • Identifying negations (“not happy”), sarcasm, irony, emojis, intensification (“very good”).
    • Handling context-dependent sentiment (“This move is amazing” vs “This amazing failure”).
  • Classification:
    • Using NLP features as input to machine learning models (Naïve Bayes, SVM, Neural Networks) or rule-based systems to classify text sentiment.
  • Topic Modeling: NLP helps link sentiments to specific topics discussed (e.g., positive sentiment about “camera” vs negative about “battery” in phone reviews).

NLP techniques allow algorithms to move beyond simple keyword matching to understand the linguistic complexities of social media text and extract meaningful sentiment.

Core NLP Tasks Involved

  • Preprocessing text.
  • Extracting sentiment features (words, phrases).
  • Handling linguistic nuances (negation, sarcasm).
  • Feeding features to classification models.

Used For

  • Brand monitoring.
  • Market research.
  • Understanding public opinion.
  • Customer feedback analysis.
  • Detecting trends.

(h) How do semantic representations help in capturing the meaning of words and sentences in NLP? 4 Marks

Definition: Semantics and Semantic Representation

Semantics in NLP is the study of meaning. Semantic representation refers to methods and formalisms used to capture the meaning of words, phrases, sentences, and larger texts in a structured, computational format that NLP systems can process and understand.

Helping Capture Meaning

Semantic representations go beyond the surface form (syntax) to encode the intended meaning. They help in:

  • Word Sense Disambiguation (WSD):
    • Capturing that words can have multiple meanings based on context (e.g., “bank” as financial vs river bank).
    • Representations help assign the correct sense by encoding relationships with other words in the sentence or context.
  • Understanding Relationships:
    • Representing relationships between entities (who did what to whom, spatial relations, temporal relations).
    • Semantic roles (agent, patient, instrument) and thematic relations are part of this. Example: In “John ate the apple,” semantic roles capture that John is the *agent* and apple is the *patient*.
  • Representing Sentence Meaning:
    • Encoding the overall meaning of a sentence independent of its grammatical structure.
    • Example formalisms: Semantic Networks, Frame-Based Semantics, Logical Forms (e.g., First-Order Logic representations).
    • They help in querying, reasoning, and tasks like question answering or machine translation.
  • Handling Synonymy/Paraphrase: Recognizing that different words or sentence structures can have the same or similar meaning.
  • Cross-Lingual Understanding: For machine translation, translating meaning representations can be more effective than word-for-word or syntactic translation.
  • Knowledge Integration: Semantic representations allow linking linguistic structures to formal knowledge bases.

Role

  • Move from structure (syntax) to meaning.
  • Disambiguate word senses.
  • Represent relationships (semantic roles).
  • Capture sentence/text meaning.
  • Support reasoning & understanding.

Example Formalism

  • Predicate Logic: Represents “John is happy” as happy(John). “John likes Mary” as likes(John, Mary).
  • Word Embeddings: Represent word meaning as numerical vectors in space where similar words are close together (king – man + woman ≈ queen).
  • Frames: Structures representing typical scenarios (e.g., a ‘CommercialTransaction’ frame with slots for Seller, Buyer, Goods, Price).

(i) How do context-free grammars (CFGs) represent the syntactic structure of natural language sentences? 4 Marks

Definition: Context-Free Grammar (CFG)

A Context-Free Grammar (CFG) is a formal grammar used to describe the structure of a language. In NLP, CFGs model the syntactic structure of natural language sentences using rules that define how categories of words and phrases can be combined to form larger structures.

Components of a CFG

  • V: A finite set of **non-terminal symbols** (variables) representing grammatical categories (e.g., S = Sentence, NP = Noun Phrase, VP = Verb Phrase, Adj = Adjective).
  • Σ: A finite set of **terminal symbols** (tokens) representing the words (vocabulary) of the language (e.g., ‘cat’, ‘the’, ‘sleeps’).
  • R: A finite set of **production rules** (rewrite rules) of the form A → β, where A is a non-terminal and β is a string of terminals and/or non-terminals. Rules show how a non-terminal can be broken down into smaller constituents. Example: S → NP VP.
  • S ∈ V: A distinguished **start symbol** (usually S for Sentence) representing the root of the grammatical structure.

Representing Syntactic Structure

CFGs represent syntactic structure by generating a **parse tree** for a sentence. The parse tree shows the hierarchical breakdown of the sentence according to the grammar’s production rules.

  • Starting from the start symbol (S), apply production rules to rewrite non-terminals.
  • Continue expanding non-terminals until only terminals (words) remain at the leaves of the tree.
  • The resulting tree structure shows how the sentence is composed of smaller grammatical units and their relationships.

Example Grammar Rules:

S → NP VP
NP → Det Noun
VP → Verb NP
VP → Verb
Det → the | a
Noun → cat | dog
Verb → sleeps | sees
                         

Using these rules, the sentence “the cat sleeps” can be parsed into a tree:

         S
        / \
       NP  VP
      / \   |
     Det Noun Verb
    /   |     |
   the cat sleeps
                          

The parse tree visually represents that “the cat” is a Noun Phrase (Det + Noun), “sleeps” is a Verb Phrase, and NP and VP combine to form a Sentence (S). This captures the sentence’s grammatical structure.

Role in NLP

  • Syntactic analysis basis.
  • Generate parse trees.
  • Model sentence composition rules.
  • Foundation for parsing algorithms.

CFG Properties

  • Rules are context-independent (LHS is single non-terminal).
  • Finite rules describe infinite sentences.

(j) Explain examples of different types of ambiguity that commonly occur in natural language. 4 Marks

Definition: Ambiguity in Natural Language

Ambiguity is the property of language where a single word, phrase, sentence, or text can have multiple possible interpretations or meanings. It is a pervasive phenomenon in human language that NLP systems must be able to identify and resolve to achieve correct understanding.

Examples of Different Types of Ambiguity

Ambiguity can manifest at various levels:

  • 1. Lexical Ambiguity:
    • Definition: A single word has multiple meanings (polysemy or homonymy).
    • Example: “bank” (financial institution vs side of river). The sentence “I went to the bank” is lexically ambiguous without context.
    • Example: “bat” (flying mammal vs sports equipment). “I saw a bat.”
  • 2. Syntactic Ambiguity (Structural Ambiguity):
    • Definition: A sentence has multiple possible grammatical structures (parse trees), leading to different interpretations of the relationships between words.
    • Example: “I saw the man with the telescope.” (Who had the telescope? I or the man?) – Could mean (I saw the man) (with the telescope) OR (I saw (the man with the telescope)).
    • Example: “The chicken is ready to eat.” (Is the chicken cooked, or does the chicken need food?).
  • 3. Semantic Ambiguity:
    • Definition: When the meaning of a sentence is unclear due to the ambiguity of a phrase or the interaction of word meanings, even if the syntax is clear. Can arise from quantifier scope, or vague predicates.
    • Example: “Every student didn’t pass the exam.” (Does NO student pass? Or did not *every* student pass, meaning *some* students passed but *not all*?). This is scope ambiguity (scope of “every” vs “didn’t”).
    • Example: “The car hit the pole while it was moving.” (What was moving? The car or the pole?) – Anaphora resolution needed, potentially semantic/referential ambiguity.
  • 4. Pragmatic Ambiguity:
    • Definition: When the intended meaning of a sentence depends on context and shared knowledge between speakers, and could be interpreted literally or non-literally (e.g., sarcasm, indirect speech).
    • Example: “Can you pass the salt?” (Literal interpretation: are you *physically able*? Intended meaning: *Please* pass the salt). Ambiguous if context doesn’t make intention clear.
    • Example: Speaker says “Lovely weather” during a storm. (Literal: good weather. Pragmatic: sarcasm meaning bad weather).
  • 5. Anaphora Resolution Ambiguity: (Often seen as part of Semantic/Discourse)
    • Definition: Ambiguity in determining what a pronoun or referential expression refers to.
    • Example: “John told Bill that he was in trouble.” (Who was ‘he’? John or Bill?).

Types Summary

  • Lexical: Single word has multiple meanings.
  • Syntactic: Sentence has multiple parse structures.
  • Semantic: Meaning unclear from word interaction/scope.
  • Pragmatic: Meaning depends on context/intention.
  • Anaphora: Pronoun reference unclear.

Challenge for NLP

  • Requires context to resolve.
  • Leads to multiple interpretations.
  • NLP systems need explicit rules or probabilities.
  • Critical for understanding.

Section-B (8 Marks Each) – Detailed Answers

Describe some emerging applications of NLP in fields such as cyber security, climate science, and urban planning. 8 Marks

Definition: Emerging Applications of NLP

Emerging applications of Natural Language Processing (NLP) refer to innovative ways NLP is being applied beyond traditional areas (like search and translation) to solve complex problems in new or rapidly evolving fields, leveraging its ability to process and understand text/speech data.

Key Concepts

  • Extending NLP techniques to specialized domains.
  • Analyzing domain-specific text or speech data.
  • Extracting insights, patterns, or actions relevant to the field.
  • Often requires adapting general NLP models to the specific jargon, context, and data characteristics of the new domain.

NLP in Cyber Security (Examples)

NLP helps security professionals analyze vast amounts of text data related to threats and vulnerabilities:

  • Threat Intelligence Analysis: Analyzing security blogs, forums, reports, social media, and dark web content to identify emerging threats, attacker tactics, techniques, and procedures (TTPs).
  • Security Incident Response: Automatically processing logs, alerts, and communication data (e.g., internal chat messages during an incident) to identify the scope, cause, and impact of a security breach.
  • Phishing Detection: Analyzing the language style, content, and urgency in emails to identify characteristics of phishing attacks more effectively than simple keyword matching.
  • Sentiment Analysis in Cybersecurity: Monitoring online discussions about a company’s security posture or recent incidents to gauge public reaction and potential vulnerability exposure.
  • Automated Vulnerability Reports Analysis: Extracting structured information about security vulnerabilities from unstructured text reports.

NLP in Climate Science (Examples)

NLP aids climate scientists in processing and analyzing large bodies of scientific literature and reports related to climate change:

  • Extracting Information from Climate Literature: Automatically identifying key findings, methodologies, datasets, and uncertainties mentioned in thousands of research papers to synthesize knowledge.
  • Analyzing Climate Policy Documents: Processing policy papers, government reports, and meeting transcripts to track policy positions, commitments, and outcomes related to climate change across different regions or organizations.
  • Monitoring Public Discourse on Climate: Analyzing news articles, social media trends, and public comments related to climate events, policies, and public opinion.
  • Information Extraction from Historical Records: Analyzing historical documents or logs that mention climate events to build longer-term climate datasets.

NLP in Urban Planning (Examples)

NLP can help urban planners process textual data reflecting citizen needs, urban conditions, and plan descriptions:

  • Analyzing Citizen Feedback: Processing public comments submitted through various channels (online forums, public meetings transcripts, social media) regarding infrastructure projects, local issues, transportation problems, or community needs to identify priorities and sentiments.
  • Monitoring Urban Conditions: Analyzing text descriptions (e.g., from social media or maintenance reports) mentioning issues like potholes, broken streetlights, waste management problems, or traffic bottlenecks to identify specific locations and types of urban deficiencies.
  • Analyzing Planning Documents: Processing complex urban planning proposals, zoning laws, environmental impact statements, and regulatory documents to extract key constraints, requirements, and implications.
  • Geolocating Information: Combining NLP with location information to map textual mentions of issues or feedback to specific geographical areas within a city.

Key Capabilities Used

  • Information Extraction.
  • Sentiment Analysis.
  • Text Classification.
  • Named Entity Recognition.
  • Topic Modeling.
  • Text Generation.

Examples in Fields

  • Cybersecurity: Threat analysis, phishing detection.
  • Climate Science: Literature review, policy analysis.
  • Urban Planning: Citizen feedback, condition monitoring.

OR

How do chatbots utilize natural language understanding to interact with users effectively? 8 Marks

Definition: Chatbots and Natural Language Understanding (NLU)

Chatbots are AI programs designed to simulate conversational interaction with human users, primarily through text or voice. Natural Language Understanding (NLU) is a subset of NLP focused specifically on reading and interpreting human language input – enabling the machine to “understand” what the user means, beyond just recognizing the words.

Key Concepts of NLU

  • Deals with ambiguities (word sense, syntax, semantics).
  • Extracts meaning and intent from user utterances.
  • Identifies key entities and relationships mentioned.
  • Translates human language into structured data or commands that a machine can process.

How Chatbots Utilize NLU for Effective Interaction

NLU is the core engine that powers a chatbot’s ability to understand user input and generate relevant responses, enabling seemingly natural conversation:

  • 1. Understanding User Intent:
    • NLU analyzes the user’s input sentence to determine what task they are trying to perform or what information they are seeking (e.g., “I want to book a flight” -> intent = BookFlight; “What’s the weather?” -> intent = GetWeather).
    • Identifying the correct intent is the first step in knowing how to respond.
  • 2. Extracting Entities (Named Entity Recognition – NER):
    • NLU identifies and extracts key pieces of information relevant to the user’s intent from the sentence.
    • Example: “Book a flight from **New York** to **London** next **Tuesday**” -> NLU extracts entities like Origin=New York, Destination=London, Date=Next Tuesday.
    • These entities provide the necessary details to fulfill the user’s request.
  • 3. Handling Utterance Variations and Ambiguity:
    • NLU deals with different ways users might phrase the same intent or provide the same information. (e.g., “flight booking” vs “book a flight” vs “reserve seat”).
    • It also tries to resolve ambiguities (e.g., using context or asking clarifying questions if multiple interpretations exist).
  • 4. Semantic Role Labeling (SLR):
    • NLU can identify the semantic roles of different parts of the sentence (e.g., “I” is the booker, “flight” is the thing being booked, “from X” is the source, “to Y” is the destination).
    • This helps understand the structure of the user’s request more accurately.
  • 5. Sentiment Analysis (Optional but adds effectiveness):
    • NLU can analyze the sentiment (positive/negative/neutral) of the user’s message.
    • Helps the chatbot respond appropriately (e.g., be empathetic if the user expresses frustration).
  • Mapping to Actions: Once intent and entities are understood, the chatbot maps this semantic understanding to specific actions or API calls to fulfill the request (e.g., call a flight booking API with extracted location/date).

Effective chatbots require a robust NLU engine that can accurately parse, interpret, and understand the nuanced and varied ways humans communicate naturally.

NLU Processes Used

  • Intent Identification.
  • Entity Extraction (NER).
  • Handling Ambiguity/Variations.
  • Semantic Role Labeling.
  • (Optional) Sentiment Analysis.
  • Map to machine commands.

Chatbot Examples

  • Customer service bots.
  • Virtual assistants (Siri, Alexa – for commands).
  • Online support bots.
  • Conversational AI.

Effectiveness Benefits

  • Understand user needs.
  • Provide relevant responses.
  • Handle natural language input.
  • Improve user experience.

NLU Challenges for Chatbots

  • Handling complex queries.
  • Maintaining context.
  • Resolving ambiguities.
  • Understanding slang/accents.
  • Domain-specific knowledge needed.

How does knowledge representation relate to NLP, and why is it important. 8 Marks

Definition: Knowledge Representation (KR)

Knowledge Representation (KR) is the study of how to think about and formally describe knowledge in a way that can be used by computer systems. It involves representing facts, objects, concepts, and relationships from the real world in a structured and explicit format amenable to machine manipulation and reasoning.

Definition: Natural Language Processing (NLP)

NLP enables machines to understand and process human language (text and speech).

Relation of Knowledge Representation to NLP

KR and NLP are deeply intertwined, especially for advanced language understanding and generation tasks. KR is essential for NLP in several ways:

  • Understanding Meaning (Semantics):
    • Capturing the meaning of language often requires mapping linguistic structures to underlying knowledge representations.
    • Representations like semantic networks, frames, ontologies, or logical forms provide structures to represent the concepts and relationships expressed in language. NLP extracts these meanings and fits them into the KR structures.
    • Example: Understanding “John is taller than Mary” requires representing entities (John, Mary) and a relation (tallerThan) within a KR framework, using height properties. NLP parses the sentence and maps it to tallerThan(John, Mary) in a logical KR.
  • Disambiguation:
    • Knowledge about the world (stored in KR) helps resolve ambiguities in language.
    • Example: Understanding which “bank” is meant often relies on knowledge about typical actions or locations associated with financial banks versus river banks (knowledge accessible via KR).
  • Reasoning and Inference:
    • Going beyond surface understanding to draw conclusions from text often requires using reasoning mechanisms applied to a KR of the information in the text and background knowledge.
    • Example: From “Socrates is a man” and knowledge (KR) “All men are mortal”, an AI needs to infer “Socrates is mortal.” NLP extracts “Socrates is a man” -> KR fact. Reasoning engine uses KR rule -> infers new KR fact.
  • Question Answering (QA):
    • To answer questions from text, NLP understands the question, uses KR to represent the query meaning and the information in the text.
    • QA system uses reasoning on the KR to find the answer.
  • Natural Language Generation (NLG):
    • Generating coherent and meaningful language from a KR.
    • NLG system receives a piece of knowledge (in KR format) and generates a natural language sentence expressing that knowledge.
  • Building Chatbots/Virtual Assistants: Understanding user requests requires mapping their language to intentions, entities, and states represented in a system’s KR.

Importance of Knowledge Representation in NLP

  • Enabling Deeper Understanding: Moves NLP beyond syntax and statistics to actually capture and process the meaning and implications of language.
  • Supporting Reasoning: Allows AI systems to use information gleaned from text to reason about the world and draw conclusions.
  • Handling Complex Tasks: Crucial for applications requiring more than surface-level analysis, such as sophisticated question answering, dialogue systems, and knowledge extraction.
  • Reducing Data Dependence: While modern NLP uses data heavily, integrating with KR can potentially reduce reliance on *just* vast quantities of text by adding explicit world knowledge.
  • Structure for Language: KR provides formal structures and schemas for representing the information content of language in a machine-processable format.
  • Resolving Ambiguities: Knowledge helps in making informed choices when language is ambiguous.

KR Basics

  • Formal description of knowledge.
  • Represents facts, concepts, relations.
  • Used for reasoning.

Relation

  • KR needed for deep meaning.
  • Helps NLP understand semantics.
  • Enables reasoning from text.
  • KR supports NLG.

Importance

  • Deeper language understanding.
  • Supports advanced AI tasks.
  • Helps disambiguate.
  • Facilitates reasoning.

KR Types

  • Semantic Networks.
  • Frames.
  • Ontologies.
  • Logical forms (Predicate Logic).

OR

How does semantics play a crucial role in improving the accuracy of machine translation system? 8 Marks

Definition: Semantics in Machine Translation

Semantics is the study of meaning in language. In machine translation (MT), incorporating semantic analysis means going beyond just translating words and syntax from a source language to a target language, focusing instead on capturing and transferring the actual *meaning*.

Key Concepts

  • Translation can be challenging due to linguistic differences (syntax, word order, vocabulary, idioms) and ambiguity.
  • Meaning can be the same even if sentence structure is different across languages.
  • Understanding the meaning allows for more flexible and accurate rephrasing in the target language.
  • Traditional rule-based or even early statistical MT often struggled with capturing full meaning.

Role of Semantics in Improving MT Accuracy

Semantics plays a crucial role in making machine translation more accurate and human-like:

  • 1. Word Sense Disambiguation (WSD):
    • Problem: Many words have multiple meanings (polysemy). Translating the wrong sense is a major source of MT error. (e.g., translating “bank”).
    • How Semantics Helps: Semantic analysis (using context, relationships with other words) helps identify the correct sense of a word *before* translating it, leading to choosing the correct corresponding word in the target language.
  • 2. Capturing Semantic Relationships:
    • Problem: Syntactic structures might differ, but the semantic relationship (e.g., who did what to whom) needs to be preserved. (e.g., Active vs Passive voice may have different structure but same core meaning).
    • How Semantics Helps: Analyzing the semantic roles (agent, patient, theme) and predicate-argument structures captures these relationships. The translation system then uses this semantic structure to generate a syntactically correct sentence in the target language that conveys the same meaning, even if the word order changes.
  • 3. Handling Paraphrase and Synonymy:
    • Problem: Different source sentences can have the same meaning, or there might be multiple valid ways to express the same meaning in the target language.
    • How Semantics Helps: Meaning representations (capturing semantics) can identify that different surface forms express similar meaning, allowing the system to choose appropriate synonyms or paraphrases in the target language based on context or style.
  • 4. Addressing Idioms and Figurative Language:
    • Problem: Idioms or figurative language have meanings not derivable from the literal words (e.g., “kick the bucket” doesn’t literally mean strike a bucket). Literal translation fails.
    • How Semantics Helps: Semantic understanding and external knowledge (linked via semantics) can help recognize idioms as single units and translate their *meaning* rather than the individual words.
  • 5. Improving Coherence and Flow:
    • Problem: Translating sentence by sentence without considering relationships across sentences can lead to awkward or incoherent translation of longer texts.
    • How Semantics Helps: Analyzing semantic dependencies and coreferences (e.g., pronoun resolution) across sentences ensures that the translation maintains consistent meaning and flow in the target text (discourse semantics).

By moving beyond direct word-to-word or structure-to-structure mapping and focusing on semantic representation and analysis, MT systems (especially modern neural MT models that learn semantic patterns) can produce translations that are not only grammatically correct but also semantically accurate and natural-sounding.

Semantics Helps With

  • Word sense ambiguity.
  • Semantic relationships.
  • Paraphrasing/Synonymy.
  • Idioms/Figurative language.
  • Text coherence.

In MT Systems

  • Analysis of source text.
  • Intermediate representation.
  • Generation of target text.

Improves MT By

  • Increased accuracy.
  • More natural phrasing.
  • Better handling of ambiguity/idioms.
  • Enhanced meaning transfer.

Challenges

  • Semantics is complex.
  • Requires large datasets or sophisticated knowledge.
  • Developing robust semantic parsers is hard.

Describe the role of parsing in machine translation systems, particularly in generating grammatically correct target language sentences? 8 Marks

Definition: Parsing in NLP

Parsing (or syntactic analysis) is the process in NLP that analyzes a sentence to determine its grammatical structure according to a formal grammar. It involves breaking down a linear sequence of words into a hierarchical structure, such as a parse tree.

Definition: Machine Translation (MT)

Machine translation is the task of automatically translating text or speech from one human language (source language) to another (target language).

Role of Parsing in Machine Translation Systems

Parsing plays a crucial role in many machine translation approaches, especially in rule-based and syntax-based statistical or neural MT systems, by providing structured information about the source sentence and guiding the generation of the target sentence:

  • 1. Analyzing Source Sentence Structure:
    • MT systems parse the source language sentence to understand its grammatical composition (identifying subject, verb, object, phrases, clauses).
    • This structure helps clarify the relationships between words, which can be ambiguous in a flat word sequence.
    • Example: Parsing “I saw the man with the telescope” to determine if “with the telescope” modifies “man” or “saw.” This syntactic clarity is essential for correct translation.
  • 2. Creating Intermediate Representation:
    • In some MT approaches, the syntactic parse tree of the source sentence is transformed into an intermediate representation that captures key structural relationships independently of the source language words.
    • Translation occurs by transforming this intermediate structure and then generating the target language sentence from it.
  • 3. Guiding Transfer and Transformation Rules:
    • Syntax-based MT systems use explicit rules that operate on the source parse tree to rearrange and transform it into a target parse tree based on grammatical differences between languages.
    • Example: Rules can specify how active voice in one language translates to passive voice structure in another, or how adjectives are positioned relative to nouns. Parsing provides the tree structure these rules apply to.
  • 4. Generating Grammatically Correct Target Language Sentences:
    • Once the system has determined the target words and their desired relationships, parsing principles (or parsing/generation components based on grammar) are used to assemble these words into a grammatically correct and fluent sentence in the target language.
    • Generating output that adheres to the grammatical rules of the target language is vital for producing high-quality translations.
  • 5. Resolving Ambiguity: Parsing can sometimes help resolve syntactic ambiguities by favoring certain tree structures based on probabilistic models or constraints, though semantic information is also often required for full disambiguation.
  • Example Translation Flow (Syntax-based): Source Sentence -> Parser (builds Source Parse Tree) -> Transfer Rules (transform to Target Parse Tree structure) -> Generator (builds Target Sentence from tree) -> Target Sentence.

While modern Neural MT systems implicitly learn these transformations without explicit parsing rules, understanding the grammatical structure remains a core part of translating accurately, and techniques based on parsing (like using dependency trees or incorporating syntactic features) still contribute to model development and analysis.

Parsing Roles in MT

  • Analyze source syntax.
  • Help disambiguate (syntactic).
  • Create intermediate structures.
  • Guide transfer rules.
  • Ensure correct target grammar.

Helps Produce

  • Structurally correct output.
  • Clear word relationships.
  • Translation reflecting grammar.

Challenges

  • Source parsing errors propagate.
  • Handling grammar differences across languages.
  • Syntactic ambiguity persists.
  • Processing efficiency.

OR

Q.4 Explain how lexicalized parsing models incorporate syntactic and lexical information to improve parsing accuracy. 8 Marks

Definition: Lexicalized Parsing Models

Lexicalized parsing models are types of parsing models (often based on formal grammars) that incorporate specific information about individual words (lexical items) into their syntactic rules or structures. This allows them to leverage the properties and preferences of words (verbs for specific argument structures, prepositions for common phrases) to improve parsing accuracy compared to purely context-free grammar (CFG) based models.

Key Concepts of CFG Limitation

  • Standard CFGs only use syntactic categories (NP, VP, Verb). Rules are general (e.g., VP -> Verb NP).
  • They don’t know *which specific verbs* tend to take NP objects (transitive) vs not (intransitive), or which prepositions go with certain nouns/verbs (collocations).
  • This lack of lexical information makes CFG parsing prone to generating grammatically valid but semantically implausible parse trees, especially when sentences are structurally ambiguous. (Example: parsing “I saw the man with the telescope” might treat “saw” as needing a NP and “with the telescope” as a Modifier, but doesn’t know if “saw” typically combines with instrumental “with”).

Incorporating Syntactic and Lexical Information (How it Improves Accuracy)

Lexicalized models enhance parsing accuracy by making parsing decisions (which rules to apply, which branches to pursue in the tree) sensitive to the actual words involved:

  • 1. Word-Specific Rules/Probabilities:
    • Instead of a general rule like VP -> Verb NP, a lexicalized grammar might have rules associated with specific verbs or heads, e.g., VP -> “eat” NP, VP -> “sleep”.
    • In statistical lexicalized parsers, these rules or rule applications are weighted based on how often a specific lexical head (like a verb or noun) appears with certain syntactic arguments or modifiers in training data. Probabilities guide the search for the most likely parse tree.
    • Example: Parser knows the verb “eat” frequently appears with a following NP object (like “an apple”), while “sleep” usually does not take a direct object. This preference influences which VP parse is preferred for a sentence.
  • 2. Incorporating Argument Structures:
    • Lexicalized grammars encode the “valency” or argument structure preferences of specific words, especially verbs (e.g., “give” takes subject, direct object, and indirect object NPs; “arrive” usually just takes a subject NP and maybe a location PP).
    • Parsing process checks if the complements found in the sentence match the lexical head’s expected argument structure.
    • Example: If the parser encounters “give”, it prefers parses where it has three complements (someone GIVES something TO someone/somewhere) over one where it just has a subject.
  • 3. Modeling Lexical Preferences for Attachment:
    • Parsing involves attaching phrases correctly (e.g., attaching a prepositional phrase). Lexicalized models know that certain words prefer certain attachments.
    • Example: The verb “saw” might prefer to attach a PP “with a tool” (“saw with a telescope”), while the noun “man” might prefer attaching a PP “with an attribute” (“man with brown hair”). If the word is “telescope” after “man”, the parser considers if “saw” is more likely to be done “with a telescope” vs “man” being described “with a telescope”.
  • 4. Improving Resolution of Structural Ambiguity: By making parsing decisions sensitive to lexical heads and their known behavior from training data, lexicalized models are significantly better at resolving syntactic ambiguity than pure CFG parsers which would treat multiple ambiguous parses as equally valid if grammar rules allow.

By moving from general grammatical categories to rules and probabilities tied to specific words, lexicalized parsers constrain the possible parses and select the most probable ones based on how words are actually used, drastically improving accuracy.

Lexicalized Parsing Basis

  • CFG + Word Info.
  • Rules tied to specific words (lexical heads).
  • Uses probabilities from training data.
  • Models word preferences.

Benefits

  • Improved accuracy over CFGs.
  • Resolves syntactic ambiguity better.
  • Captures lexical preferences.
  • Basis for high-performance parsers.

Challenges

  • Requires large datasets with lexical info.
  • More complex grammar/model structure.
  • Can be computationally expensive.
  • Data sparseness (rare words).

Grammar Types

  • Lexicalized CFG (LCFG).
  • Combinatory Categorial Grammar (CCG).
  • Tree-Adjoining Grammar (TAG).
  • Head-Driven Phrase Structure Grammar (HPSG).

Describe any linguistic phenomena related to auxiliary verbs that pose challenges of CFG-based parsing. 8 Marks

Definition: Auxiliary Verbs

Auxiliary verbs (or helping verbs) are verbs that precede a main verb in a sentence to help express grammatical mood, tense, voice, or aspect (e.g., ‘be’, ‘have’, ‘do’, modal verbs like ‘can’, ‘will’, ‘may’, ‘must’). They often form part of the verb phrase.

Definition: CFG-Based Parsing

Parsing using a Context-Free Grammar (CFG) to build a hierarchical structure (parse tree) of a sentence based on grammatical rules that replace non-terminal categories with other non-terminals or terminals.

Linguistic Phenomena of Auxiliary Verbs Challenging CFG-Based Parsing

Auxiliary verbs exhibit behaviors that deviate from the simple constituent structure assumptions of basic CFGs, posing parsing challenges:

  • 1. Inversion (Subject-Auxiliary Inversion):
    • Problem: In many sentence types (questions, negative inversion), the auxiliary verb moves to a position *before* the subject, disrupting the standard Subject-Auxiliary-Verb order typically captured by simple CFG rules like S → NP VP.
    • Example: Statement: “John **is** swimming.” (S → NP VP) vs Question: “**Is** John swimming?” (Structure is Aux NP VP). Simple CFG rules often need special, sometimes redundant, rules to handle such inversions.
    • This creates multiple potential grammar rule applications for different sentence types.
  • 2. Multiple Auxiliaries and Their Ordering:
    • Problem: Sentences can have multiple auxiliaries in a fixed order (e.g., modal + have + be). Simple CFG rules might represent this linearly (VP -> Modal HaveAux BeAux MainVerb) but don’t naturally capture the nested relationship where each auxiliary governs the subsequent verb structure.
    • Example: “John **must** **have** **been** swimming.” Order (must, have, been) is fixed, each preceding one governing the next. Standard CFG doesn’t strongly capture this nesting without specific, less general rules.
  • 3. Verb Phrase Ellipsis (VPE):
    • Problem: A subsequent verb phrase containing a repeated auxiliary is omitted when its meaning is recoverable from context.
    • Example: “John can swim, and Mary can **too**.” (Implies “Mary can swim too”). The structure “Mary can too” contains an auxiliary (“can”) followed by “too”, where the main verb phrase (“swim”) is missing. CFG needs special rules or mechanisms to handle such missing constituents while correctly associating “can” with the elided “swim.”
  • 4. ‘Do’-Support:
    • Problem: The auxiliary ‘do’ is inserted in negative and question formations when no other auxiliary or form of ‘be’/’have’ is present. Its appearance is syntactically motivated rather than semantic.
    • Example: “John **does not** swim.” (Do-support for negation) vs “John swims.” Handling this insertion (or removal) of ‘do’ depending on sentence structure adds complexity to CFG rules aiming to cover both sentence types.
  • 5. Argument Structure and Verb Complements: While more a challenge for combining syntax/semantics, auxiliaries influence the form of the main verb and sometimes which complements are required (e.g., modal + base form, have + past participle). Capturing this dependent form within a CFG (where rules are context-free regarding words outside the replaced symbol) is tricky; auxiliary presence is context.

These phenomena require CFG grammars to be either very complex (proliferation of rules to cover permutations) or use extensions beyond the basic context-free definition (like features or transformations), which are handled in more sophisticated parsing frameworks.

Challenges List

  • Inversion (Aux before subject).
  • Multiple auxiliary ordering.
  • Verb Phrase Ellipsis (missing VPs).
  • ‘Do’-Support insertion.
  • Auxiliary verb form dependency.

Impact on CFG

  • Rule proliferation needed.
  • Breaks simple constituent assumption.
  • Makes parsing complex.
  • Often requires CFG extensions.

Example Sentence Patterns

  • “Can he come?” (Inversion).
  • “He must have been studying.” (Multiple).
  • “She did not go.” (Do-support).

OR

Q.5 Explain how dependency grammars represent the syntactic relationship between auxiliary verbs and other words in a sentence. 8 Marks

Definition: Dependency Grammar

Dependency Grammar is a theory of sentence structure that represents the grammatical relationships between words as dependencies, typically shown as directed links between a ‘head’ word and words that depend on it (its dependents). Unlike constituency grammar (like CFGs) which groups words into nested phrases, dependency grammar focuses on binary head-dependent relations.

Key Concepts

  • Focus is on words and their direct links, not nested phrases (though phrases can be inferred from dependencies).
  • Relationships are labeled (e.g., nsubj for nominal subject, obj for object, aux for auxiliary).
  • Every word (except often the main predicate/verb or sentence root) has exactly one head, on which it depends.
  • Represented graphically as a set of directed edges between words, forming a tree structure where nodes are words.
  • Often considered better suited for representing sentence meaning (semantic relations often align with dependency relations) and processing languages with more flexible word order than English.

Representing Relationship between Auxiliary Verbs and Other Words

In dependency grammar, auxiliary verbs are treated as heads or dependents depending on the specific type of auxiliary and convention, but their core role is to depend on or modify the meaning/tense of the main verb.

  • Auxiliary Depending on the Main Verb:
    • In many dependency frameworks, auxiliary verbs (non-modal) depend on the main verb of the clause. The main verb is the head.
    • The relationship label often signifies ‘aux’ or ‘auxpass’ (for passive voice auxiliaries).
    • Example: “John **is** swimming.” ‘swimming’ is the main verb. ‘is’ depends on ‘swimming’ with relation label ‘aux’. Dependency: `swimming <-[aux]- is`.
    • Example: “John **has** eaten the apple.” ‘eaten’ is the main verb. ‘has’ depends on ‘eaten’ with label ‘aux’. Dependency: `eaten <-[aux]- has`.
    • This directly represents the idea that the auxiliary modifies or helps the main verb.
  • Modal Auxiliary Acting as Head of VP (Sometimes):
    • In some models, modal auxiliaries (can, will, may, must) can act as the head of the Verb Phrase structure, with the main verb depending on the modal.
    • Example: “John **can** swim.” ‘can’ is the head. ‘swim’ depends on ‘can’ with label ‘vcomp’ (verb complement) or similar. Dependency: `can ->[vcomp]-> swim`.
    • However, other models consistently make the main verb the head, with the modal as an ‘aux’. Both approaches aim to capture the modifier/head relationship.
  • Dependency Tree Structure: A full dependency parse of a sentence containing an auxiliary would show the auxiliary linked directly to its main verb (or governing verb), and that verb linked to its subject, objects, etc., clearly indicating which word modifies which.
  • Example Sentence: “The cat **has** been sleeping.”
    • ‘has’ depends on ‘sleeping’ (aux).
    • ‘been’ depends on ‘sleeping’ (auxpass).
    • ‘sleeping’ depends on some implicit root or is the main predicate of the clause (or depends on ‘has’ in some systems).
    • ‘cat’ is the nominal subject (nsubj) of ‘sleeping’ (or ‘has’).
    • ‘the’ is the determinant (det) of ‘cat’.
  • The core idea is ‘has been sleeping’ involves ‘has’ and ‘been’ directly related to the concept of ‘sleeping’. Dependency captures these specific word-to-word links.

DG Basics

  • Words linked by dependencies.
  • Head -> Dependent links.
  • Relations are labeled.
  • Focus on word-to-word ties.

Benefits (for Auxiliaries)

  • Directly shows which verb is being modified.
  • Handles inversion more naturally than CFG.
  • Captures relationships regardless of word order shifts.

Challenges

  • Defining consistent heads (e.g., aux vs modal).
  • Long-distance dependencies still complex.
  • Parsing algorithm complexity.

Example Relationships

  • MainVerb <–[aux]– Auxiliary.
  • HeadVerb <–[auxpass]– ‘be’ in passive.
  • MainVerb <–[neg]– ‘not’.

Discuss the parsing techniques used for resolving syntactic ambiguity in natural language sentences. 8 Marks

Definition: Syntactic Ambiguity

Syntactic ambiguity (or structural ambiguity) occurs when a sentence or phrase can be parsed in multiple different ways according to the rules of a grammar, leading to multiple possible structural interpretations and potentially different meanings. (e.g., “I saw the man with the telescope”).

Definition: Parsing Techniques (Resolving Ambiguity)

Parsing techniques in NLP build structural representations of sentences. When faced with syntactic ambiguity, these techniques must employ strategies to identify all possible valid parse trees and then, for practical applications, select the most likely or correct interpretation based on various sources of information.

Parsing Techniques Used for Resolving Syntactic Ambiguity

Resolving syntactic ambiguity typically involves generating the ambiguous parses and then using additional constraints or probabilistic information:

  • 1. Generating Multiple Parse Trees (Chart Parsing, etc.):
    • Techniques like chart parsing (e.g., CYK algorithm) can find *all* valid parse trees for an ambiguous sentence given a CFG.
    • This makes the ambiguity explicit but doesn’t choose the best parse.
  • 2. Using Disambiguation Rules or Heuristics:
    • Apply pre-defined rules or preferences to favor certain structures. (e.g., Prefer to attach prepositional phrases to the nearest possible constituent, or based on known preferences for verbs/nouns).
    • Example: “I saw the man with the telescope”. A heuristic might prefer attaching “with the telescope” to the noun “man” over the verb “saw” (Noun Phrase Attachment heuristic), which works for sentences like “I saw the man with the beard.” (But fails for instrumentals!).
    • Limitations: Rules are often heuristic, hand-crafted, don’t cover all cases, and fail with complex ambiguities.
  • 3. Incorporating Lexical Preferences (Lexicalized Parsing):
    • As discussed (related Q4 OR), include information about specific words. Parsers learn *from data* which attachments or structures are statistically more likely for given lexical heads (verbs, nouns, prepositions).
    • Example: A lexicalized parser knows “saw” commonly takes “with instrument” attachments (“saw with my eyes”, “saw with a microscope”), while “man” doesn’t. It also knows “man” takes “with attribute” (“man with brown hair”). Based on “telescope,” it evaluates if “saw” is more likely to be done “with a telescope” vs “man” being described “with a telescope”.
    • Probability distributions learned from large corpora (collections of text) guide the choice among valid parses.
  • 4. Probabilistic Context-Free Grammars (PCFGs) and Probabilistic Parsing:
    • Assign probabilities to CFG production rules based on their frequency in a training corpus. (e.g., P(S -> NP VP), P(NP -> Det Noun)).
    • Calculate the probability of a parse tree as the product of probabilities of rules used to generate it.
    • Parsing aims to find the most probable parse tree (Viterbi parse) or compute probabilities for all parses.
    • Improves over rule-based but probabilities are purely structural unless combined with lexicalization (Lexicalized PCFGs). Can still favor syntactically common but semantically odd parses without semantic input.
  • 5. Using Statistical Machine Learning Models:
    • Train supervised machine learning models on labeled corpora (sentences manually annotated with correct parse trees).
    • These models learn features related to the context and relationships in the sentence to predict the correct parsing decisions, including resolving ambiguities based on complex learned patterns. Modern deep learning parsers fall into this category.
    • Benefit: Learn complex patterns automatically. Requires large labeled datasets.
  • 6. Incorporating Semantic Information (Integrated Systems):
    • Integrate syntactic parsing with semantic analysis or knowledge bases during the process.
    • Discard parses that are semantically implausible or violate known facts. (e.g., “The rock ate the cookie” is syntactically valid but semantically invalid, a robust parser might discard or lower its probability).

Modern NLP systems often combine several techniques (e.g., statistical parsing models trained with large datasets, sometimes incorporating lexical features) to achieve high accuracy in resolving syntactic ambiguity.

Resolution Strategies

  • Generate multiple parses.
  • Apply rules/heuristics.
  • Use lexical info (lexicalized parsing).
  • Use probabilities (PCFGs).
  • Train ML models (statistical parsing).
  • Integrate semantic/knowledge.

Purpose

  • Choose correct grammatical structure.
  • Select most likely interpretation.
  • Handle ambiguity challenges.
  • Improve understanding.

Challenges

  • Requires complex grammars/models.
  • Needs vast data.
  • Ambiguities can interact.
  • Heuristics are error-prone.
  • Computationally expensive.

Example Ambiguities

  • “saw the man with the telescope.”
  • “The chicken is ready to eat.”
  • PP attachment.
  • Coordination ambiguity.
  • Scope ambiguity.

OR

Q.6 Explain the concept of probabilistic modeling in ambiguity resolution and its applications in NLP. 8 Marks

Definition: Probabilistic Modeling

Probabilistic modeling in NLP involves using principles of probability and statistics to model linguistic phenomena. Instead of defining absolute rules (like in strict grammars), it assigns probabilities to outcomes (like word sequences, parse tree structures, word senses, translations).

Key Concepts in Ambiguity Resolution

  • Natural language is highly ambiguous (syntactic, lexical, semantic, etc.).
  • Humans often resolve ambiguity unconsciously using context, world knowledge, and statistical likelihoods learned from experience.
  • Probabilistic models aim to replicate this by assigning probabilities to the possible interpretations of an ambiguous input and choosing the interpretation with the highest probability.
  • Relies on learning probabilities from large corpora of text (training data).

Concept of Probabilistic Modeling in Ambiguity Resolution

Probabilistic models estimate the likelihood of different linguistic events or structures. When applied to ambiguity resolution, this involves:

  • 1. Enumerating Possible Interpretations:
    • For a given ambiguous input (e.g., a word, phrase, or sentence), identify all grammatically or lexically valid possible interpretations according to the grammar or lexicon.
    • Example: For “bank”, identify “financial bank” and “river bank”. For “I saw the man with the telescope”, identify “I used the telescope” vs “the man had the telescope” parses.
  • 2. Assigning Probabilities to Each Interpretation:
    • Train a statistical model on a large corpus to learn how often each interpretation occurs in different contexts.
    • Calculate the probability of each possible interpretation for the current ambiguous input, often considering the surrounding words or sentence structure.
    • Example (simplified): Model learns from data that in sentences where “saw” involves seeing someone (common), if a phrase “with a tool” follows the verb, it’s more likely the tool was used for seeing. If the phrase follows a noun, it’s more likely the noun possesses the attribute. The probabilities P(Saw | context) vs P(Man | context) guide resolution.
  • 3. Selecting the Most Probable Interpretation:
    • Choose the interpretation with the highest assigned probability as the most likely intended meaning.
    • Example: For “I went to the bank on the river”, the probability of “river bank” is estimated as much higher than “financial bank” based on co-occurrence patterns in training data (“river” frequently co-occurs with “river bank”, rarely with “financial bank”).
  • Different types of probabilistic models (e.g., n-gram models, Hidden Markov Models, Probabilistic Context-Free Grammars (PCFGs), neural language models) learn different types of linguistic probabilities.

Probabilistic modeling allows NLP systems to make “best guesses” in the face of ambiguity based on what is statistically common in language, rather than failing or requiring rigid, exception-laden rules.

Applications in NLP Ambiguity Resolution

  • Part-of-Speech (POS) Tagging: Resolving if a word like “bank” is a noun or a verb in context by checking probability of sequences like “the [Noun]” vs “to [Verb]”.
  • Word Sense Disambiguation (WSD): Determining the correct meaning of a word from its context using probabilities of senses given surrounding words.
  • Syntactic Disambiguation (Parsing): Choosing the most likely parse tree for a syntactically ambiguous sentence using PCFGs or other statistical parsers that assign probabilities to tree structures.
  • Coreference Resolution: Identifying what pronouns (he, she, it) refer to by checking the probabilities of a pronoun referring to different possible noun phrases based on distance, gender, number, and grammatical role.
  • Machine Translation: Resolving source language ambiguities and choosing the most probable translation in the target language by considering the probabilities of different word choices or structures given the context.
  • Speech Recognition: Choosing among similar-sounding words or phrases based on the most probable word sequence given acoustic input.

Model Process

  • List possible meanings.
  • Assign probability to each (from data).
  • Choose highest probability.

Used For

  • POS tagging.
  • Word sense disambiguation.
  • Syntactic parsing.
  • Coreference resolution.
  • Machine translation.
  • Speech recognition.

Benefits

  • Handle pervasive ambiguity.
  • Learn patterns from data.
  • Achieve high accuracy (in context).
  • Make “best guesses” like humans.

Challenges

  • Requires large amounts of labeled data.
  • Data sparseness for rare events.
  • Models can struggle with unexpected contexts.
  • Capturing complex world knowledge implicitly.

Scroll to Top