Unit 5 - Practice Quiz

INT428 60 Questions
0 Correct 0 Wrong 60 Left
0/60

1 What is the primary function of Generative AI?

Fundamentals of Generative AI Easy
A. To store and retrieve large volumes of information
B. To detect anomalies in a dataset
C. To classify existing data into predefined categories
D. To create new and original content

2 In the context of AI, what does LLM stand for?

LLMs Easy
A. Large Language Model
B. Logical Linguistic Module
C. Long Learning Machine
D. Low-Level Memory

3 A Generative Adversarial Network (GAN) is composed of two main neural networks. What are they called?

GANs Easy
A. A Generator and a Discriminator
B. A Predictor and a Corrector
C. An Encoder and a Decoder
D. A Master and a Follower

4 How do diffusion models typically begin the process of generating an image?

Diffusion Easy
A. By starting with a blank white canvas
B. By combining several smaller images
C. By starting with an image of pure random noise
D. By using a pre-existing high-resolution photograph

5 How do generative models learn to create new content?

How generative models work Easy
A. By accessing a live database of real-time information
B. By following a strict set of human-written rules and logic
C. By making random guesses until the output is correct
D. By identifying patterns and structures in large training datasets

6 Which of the following is a common business application of generative AI in marketing?

Industrial applications (content creation, automation) Easy
A. Processing financial transactions
B. Automating the creation of ad copy and social media posts
C. Optimizing factory floor operations
D. Managing supply chain logistics

7 What is a major ethical concern related to generative AI creating realistic but fake images or videos?

Ethics and Responsible Use of Generative AI Easy
A. The slow speed at which content is generated
B. The limited range of languages the AI understands
C. The high cost of the required computer hardware
D. The potential for creating 'deepfakes' to spread misinformation

8 In the context of using a large language model, what is a 'prompt'?

Prompt Engineering: Fundamentals of prompt Easy
A. The input or instruction provided by the user to the model
B. The final text output generated by the model
C. The software algorithm that powers the model
D. A measure of the model's accuracy

9 What is the primary reason why prompt engineering is important?

Importance of Prompt Engineering Easy
A. It is required to install the AI model on a computer
B. It reduces the amount of data the AI needs for training
C. It helps to get more accurate, relevant, and desired outputs from the AI
D. It makes the AI model run faster

10 Language models like GPT-4 are 'pre-trained'. What does this typically involve?

Overview of language models Easy
A. Installing a set of pre-configured software plugins
B. Training the model on a small, specific set of questions and answers
C. Training the model on a vast and diverse dataset of text from the internet
D. Manually programming the model with grammar rules

11 Which characteristic is essential for a good prompt?

Key elements of a good prompt Easy
A. Vagueness and ambiguity
B. Maximum possible length
C. Clarity and specificity
D. Using only single-word commands

12 Instructing an AI model to 'Act as a senior software engineer' before asking a coding question is an example of which prompt pattern?

Prompt Patterns Easy
A. The Repetition Pattern
B. The Question Refinement Pattern
C. The Persona Pattern
D. The Inverted Pattern

13 What is 'Prompt Tuning'?

Prompt Tuning Easy
A. Upgrading the hardware that the AI runs on
B. Changing the internal parameters of the AI model's code
C. Translating a prompt into different languages
D. The process of refining and iterating on a prompt to improve the AI's output

14 If an AI model is trained on data that contains historical biases, what is a likely consequence?

Ethics and Responsible Use of Generative AI Easy
A. The model will perform its tasks much more slowly
B. The model may generate outputs that perpetuate or amplify those biases
C. The model will automatically identify and remove all biases
D. The model will be unable to generate any output

15 In a GAN, what is the main objective of the 'Discriminator' network?

GANs Easy
A. To reduce the memory usage of the model
B. To speed up the training process of the entire network
C. To accurately distinguish between real data and fake data from the generator
D. To create new data that is as realistic as possible

16 A core capability of most modern LLMs is their ability to understand the ___ of a user's request.

LLMs Easy
A. context
B. emotion
C. location
D. hardware

17 In software development, tools like GitHub Copilot use generative AI to do what?

Industrial applications (content creation, automation) Easy
A. Test the physical hardware of a computer
B. Suggest and write code snippets automatically
C. Design the user interface of an application
D. Manage project budgets and timelines

18 Which of the following tasks is a classic example of what Generative AI does?

Fundamentals of Generative AI Easy
A. Identifying whether a picture contains a cat or a dog
B. Calculating the fastest route between two cities
C. Sorting a list of names alphabetically
D. Writing a new poem based on a theme

19 Providing a model with an example of a task before asking it to perform that task is known as what?

Prompt Patterns Easy
A. Role-play prompting
B. Context-free prompting
C. Zero-shot prompting
D. Few-shot or one-shot prompting

20 To get a list of bullet points from an AI model, what should you explicitly include in your prompt?

Key elements of a good prompt Easy
A. A request for the model to work as fast as possible
B. Instructions on the desired format, like "Provide the answer as a bulleted list"
C. The word 'generate' repeated many times
D. A very long and detailed paragraph of text

21 A data scientist training a Generative Adversarial Network (GAN) for creating realistic human faces notices that the generator starts producing only a few very similar-looking faces, failing to capture the diversity of the training data. What is this common GAN training problem called?

GANs Medium
A. Discriminator fatigue
B. Mode collapse
C. Vanishing gradient problem
D. Overfitting the latent space

22 In the context of a diffusion model, what is the primary objective of the neural network during the reverse process (denoising)?

Diffusion Medium
A. To upscale the low-resolution noisy image into a high-resolution clean image.
B. To predict the noise that was added to the image at a specific timestep, so it can be subtracted.
C. To accurately classify the object in the noisy image at each timestep.
D. To predict the final, clean image from the pure noise input in a single step.

23 A company uses a generative AI to create marketing images. The model, trained on a large public dataset, consistently generates images of CEOs as white men and nurses as women. This is a clear example of which ethical issue?

Ethics and Responsible Use of Generative AI Medium
A. Model hallucination
B. Data poisoning
C. Intellectual property infringement
D. Algorithmic bias amplification

24 Which of the following best describes the 'attention mechanism' in a Transformer-based Large Language Model?

LLMs Medium
A. A technique that allows the model to weigh the importance of different words in the input sequence when producing an output for a specific position.
B. A regularization technique to prevent the model from focusing too much on any single input token.
C. A fixed-size memory buffer where the model stores important contextual information.
D. A method to increase the model's vocabulary size by paying attention to new words.

25 A user wants an LLM to solve a complex multi-step physics problem. Which prompt pattern would be most effective in guiding the model to a more accurate and transparent solution?

Prompt Patterns Medium
A. The Persona Pattern (e.g., "Act as a physicist...")
B. The Question Refinement Pattern (e.g., "You are a helpful assistant...")
C. The Template Pattern (e.g., "Problem: [insert problem], Solution: [provide solution]")
D. The Chain-of-Thought (CoT) Pattern (e.g., "...Let's think step by step.")

26 A software development company wants to use a generative AI model to automatically generate unit tests for their code. Which capability of generative models is most directly being leveraged in this application?

Industrial applications (content creation, automation) Medium
A. Data augmentation
B. Code generation and understanding
C. Sentiment analysis
D. Image synthesis

27 What is the key difference between 'Prompt Tuning' (or soft prompting) and 'Full Fine-Tuning' of a Large Language Model?

Prompt Tuning Medium
A. Prompt Tuning updates all the weights of the model, while Full Fine-Tuning only updates the final layer.
B. Prompt Tuning keeps the original LLM's weights frozen and only trains a small set of prompt-specific parameters, making it highly parameter-efficient.
C. Prompt Tuning only works for text generation tasks, whereas Full Fine-Tuning works for both classification and generation.
D. Prompt Tuning requires far more data than Full Fine-Tuning to be effective.

28 When a Variational Autoencoder (VAE) generates a new data point, what is it sampling from?

How generative models work Medium
A. A fixed, deterministic vector representing the most common data point.
B. The original input data distribution directly.
C. The output of a competing discriminator network.
D. A learned latent space, which is typically modeled as a Gaussian distribution.

29 A user provides the prompt: "Write about cars." The LLM's output is broad and unhelpful. Which key element of a good prompt is most obviously missing?

Key elements of a good prompt Medium
A. Persona
B. Format
C. Context and Constraints
D. Tone

30 A journalist uses an LLM to help write an article and the model generates a factual-sounding but entirely incorrect statistic about a public figure. This phenomenon, where the model confidently produces false information, is known as:

Ethics and Responsible Use of Generative AI Medium
A. Hallucination
B. Overfitting
C. Bias
D. Data leakage

31 In a well-functioning Generative Adversarial Network (GAN), what is the ideal state of the discriminator's accuracy on a mixed set of real and generated data as training converges?

GANs Medium
A. Approaching 100%, as it perfectly identifies fakes.
B. Approaching 0%, as it is completely fooled by the generator.
C. Fluctuating wildly between 0% and 100%.
D. Approaching 50%, as it cannot reliably distinguish between real and fake samples.

32 Which statement best differentiates a generative AI model from a discriminative AI model?

Fundamentals of Generative AI Medium
A. Generative models are always unsupervised, while discriminative models are always supervised.
B. Generative models learn the underlying distribution of the data, , to create new samples, while discriminative models learn the decision boundary between classes, .
C. Generative models are less computationally expensive to train than discriminative models.
D. Generative models can only be used for creating images, while discriminative models are used for text classification.

33 What is the primary purpose of the 'temperature' parameter when generating text from a Large Language Model?

LLMs Medium
A. To adjust the model's processing speed and computational cost.
B. To control the randomness and creativity of the model's output.
C. To control the length of the generated output.
D. To set the maximum number of tokens the model can consider as context.

34 A marketing specialist wants to generate several variations of an advertisement slogan. They want to provide the LLM with a clear structure and examples to follow. Which prompt pattern is best suited for this task?

Prompt Patterns Medium
A. Chain-of-Thought Pattern
B. Flipped Interaction Pattern
C. Persona Pattern
D. Few-Shot or In-Context Learning Pattern

35 In an autoregressive generative model like GPT, how is a new token (word or sub-word) generated?

How generative models work Medium
A. It is generated by a separate discriminator model that suggests the most likely next token.
B. All tokens in the sequence are generated simultaneously based on a single latent vector.
C. It is randomly selected from the entire vocabulary with uniform probability.
D. It is generated based on the sequence of tokens that have been generated previously in the same sequence.

36 A pharmaceutical company uses a generative model to design novel protein structures that might bind to a specific disease target. This application primarily falls under the category of:

Industrial applications (content creation, automation) Medium
A. Synthetic data generation for privacy
B. Drug discovery and materials science
C. Automated content summarization
D. Creative content creation for marketing

37 Compared to Generative Adversarial Networks (GANs), what is a widely recognized advantage of Diffusion Models regarding training stability?

Diffusion Medium
A. Diffusion models require much less training data than GANs.
B. Diffusion models do not require the use of neural networks.
C. Diffusion models train significantly faster than GANs.
D. Diffusion models generally have a more stable training process and are less prone to issues like mode collapse.

38 A company wants to adapt a single, large, pre-trained language model to perform several different, specialized tasks (e.g., legal document analysis, medical chatbot, marketing copy generation) without creating and storing a full copy of the model for each task. Which technique would be most efficient?

Prompt Tuning Medium
A. Training a separate, smaller model from scratch for each task.
B. Using the base model with only zero-shot prompts.
C. Using Prompt Tuning to create a small, task-specific 'soft prompt' for each task.
D. Full Fine-Tuning for each task.

39 To mitigate the risk of a generative AI creating harmful or biased content, developers often employ a technique where the model's outputs are checked against a set of predefined safety rules or filtered by another model before being shown to the user. This practice is an example of:

Ethics and Responsible Use of Generative AI Medium
A. Unsupervised learning
B. Data augmentation
C. Output moderation and safety filtering
D. Latent space interpolation

40 Why is 'prompt engineering' considered a critical skill for effectively using modern large language models, even though the models are pre-trained on vast amounts of data?

Importance of Prompt Engineering Medium
A. Because LLMs can only understand prompts written in a specific programming language.
B. Because models will refuse to generate a response if the prompt is not engineered correctly.
C. Because the quality, specificity, and structure of the input prompt directly and significantly influence the relevance and accuracy of the model's output.
D. Because prompt engineering is the only way to update the weights of the model.

41 In the context of Generative Adversarial Networks (GANs), the original formulation uses a minimax objective function which is equivalent to minimizing the Jensen-Shannon (JS) divergence. Why does switching to the Wasserstein-1 distance, as in WGANs, lead to more stable training and mitigate mode collapse?

GANs Hard
A. The Wasserstein distance provides a smoother gradient signal everywhere, even when the generator and real distributions have disjoint supports, whereas the JS divergence gradient vanishes in this case.
B. The Wasserstein distance is computationally cheaper to approximate than the JS divergence, allowing for larger batch sizes and faster convergence.
C. The Wasserstein distance forces the generator to produce samples that are exactly identical to the training data, thus eliminating mode collapse by definition.
D. The Wasserstein distance requires the discriminator (critic) to be a simple binary classifier, which is easier to optimize than the complex function required for JS divergence.

42 A key component of a Denoising Diffusion Probabilistic Model (DDPM) is the noise schedule, , which controls the variance of noise added at each forward step . How does the choice of a cosine-based noise schedule, compared to a linear schedule, typically impact the generation process?

Diffusion Hard
A. A cosine schedule adds noise more slowly at the beginning of the forward process and more rapidly at the end, which prevents the model from destroying information about the data distribution too quickly and often leads to better sample quality.
B. A cosine schedule makes the forward process non-Markovian, which simplifies the mathematical derivation of the loss function but requires more memory to train.
C. A cosine schedule significantly speeds up the inference (sampling) time by allowing for larger steps in the reverse process, but at the cost of lower sample diversity.
D. A linear schedule is provably optimal for all datasets, while a cosine schedule is a heuristic that only works for low-resolution images.

43 When fine-tuning a large language model (LLM) on a new, smaller dataset, one major risk is 'catastrophic forgetting', where the model loses its pre-trained general capabilities. How does a Parameter-Efficient Fine-Tuning (PEFT) method like Low-Rank Adaptation (LoRA) specifically address this problem at an architectural level?

LLMs Hard
A. LoRA retrains only the final classification head of the LLM and keeps the entire Transformer body frozen, which is a technique known as transfer learning.
B. LoRA identifies and freezes the neurons responsible for general knowledge while only allowing the neurons related to the new task to be updated.
C. LoRA creates a complete copy of the model for the new task and uses a regularization term to penalize deviations from the original model's outputs.
D. LoRA freezes all the original pre-trained weights and injects small, trainable low-rank matrices into the layers of the Transformer. Only these new matrices are updated, preserving the vast knowledge in the original weights.

44 Consider a 'data poisoning' attack on a generative model being trained for a commercial application. How does this attack vector fundamentally differ from the more commonly discussed issue of inherent societal bias in the training data?

Ethics and Responsible Use of Generative AI Hard
A. Inherent bias only affects discriminative models, while data poisoning is a unique vulnerability of generative models.
B. Data poisoning is a deliberate, adversarial attack where malicious data is intentionally injected to cause specific, targeted failures (e.g., generating harmful content for specific prompts), whereas inherent bias is an unintentional reflection of systemic societal prejudices present in the source data.
C. Inherent bias can be completely eliminated by carefully filtering the training data, while data poisoning is a permanent and irreversible corruption of the model's weights.
D. Data poisoning aims to improve the model's performance on a hidden task known only to the attacker, whereas inherent bias universally degrades the model's performance.

45 The ReAct (Reason and Act) prompt pattern significantly enhances LLM capabilities for complex tasks. What is the core mechanism of ReAct that distinguishes it from a simpler Chain-of-Thought (CoT) pattern?

Prompt Patterns Hard
A. ReAct interleaves reasoning traces (thoughts) with actions that can query external tools (e.g., a search engine or API), allowing the model to dynamically gather information to inform its next reasoning step.
B. ReAct requires the model to generate responses in a structured format like JSON or XML, whereas CoT produces free-form text.
C. ReAct is exclusively used for code generation, while CoT is used for natural language reasoning tasks.
D. ReAct is a form of unsupervised learning where the model generates its own prompts, whereas CoT relies on a human-provided prompt with few-shot examples.

46 From a computational and deployment perspective, what is the most significant advantage of using Prompt Tuning (which learns a soft prompt) over full fine-tuning for serving multiple customized tasks on a single base LLM?

Prompt Tuning Hard
A. Prompt Tuning requires no labeled data for the new task, making it a form of zero-shot learning.
B. Soft prompts generated by Prompt Tuning are human-readable and can be easily edited, offering better interpretability than the weights modified during full fine-tuning.
C. A single, unmodified base LLM can be served, and task-specific capabilities are enabled by loading only a very small (e.g., <1MB) task-specific soft prompt, drastically reducing storage and memory costs compared to storing a full model copy for each task.
D. Prompt Tuning achieves significantly higher accuracy on all tasks compared to full fine-tuning because it avoids overfitting.

47 Comparing the fundamental mechanisms of major generative model families, which statement accurately synthesizes their differences in modeling the data distribution ?

How generative models work Hard
A. Autoregressive models and Normalizing Flows compute a tractable density for , while GANs and Diffusion Models learn an implicit distribution that can be sampled from but whose density cannot be easily evaluated.
B. All generative models aim to maximize the exact log-likelihood of the data, but GANs use a surrogate objective that makes optimization easier.
C. Diffusion models are the only class of models that can generate data in a single forward pass, making them the fastest at inference time.
D. GANs and VAEs both use an explicit latent variable to model , while Autoregressive and Diffusion models generate data directly in the pixel/token space without a latent representation.

48 A software company integrates an LLM-based code generation tool into its development workflow to increase productivity. Beyond the risk of generating buggy code, what represents the most significant, long-term strategic risk to the company's engineering culture and intellectual property?

Industrial applications (content creation, automation) Hard
A. The model may inadvertently generate code that is licensed under a restrictive open-source license (e.g., GPL), creating complex legal and compliance issues for the company's proprietary codebase.
B. The gradual erosion of deep problem-solving and system design skills among junior and mid-level developers, leading to a workforce that is proficient at prompting but lacks fundamental engineering expertise, creating critical skill gaps and technical debt.
C. An increase in the company's cloud computing costs due to the high computational requirements of running the LLM-based tool.
D. The model will automate all simple coding tasks, forcing the company to lay off its entire junior developer team.

49 What is the primary cause of 'mode collapse' in a Generative Adversarial Network (GAN), and how does it manifest in the generator's output?

GANs Hard
A. The generator's learning rate is too high, causing its weight updates to be unstable and oscillate between a few different output modes.
B. The latent space is too low-dimensional, preventing the generator from having enough capacity to represent the full diversity of the training data.
C. The loss function for the generator saturates, meaning its value stops changing even as the outputs get worse, leading to random, unstructured outputs.
D. The discriminator becomes too powerful too quickly, learning to perfectly distinguish real from fake samples. It provides a narrow, uninformative gradient signal that pushes the generator to produce only a few highly plausible samples that can fool the discriminator.

50 The concept of 'emergent abilities' in Large Language Models describes capabilities that are not present in smaller models but appear suddenly in larger models. What is a key implication of this phenomenon for AI development and safety?

LLMs Hard
A. It implies that the behavior of future, more powerful models is fundamentally unpredictable based on the scaling of current models, meaning new capabilities and potential risks could arise unexpectedly without being explicitly designed or trained for.
B. It means that emergent abilities can be programmed into a model by carefully curating the training data, giving developers full control over their appearance.
C. It suggests that smaller, more efficient models are inherently safer because they cannot develop these unpredictable abilities.
D. It proves that model performance scales linearly with the number of parameters, allowing for precise prediction of future abilities.

51 The 'Stochastic Parrots' critique of LLMs argues against viewing them as intelligent beings. What is the central thesis of this argument, moving beyond simple bias concerns?

Ethics and Responsible Use of Generative AI Hard
A. That LLMs are systems for 'stitching together' sequences of linguistic forms based on probabilistic information, but without any true understanding or grounding in the real world, which can lead to them amplifying biases, being environmentally costly, and creating a misleading illusion of meaning.
B. That the random, or 'stochastic', nature of LLM outputs makes them too unreliable for any serious application.
C. That LLMs can only repeat, or 'parrot', information they have seen in their training data and are incapable of generating novel sentences or ideas.
D. That the primary ethical risk of LLMs is that they will learn to perfectly mimic human speech and deceive users into thinking they are conscious.

52 What fundamental architectural limitation of Recurrent Neural Networks (RNNs) and LSTMs was the Transformer architecture's self-attention mechanism specifically designed to overcome?

Overview of language models Hard
A. The sequential nature of RNNs, which prevents parallelization of computations within a single training example and makes it difficult to model long-range dependencies due to the path length between distant tokens.
B. The fixed-size context window of RNNs, which limits their input to a few hundred words.
C. The inability of RNNs to process text data, as they were originally designed for numerical time-series data.
D. The excessive number of parameters in RNNs, which the Transformer architecture significantly reduces through weight sharing.

53 The objective function of a Diffusion Model is typically framed as a variational lower bound (ELBO) on the log-likelihood of the data, similar to a Variational Autoencoder (VAE). However, their approaches to the latent space differ significantly. What is a key distinction?

Diffusion Hard
A. VAEs use a simple Gaussian prior for the latent space, while Diffusion Models require a complex, learned prior that is more expressive.
B. Diffusion Models can only model discrete data, while VAEs are designed for continuous data like images.
C. The decoder in a VAE is probabilistic, while the decoder (reverse process) in a Diffusion Model is deterministic.
D. In Diffusion Models, the latent space has the same dimensionality as the input data, and the encoder is a fixed, non-learned process (adding noise), whereas in VAEs, the encoder is a learned neural network that maps data to a low-dimensional latent space.

54 In the context of LLMs, what is 'AI alignment' and how does it differ from traditional performance metrics like perplexity or accuracy?

LLMs Hard
A. Alignment is the process of steering a model's behavior to follow human intent and ethical principles (e.g., be helpful, honest, and harmless), which is a qualitative and normative goal, whereas metrics like perplexity measure the model's statistical prediction quality on a dataset.
B. Alignment measures how well the model's internal representations match human neural activity, while accuracy measures its performance on a downstream task.
C. Alignment refers to ensuring the model's output is grammatically correct and coherent, while accuracy measures its factual correctness.
D. Alignment is a specific fine-tuning technique, such as RLHF, whereas perplexity is a pre-training metric.

55 A user provides the prompt: "Write about electric cars." The LLM produces a generic, high-level summary. To elicit a detailed analysis comparing the 5-year total cost of ownership (TCO) for a Tesla Model 3 vs. a Toyota Camry, including depreciation, fuel, insurance, and maintenance, which combination of prompt elements would be most effective?

Key elements of a good prompt Hard
A. Using a one-shot example showing a TCO analysis for two different cars.
B. Combining a Persona ('Act as a senior automotive financial analyst'), specific Constraints ('Focus only on the US market for the 2023 models'), a detailed Task Description ('Analyze the 5-year TCO...'), and an Output Format ('Present the final comparison in a markdown table').
C. Simply adding more keywords like 'cost', 'finance', 'Tesla', and 'Camry' to the original prompt.
D. Increasing the temperature parameter of the model to encourage more creative and detailed output.

56 A key distinction in machine learning is between generative and discriminative models. If you are building a system to generate entirely new, photorealistic images of human faces, you would use a generative model. If you are building a system to determine whether an existing image contains a human face or not, you would use a discriminative model. This is because:

Fundamentals of Generative AI Hard
A. Generative models require significantly more data to train than discriminative models, which is why they are only used for content creation.
B. Generative models use unsupervised learning exclusively, while discriminative models use supervised learning exclusively.
C. The generative model learns the joint probability distribution (or just for unconditional generation), allowing it to sample new data points, while the discriminative model learns the conditional probability to predict a label for a given input .
D. Discriminative models are a subclass of generative models that have been fine-tuned for classification tasks.

57 You are tasked with creating a prompt that generates a fictional historical narrative from the perspective of a specific character, but you want to explicitly prevent the narrative from including anachronistic technology (e.g., smartphones in ancient Rome). Which advanced prompting technique would be most direct and effective for this specific type of content steering?

Prompt Patterns Hard
A. Employing a 'Chain of Thought' prompt to force the model to reason about the historical accuracy of each sentence it writes.
B. Using a 'Template' prompt that provides a strict fill-in-the-blanks structure, leaving little room for the model to introduce new concepts.
C. Using a 'Negative Prompt' in combination with a 'Persona' prompt. The persona sets the character ('Act as a Roman centurion...'), and the negative prompt explicitly forbids certain concepts ('Do not include any technology invented after the 1st century AD...').
D. Fine-tuning the model on a dataset of historically accurate narratives, which is a training method, not a prompting technique.

58 A company deploys a generative AI to create marketing copy. The AI generates text that, while factually correct about the product, is found to be subtly manipulative by using phrasing and emotional appeals that exploit known cognitive biases (e.g., scarcity, authority bias) to an unethical degree. This scenario highlights which complex ethical challenge?

Ethics and Responsible Use of Generative AI Hard
A. The problem of model hallucination, as the AI is generating content that is not strictly based on its input data.
B. The problem of data privacy, as the AI must have been trained on private user data to understand cognitive biases.
C. The issue of copyright infringement, as the manipulative phrases were likely copied from other marketing campaigns.
D. The challenge of 'persuasive AI' and dual-use technology, where the line between effective marketing and unethical manipulation is subjective and difficult to codify into model guardrails, raising questions of accountability for the AI's persuasive impact.

59 Why is 'prompt engineering' considered a critical skill for effectively using large language models, even though the models themselves are designed to understand natural language? What underlying aspect of LLM behavior necessitates this practice?

Importance of Prompt Engineering Hard
A. LLMs have a limited vocabulary and can only understand specific keywords, so prompt engineering is about finding the exact sequence of keywords the model was trained on.
B. Prompt engineering is a temporary workaround for bugs in current LLMs and will become obsolete as models become more intelligent.
C. LLMs operate within a high-dimensional latent space, and the prompt acts as a vector that guides the model to a specific region of that space. Small changes in the prompt can lead to large, non-intuitive changes in the output trajectory, making precise phrasing essential to constrain the model to the desired solution manifold.
D. The primary purpose of prompt engineering is to reduce the computational cost of running the LLM by making the input sequence as short as possible.

60 When comparing the industrial application of generative AI for content summarization versus content synthesis, what is the key value proposition of synthesis that summarization cannot offer?

Industrial applications (content creation, automation) Hard
A. Synthesis can create novel insights and conclusions by integrating information from multiple, disparate sources into a new, coherent document, whereas summarization is fundamentally reductive and limited to the information within a single source.
B. Synthesis is much faster and computationally cheaper than summarization because it does not need to analyze the entire source document.
C. Summarization can only be applied to text, while synthesis can be applied to text, images, and audio.
D. Summarization is always less accurate than synthesis because it involves compression, which loses information.