Turn Articles and Blog Posts into Quizzes with AI

Tested prompts for generate quiz from article compared across 5 leading AI models.

If you have an article, blog post, study guide, or training document and need to test whether readers actually absorbed the content, generating a quiz manually takes time you probably do not have. AI can read the source text and return multiple-choice questions, true/false items, or short-answer prompts in seconds. That is the core problem this page solves.

The prompt and model outputs shown on this page were built specifically for the task of converting written content into quiz questions. You paste your article, specify the format and difficulty, and the model extracts key concepts, turns them into testable questions, and writes plausible wrong answers so the quiz actually works as a learning check.

Before you scroll to the outputs, it helps to know what separates a useful quiz from a weak one. Good questions test comprehension of the article's main ideas, not trivia buried in a footnote. The examples and tips below will help you calibrate the prompt so the AI produces questions you can actually use without editing every single one.

When to use this

This approach fits any situation where you have a body of text and need to measure or reinforce understanding. It works whether you are an educator building a lesson assessment, a content marketer creating an interactive piece, a corporate trainer packaging a policy document, or a developer prototyping a learning product.

  • Converting a long-form blog post into a reader engagement quiz for your website
  • Turning a company policy or compliance document into a knowledge-check assessment for onboarding
  • Creating study questions from a textbook chapter or research article for students
  • Building a quick comprehension check after a newsletter issue to increase subscriber interaction
  • Generating practice questions from a product FAQ or knowledge base article for customer education

When this format breaks down

  • The source article is shorter than 200 words. There is not enough material for the AI to generate varied, non-repetitive questions without padding or making things up.
  • You need psychometrically validated assessments for high-stakes testing, such as professional certification exams. AI-generated questions require expert review before being used in that context.
  • The article covers a highly technical or niche domain where factual errors are dangerous, such as medical dosing or legal statutes, and you do not have a subject-matter expert to review every output.
  • Your source content is paywalled, proprietary, or under NDA and you are pasting it into a third-party AI interface without confirming your organization's data policy permits that.

The prompt we tested

You are an expert quiz designer who transforms written content into engaging, accurate quizzes that test comprehension and retention.

Follow these instructions carefully:
Generate exactly 8 questions: 5 multiple-choice (4 options labeled A-D, only one correct), 2 true/false, and 1 short-answer. Format as a numbered list with the question type in brackets (e.g., '1. [Multiple Choice]'), then list options, then 'Answer:' and 'Explanation:' on separate lines. Keep language clear, neutral, and at the same reading level as the source article.

Here is the article to turn into a quiz:
---
The James Webb Space Telescope, launched in December 2021, is the largest and most powerful space telescope ever built. Unlike Hubble, which observes primarily in visible and ultraviolet light, Webb is optimized for infrared observations, allowing it to peer through cosmic dust and study the earliest galaxies formed after the Big Bang. It orbits the Sun at the second Lagrange point (L2), about 1.5 million kilometers from Earth, where its massive sunshield keeps its instruments cooled to around -223°C.
---

Produce the quiz now. Base every question strictly on the article's content (no outside knowledge), cover the most important facts and ideas, and vary question difficulty from recall to inference. After each question, provide the correct answer and a one-sentence explanation citing the relevant part of the article.

Model comparison

ModelQualityCostLatencyTokens
Claude Opus 4.7Not run yet
Claude Haiku 4.5Not run yet
GPT-5.4Not run yet
Gemini 2.5 Pro $0.0021 17.8s 301 / 338
Grok 4.1 Fast Reasoning $0.0004 7.2s 439 / 540

Outputs

Gemini 2.5 Pro
1. [Multiple Choice] When was the James Webb Space Telescope launched?
A. December 2020
B. December 2021
C. January 2022
D. January 2023
Answer: B
Explanation: The article's first sentence states that the telescope was "launched in December 2021."

2. [Multiple Choice] According to the article, a primary difference between the Webb and Hubble telescopes is...
A. Webb orbits the Earth, while Hubble orbits the Sun.
B. Webb is smaller and less powerful than Hubble.
C. Webb observes in infrared, while Hubble focuses on visible and ultraviolet light.
D. Webb requires a higher operating temperature than Hubble.
Answer: C
Explanation: The text notes that "Unlike Hubble, which observes primarily in visible and ultraviolet light, Webb is optimized for infrared observations."

3. [Multiple Choice] What is the function of the Webb telescope's massive sunshield?
A. To power the telescope using solar energy.
B. To protect the instruments from space debris.
C. To keep the instruments at their required cool temperature.
D. To focus light into the telescope's primary mirror.
Answer: C
Explanation: The article mentions that the "massive sunshield keeps its instruments cooled to around -223°C."

4. [Multiple Choice] Where is the James Webb Space Telescope located?
A. In a low orbit around the Earth.
B. In an orbit around the Moon.
C. At the second Lagrange point (L2), orbiting the
Grok 4.1 Fast Reasoning
1. [Multiple Choice] When was the James Webb Space Telescope launched?  
A. December 2020  
B. December 2021  
C. December 2022  
D. January 2021  
Answer: B  
Explanation: The article states that the James Webb Space Telescope was launched in December 2021.

2. [Multiple Choice] Compared to the Hubble telescope, what is the James Webb Space Telescope optimized for?  
A. Visible light observations  
B. Ultraviolet light observations  
C. Infrared observations  
D. X-ray observations  
Answer: C  
Explanation: Unlike Hubble, which observes primarily in visible and ultraviolet light, Webb is optimized for infrared observations.

3. [Multiple Choice] Where does the James Webb Space Telescope orbit the Sun?  
A. At the first Lagrange point (L1)  
B. At the second Lagrange point (L2)  
C. Directly around Earth  
D. Around the Moon  
Answer: B  
Explanation: It orbits the Sun at the second Lagrange point (L2), about 1.5 million kilometers from Earth.

4. [Multiple Choice] Approximately how far is the James Webb Space Telescope from Earth?  
A. 1.5 thousand kilometers  
B. 15 thousand kilometers  
C. 1.5 million kilometers  
D. 15 million kilometers  
Answer: C  
Explanation: It orbits the Sun at the second Lagrange point (L2), about 1.5 million kilometers from Earth.

5. [Multiple Choice] What advantage do the infrared observations of the James Webb Space Telescope provide?  
A. Observing in ultraviolet light  
B. Cooling the instruments  
C. Peering through cosmic dust  
D. Orbiting closer to Earth  
Answer: C  
Explanation: Webb is optimized for infrared observations, allowing it to peer through cosmic dust and study the earliest galaxies formed after the Big Bang.

6. [True/False] The James Webb Space Telescope is the largest and most powerful space telescope ever built.  
Answer: True  
Explanation: The article describes the James Webb Space Telescope as the largest and most powerful space telescope ever built.

7. [True/False] The James Webb Space Telescope's massive sunshield keeps its instruments cooled to around -223°C.  
Answer: True  
Explanation: The article states that its massive sunshield keeps its instruments cooled to around -223°C.

8. [Short Answer] What specific cosmic phenomena can the James Webb Space Telescope study due to its infrared optimization?  
Answer: The earliest galaxies formed after the Big Bang  
Explanation: Infrared observations allow it to peer through cosmic dust and study the earliest galaxies formed after the Big Bang.

What makes these work

  1. 01
    Specify question type and count upfront

    Telling the model exactly how many questions you need and what format (multiple-choice with 4 options, true/false, short-answer) eliminates guessing and produces output you can use immediately. If you leave format open, the model may mix types inconsistently or produce more questions than you need, requiring extra editing.

  2. 02
    Set difficulty relative to your audience

    Words like 'beginner,' 'intermediate,' or 'advanced' meaningfully change output. A beginner question tests recall of a fact stated directly in the article. An advanced question asks the reader to apply, compare, or infer. Name your audience explicitly in the prompt, such as 'for 10th-grade students' or 'for certified professionals,' and the model calibrates accordingly.

  3. 03
    Ask for the correct answer to be labeled

    Unless you tell the model to mark which answer is correct, some outputs will omit that information and force you to manually check every item. Add a line to your prompt like 'mark the correct answer with (correct)' or 'provide an answer key below the questions.' This saves significant post-processing time when generating more than five questions.

  4. 04
    Paste the most information-dense section of the article

    Introductions and conclusions are often thin on testable facts. If your article is long, paste the body paragraphs that contain the core arguments, data, or processes. The model will generate sharper questions from content-rich text than from summary or transitional passages, and you avoid questions based on the author's framing rather than actual knowledge.

More example scenarios

#01 · Online course creator turning a nutrition blog post into a module quiz
Input
Article excerpt: 'Protein synthesis requires a positive nitrogen balance, meaning protein intake must exceed protein breakdown. For most active adults, research supports consuming 1.6 to 2.2 grams of protein per kilogram of body weight per day. Timing also matters: studies show that consuming 20 to 40 grams of protein within two hours post-exercise maximizes muscle protein synthesis.' Generate 3 multiple-choice questions at a beginner difficulty level.
Expected output
Q1: What does a positive nitrogen balance mean? A) Protein intake exceeds protein breakdown (correct) B) Carbohydrate intake exceeds fat intake C) Caloric intake exceeds caloric expenditure D) Nitrogen is stored in muscle cells. Q2: What is the recommended daily protein range for active adults? A) 0.8g/kg B) 1.0-1.2g/kg C) 1.6-2.2g/kg (correct) D) 3.0-4.0g/kg. Q3: When should protein ideally be consumed after exercise? A) Within 30 minutes B) Within 2 hours (correct) C) 4-6 hours later D) Timing does not matter.
#02 · HR manager converting a remote work policy document into a compliance quiz
Input
Policy text: 'Employees working remotely must be available via Slack between 9 AM and 3 PM in their local time zone. All work-related data must be stored on company-approved cloud platforms only. Personal devices may be used only if enrolled in the company MDM program. Expenses for home office equipment up to $500 per year are reimbursable with a receipt submitted within 30 days of purchase.' Generate 4 true/false questions.
Expected output
1. Remote employees must be available on Slack from 9 AM to 5 PM. FALSE - availability is required 9 AM to 3 PM. 2. Work data may be stored on any cloud platform the employee prefers. FALSE - only company-approved platforms are permitted. 3. Personal devices can be used for work if enrolled in the MDM program. TRUE. 4. Home office equipment reimbursements require a receipt submitted within 30 days. TRUE.
#03 · High school teacher generating reading comprehension questions from a history article
Input
Article summary: 'The Marshall Plan, officially the European Recovery Program, was a U.S. initiative launched in 1948 to provide over $13 billion in economic aid to rebuild Western European economies after World War II. Secretary of State George Marshall proposed the plan in a 1947 Harvard speech. The plan was partly motivated by concern that economic instability would make European countries vulnerable to communist influence.' Generate 3 short-answer questions for 10th graders.
Expected output
1. What was the official name of the Marshall Plan, and what year did it launch? (Expected: European Recovery Program, 1948) 2. Who proposed the Marshall Plan, and where did he announce it? (Expected: Secretary of State George Marshall, in a Harvard speech in 1947) 3. Besides economic recovery, what political concern motivated the United States to create the plan? (Expected: Fear that economic instability would increase communist influence in Western Europe)
#04 · SaaS marketer building an interactive quiz from a product explainer article
Input
Article focus: A 600-word explainer on how a project management tool handles task dependencies, Gantt chart views, and automated deadline reminders. Generate 4 multiple-choice questions that a potential customer could answer to see how much they already know about project management software, with difficulty set to intermediate.
Expected output
Q1: What is a task dependency in project management software? A) A task that requires budget approval B) A task that cannot start until a linked task is complete (correct) C) A recurring task set on a timer D) A task assigned to more than one user. Q2: What does a Gantt chart primarily display? A) Team member availability B) Budget allocation by department C) Task timelines and overlaps across a project (correct) D) Client communication logs. Q3: Automated deadline reminders are typically triggered by? A) Manual entry each week B) Predefined rules based on due dates (correct) C) Random intervals D) Manager approval. Q4: Which scenario best illustrates a task dependency? A) Two tasks due on the same day B) Testing cannot begin until development is complete (correct) C) A task assigned to a remote employee D) A recurring weekly status report.
#05 · Medical education platform generating questions from a clinical article for nursing students
Input
Article excerpt: 'Sepsis is defined as life-threatening organ dysfunction caused by a dysregulated host response to infection. Early identification relies on the Sequential Organ Failure Assessment (SOFA) score. The Sepsis-3 definition replaced the older SIRS criteria. First-line treatment includes broad-spectrum antibiotics administered within one hour of recognition, aggressive IV fluid resuscitation, and vasopressors for refractory hypotension.' Generate 3 multiple-choice questions at advanced difficulty for nursing students. Note: outputs require clinical review before use.
Expected output
Q1: According to Sepsis-3, what defines sepsis? A) Presence of infection with a fever above 38C B) Two or more SIRS criteria met C) Life-threatening organ dysfunction from a dysregulated host response to infection (correct) D) Bacteremia confirmed by blood culture. Q2: Which scoring tool is central to the Sepsis-3 definition? A) APACHE II B) Glasgow Coma Scale C) SOFA score (correct) D) NEWS2. Q3: What is the recommended window for administering broad-spectrum antibiotics after sepsis recognition? A) Within 30 minutes B) Within 1 hour (correct) C) Within 3 hours D) Before vasopressors are initiated.

Common mistakes to avoid

  • Pasting the entire article without trimming

    Very long inputs push the model toward generating surface-level or repetitive questions because it tries to cover too much ground. Trim your input to the 300 to 600 words that contain the most testable content. You will get tighter, more varied questions and use fewer tokens.

  • Not asking for plausible distractors

    Default multiple-choice distractors from AI are often obviously wrong, which makes the quiz too easy and defeats its purpose. Prompt the model explicitly to 'write distractors that are plausible but clearly incorrect on close reading.' This produces harder questions that actually test comprehension rather than elimination by absurdity.

  • Skipping human review before publishing

    AI can misread a nuanced article and generate a question with a factually wrong correct answer, especially in technical domains. Every output should be reviewed by someone who knows the source material before the quiz goes live. This is not optional for high-stakes or professional contexts.

  • Ignoring question diversity across a single article

    If you ask for ten questions from one article without constraints, the model often clusters around the first or most prominent topic and ignores later sections. Ask for 'two questions from each of the following sections' or break the article into chunks and run separate prompts. This distributes coverage and avoids repetitive question clusters.

  • Using vague difficulty labels with no context

    Difficulty is relative to the audience. Telling the model 'make it hard' without specifying who is taking the quiz produces unpredictable results. Always pair a difficulty label with an audience descriptor so the model has an actual reference point for calibration.

Related queries

Frequently asked questions

Can AI generate quiz questions from a PDF or URL?

Most AI chat interfaces require you to paste text directly. However, several tools built on top of models like GPT-4 accept PDF uploads or can read URLs via plugins or integrations. If you are using a raw API or chat interface, copy the article text and paste it into the prompt. For large documents, break them into sections and run multiple prompts.

How many questions can I generate from one article?

As a rule of thumb, a 500-word article can support 5 to 8 distinct questions before the model starts repeating itself or generating trivial items. A 1,500-word article can support 15 to 20 questions. If you need more, ask the model to vary question types across recall, comprehension, and application levels to extend range without repetition.

What is the best AI tool for generating quizzes from articles?

General-purpose models like GPT-4, Claude, and Gemini all handle this task well when prompted correctly. Specialized tools like Quizgecko, Quizizz AI, and Formative are purpose-built for this workflow and often include export options for platforms like Google Forms or LMS systems. The right choice depends on whether you need raw output or direct integration with a teaching or testing platform.

Can I generate quizzes from articles in languages other than English?

Yes. Current large language models handle quiz generation in most major languages including Spanish, French, German, Portuguese, Japanese, and Chinese with reasonable quality. Paste the article in the target language and write your prompt in that same language for best results. Quality may drop slightly for lower-resource languages.

How do I make sure the AI does not hallucinate quiz answers?

Grounding the model in the source text reduces hallucination significantly. Paste the article directly into the prompt and include an instruction such as 'only generate questions based on information explicitly stated in the text above.' After generation, cross-check every correct answer against the source article before using the quiz. Never skip that review step.

Can I export AI-generated quiz questions to Google Forms or an LMS?

Not directly from a plain AI chat interface, but several intermediary tools and workflows make this easy. Tools like Quizgecko export to Google Forms natively. For LMS platforms like Canvas or Moodle, you can ask the AI to format output as QTI-compatible text, then import it. Alternatively, a simple spreadsheet copy-paste into Google Forms works fine for smaller quizzes.