Why AI engines ignore most university websites
ChatGPT, Perplexity and Google AI Overviews do not rank web pages. They synthesise answers from vast corpora and cite sources they deem reliable, structured and factually verifiable. Most college content still fails on all three counts.
In Ireland, only 8% of ChatGPT responses about Irish private higher education mention a specific college (Source: Skolbot GEO Monitoring summary for Ireland, Feb 2026). Public universities such as Trinity College Dublin and UCD are cited far more often than private colleges, and Perplexity only partially closes the gap. Your content exists online, but AI engines often cannot extract anything citable from it.
Four factors separate citable content from invisible content: technical structure, data specificity, source authority and answer clarity. Each one is within your college's control, and each matters especially in Ireland because local providers compete directly against larger UK institutions in the same English-language query space.
What makes content "citable" by an LLM
Structure beats length every time
An LLM does not read a page from top to bottom. It extracts answer fragments from recognisable patterns: question-answer pairs, comparison tables, definitions framed by semantic markup. A page with no obvious structure is less likely to be cited than a shorter page with descriptive H2s, a data table and a marked-up FAQ.
Structural signals that LLMs exploit:
| Signal | Impact on citability | Implementation difficulty |
|---|---|---|
| FAQ marked up in JSON-LD | High: direct extraction | Low |
| Tables with descriptive headers | High: comparable data | Low |
| H2/H3 phrased as questions | Medium: semantic matching | Low |
| Schema.org EducationalOrganization | High: entity identification | Medium |
| Sourced numerical data | High: verifiable facts | Medium |
Specificity wins over superlatives
Content claiming "our college offers a supportive learning environment" will never be cited. Content stating "89% of 2025 graduates secured employment within 6 months, tuition EUR 7,250 per year, QQI-validated Level 8 award, 236 survey respondents" gives AI engines factual material they can reuse.
Data points AI engines actively look for on Irish college websites:
- Graduate outcomes, with methodology and sample size
- Tuition fees and payment structure by programme
- Official validation and qualification information (QQI, NFQ)
- Rankings or sector comparisons where available
- CAO route, campus location, work placement details and international student support
4 techniques to make your content citable
1. Implement Schema.org on key pages
Colleges with structured Schema.org markup achieve an average of +12 visibility points in AI engine responses (Source: Skolbot GEO Monitoring, 500 queries x 6 countries x 3 AI engines, Feb 2026). The EducationalOrganization markup turns your institution into an identifiable entity. The Course schema does the same for each Level 6, 7, 8 or 9 programme.
For the full technical implementation guide, see our Schema.org guide for universities.
The minimum implementation covers three schemas:
- EducationalOrganization on your homepage and About page
- Course on each programme page
- FAQPage on FAQ pages and blog articles containing Q&A sections
The fields that matter most to LLMs: accreditation, numberOfStudents, aggregateRating, alumni and programPrerequisites. These are the data points ChatGPT cross-references with the Central Applications Office, Quality and Qualifications Ireland and the Higher Education Authority to validate reliability.
2. Structure every page with direct answers
AI engines operate on a question-answer model. To maximise citation probability, each H2 should pose or imply a question, and the first 1-2 sentences must answer it directly. The rest of the paragraph adds context and nuance.
Before:
"Our college is known for practical teaching and close industry links across Dublin."
After:
"The Level 8 Honours Bachelor's in Computing at [College] runs for 4 years, costs EUR 7,250 per year, offers a mandatory work placement and reports 89% graduate employment within 6 months (2025 survey, 236 respondents). The programme is QQI-validated, referenced on the NFQ and available through the CAO or direct entry route."
The second version contains six verifiable data points. The first contains none.
3. Build comparison tables with your data
Tables are the most extractable format for an LLM. A clean table with clear headers and numerical data will be preferred over a narrative paragraph containing the same information.
Example of a citable table for a programme page:
| Criterion | Honours Bachelor's in Computing | MBA |
|---|---|---|
| Duration | 4 years | 2 years |
| Annual tuition | EUR 7,250 | EUR 12,900 |
| Employment rate at 6 months | 89% | 94% |
| Median starting salary | EUR 34,000 | EUR 52,000 |
| Qualification level | NFQ Level 8 | NFQ Level 9 |
| Intake size | 120 | 35 |
Publish these tables on your programme pages, not buried in downloadable prospectuses. AI engines do not reliably parse PDFs hidden behind enquiry forms.
4. Add marked-up FAQ sections
An FAQ section serves two purposes: it answers the questions prospects ask AI engines, and FAQPage JSON-LD markup enables structured extraction.
The common mistake is writing brand FAQs ("Why choose our college?") instead of informational FAQs ("What CAO points are typical?", "Is the programme QQI-validated?", "What are the tuition fees for international students?"). AI engines favour the latter.
To diagnose your current situation, use our ChatGPT visibility diagnostic tool.
How to measure whether your content is being cited
Checking whether AI engines cite your college requires a systematic approach.
3-step testing protocol
-
Identify your 20 strategic queries: the questions prospects ask about your institution, programmes, city and sector. Examples: "best private college in Dublin for business", "QQI validated computing degree Ireland", "fees [college] 2026".
-
Test across 3 AI engines: submit each query to ChatGPT, Perplexity and Gemini. Record whether your institution is mentioned, whether the information is accurate, and whether sources are cited.
-
Track monthly evolution: LLM corpora evolve. Content published or updated today may take 4-8 weeks to be integrated. Measure monthly to identify trends.
Key metrics to track
| Metric | Target | Measurement frequency |
|---|---|---|
| Mention rate (brand queries) | >80% | Monthly |
| Mention rate (generic queries) | >20% | Monthly |
| Accuracy of cited information | 100% | Monthly |
| Sources cited (Perplexity) | >2 pages from your site | Monthly |
For a complete methodology on tracking your AI visibility, see our GEO guide for universities.
Before and after: optimising a programme page
An Irish private college wanted ChatGPT to mention its Level 8 business degree when prospects asked "best private business college in Dublin".
Before optimisation:
- Programme page without Schema.org
- Narrative text with no numerical data
- No FAQ section
- No comparison table
Result: ChatGPT never mentioned the college for this query.
After optimisation:
Coursemarkup witheducationalLevel,provider,accreditation- Table with tuition, duration, graduate outcomes and qualification level
- Marked-up FAQ with 5 questions (CAO, QQI validation, work placement, fees, entry routes)
- Link to QQI and CAO as authoritative references
Result at 8 weeks: ChatGPT cites the college in 3 out of 5 responses for the same query. Perplexity links to the programme page as a source in 4 out of 5 cases.
This correlation between structured markup and citability holds across our full panel. For the technical mechanisms, our article on structured data for universities details each schema.
FAQ
How do I check if ChatGPT already cites my university?
Test 20 strategic queries directly in ChatGPT. Record every mention of your institution, the accuracy of the data and the presence of links. Repeat monthly to track changes. Perplexity is simpler to audit because it displays its sources beneath each response.
How long before optimised content gets cited?
Between 4 and 8 weeks after publication or modification. Perplexity tends to react faster because it queries the live web. But the page still needs strong structured data and credible Irish sources to be reused consistently.
Is Schema.org markup enough to get cited?
No, but it is necessary. Markup identifies your institution as a verifiable entity. Without it, AI engines must extract this information from raw text, with a high error rate. Markup alone does not replace specific, data-rich, well-structured content.
Should I optimise for ChatGPT or Perplexity first?
Both, as the techniques overlap. But if you must prioritise, start with Perplexity: it cites sources explicitly, making tracking straightforward. Optimisations that work for Perplexity also benefit ChatGPT.
Which pages on my site should I optimise first?
Your homepage, the three most-enquired programme pages, your admissions page and your FAQ page. In Ireland, the highest-priority pages are usually the ones that explain QQI status, NFQ level, CAO route, fees and graduate outcomes.
Is your university cited by ChatGPT? Test your AI visibility for free


