Montreal, Canada

McGill University

McGill University is listed as QS 2026 rank 27. McGill University has 5 source-backed AI policy claim records from 5 official source attributions. The public record preserves original-language evidence snippets, source URLs, snapshot hashes, confidence, and review state.

Short answer

v1 public contract

McGill University is listed as QS 2026 rank 27. McGill University has 5 source-backed AI policy claim records from 5 official source attributions. The public record preserves original-language evidence snippets, source URLs, snapshot hashes, confidence, and review state.

Policy statusReviewed evidence-backed recordReview: Agent reviewedEvidence-backed claims5Reviewed5Candidate0Official sources5

This reference record summarizes visible public data only. Official sources and original-language evidence remain canonical; confidence is separate from review state.

This page is not legal advice, not academic integrity advice, and not an official university statement unless a linked source is the university's own official page.

Policy profile

Deterministic source-backed dimensions derived from this record's public claims.

Coverage score85/100Coverage labelbroad public coverageReview: Machine candidateAnalysis confidence82%

Policy profile rows are machine-candidate derived metadata. They are not final policy conclusions; inspect the linked claim evidence before reuse.

Analysis page-quality metadata is available at /api/public/v1/analysis/page-quality.json.

AI disclosure

No source-backed public claim about AI disclosure or acknowledgement is present in this profile.

The current public tracker record does not contain claim evidence about disclosing, acknowledging, citing, or declaring AI use.

Not MentionedMachine candidateConfidence0%Evidence0Sources0

Academic integrity

McGill University has 1 source-backed public claim for academic integrity; deterministic analysis status: required.

RequiredMachine candidateConfidence82%Evidence1Sources1

Research guidance

McGill University has 1 source-backed public claim for research guidance; deterministic analysis status: restricted.

RestrictedMachine candidateConfidence83%Evidence1Sources1

Coverage score measures breadth of public, source-backed coverage only. It is not a policy quality, strictness, legal adequacy, safety, or compliance score.

Evidence-backed claims

5 reviewed evidence-backed public claim

Ai Tool Treatment

McGill lists Microsoft 365 Copilot Chat as an available AI tool for staff, faculty, and students, and says a secure version with enterprise data protection is available for all McGill users.

Review: Agent reviewedConfidence98%

Normalized value: microsoft_copilot_commercial_data_protection_approved

Original evidence

Evidence 1
Microsoft 365 Copilot Chat is an AI-powered feature integrated into Microsoft Edge and accessible through other browsers. Microsoft 365 Copilot Chat can answer questions, generate content, condense long texts and more. A secure version with enterprise data protection is available for all McGill users​. Audience: Staff, Faculty, and Students Price: Free

Ai Tool Treatment

McGill explicitly rejects DeepSeek AI for McGill-managed or research-funded devices, rejects Read.AI and other AI meeting bots for McGill use, and says tools not mentioned in the available AI tools list are automatically considered rejected.

Review: Agent reviewedConfidence98%

Normalized value: deepseek_readai_rejected_prohibited

Original evidence

Evidence 1
If a tool is not mentioned in the "Available AI tools" list, it is automatically considered rejected , even if it is not listed among these prohibited tools DeepSeek AI: This tool has raised serious data exposure risks and prompt injection vulnerabilities. Its use is not permitted for any McGill-managed or research-funded device. This decision follows cybersecurity directives from the Government of Quebec and the Government of Canada.

Privacy

McGill guidance says users should mitigate potential privacy concerns by removing personally identifying information when using AI tools, be careful with sensitive or restricted material, and avoid using Personal Health Information (PHI) or Payment Card Industry (PCI) data with AI tools.

Review: Agent reviewedConfidence97%

Normalized value: remove_pii_phi_pci_from_ai_use

Original evidence

Evidence 1
Mitigate potential privacy concerns by removing personally identifying information (e.g., names, email addresses, phone numbers). For example, when writing a prompt to draft an email to Joe Smith, replace "Joe Smith" with "XYZ."

Academic Integrity

McGill's Provost-endorsed principles state that instructors remain responsible for comporting themselves according to the highest standards of academic integrity in their use of generative AI tools. Instructors must be explicit in course outlines about the expectations for use of generative AI tools and may set limits on their use in assessment tasks.

Review: Agent reviewedConfidence96%

Normalized value: instructor_responsible_academic_integrity_ai

Original evidence

Evidence 1
Fourth principle: Instructors remain responsible for comporting themselves according to the highest standards of academic integrity in their use of generative AI tools. Instructors maintain responsibility and accountability for all of their instructional materials whether independently created, third-party generated, supported by generative AI tools, or derived from other resources. Instructors must be explicit in course outlines about the expectations for use of generative AI tools and may set limits on their use in assessment tasks.

Teaching

McGill recommends that instructors explain to students in their course outline what the appropriate use or non-use is of generative AI tools in the context of that course. The use or non-use of these tools should align with the learning outcomes associated with the course.

Review: Agent reviewedConfidence95%

Normalized value: course_outline_statement_recommended

Original evidence

Evidence 1
There should be no default assumption as to the use of generative AI tools. Therefore, McGill recommends that instructors explain to students in their course outline what the appropriate use or non-use is of generative AI tools in the context of that course. The use or non-use of these tools should align with the learning outcomes associated with the course. For this reason, instructors will need to write their own context-appropriate course outline statements.

Candidate claims

0 machine or needs-review claim

Candidate claims are not final policy conclusions. They preserve source URL, source snapshot hash, evidence, confidence, and review state so the record can be audited before review.

Official sources

5 source attribution

Change log

Source-check timeline and diff-style claim/evidence preview.

View the public change record for this university, including source snapshot hashes, claim review states, and a diff-style preview of current source-backed evidence.

Last checkedMay 11, 2026Last changedMay 10, 2026Open change log

Corrections and missing evidence

Corrections create review tasks and do not directly change this public record.

If an official source is missing, stale, moved, blocked, or incorrectly summarized, submit a source URL, policy change report, or institution correction for review. Corrections must preserve source URLs, source language, original evidence, review state, and audit history.

Back to universities