Platform

  • Search
  • Seminars
  • Conferences
  • Jobs

Resources

  • Submit Content
  • About Us

© 2025 World Wide

Open knowledge for all • Started with World Wide Neuro • A 501(c)(3) Non-Profit Organization

Analytics consent required

World Wide relies on analytics signals to operate securely and keep research services available. Accept to continue, or leave the site.

Review the Privacy Policy for details about analytics processing.

World Wide
SeminarsConferencesWorkshopsCoursesJobsMapsFeedLibrary
Back to SeminarsBack
Seminar✓ Recording AvailableArtificial Intelligence

Improving Language Understanding by Generative Pre Training

Amgad Hasan
Schedule
Tuesday, April 23, 2024

Showing your local timezone

Schedule

Tuesday, April 23, 2024

2:00 PM Europe/Athens

Watch recording
Host: LLM Paper Club

Watch the seminar

Recording provided by the organiser.

Event Information

Domain

Artificial Intelligence

Original Event

View source

Host

LLM Paper Club

Duration

70 minutes

Abstract

Natural language understanding comprises a wide range of diverse tasks such as textual entailment, question answering, semantic similarity assessment, and document classification. Although large unlabeled text corpora are abundant, labeled data for learning these specific tasks is scarce, making it challenging for discriminatively trained models to perform adequately. We demonstrate that large gains on these tasks can be realized by generative pre-training of a language model on a diverse corpus of unlabeled text, followed by discriminative fine-tuning on each specific task. In contrast to previous approaches, we make use of task-aware input transformations during fine-tuning to achieve effective transfer while requiring minimal changes to the model architecture. We demonstrate the effectiveness of our approach on a wide range of benchmarks for natural language understanding. Our general task-agnostic model outperforms discriminatively trained models that use architectures specifically crafted for each task, significantly improving upon the state of the art in 9 out of the 12 tasks studied. For instance, we achieve absolute improvements of 8.9% on commonsense reasoning (Stories Cloze Test), 5.7% on question answering (RACE), and 1.5% on textual entailment (MultiNLI).

Topics

commonsense reasoningdiscriminative fine-tuningdocument classificationgenerative pre-traininglanguage modelnatural language understandingquestion answeringsemantic similaritytask-aware input transformationstextual entailment

About the Speaker

Amgad Hasan

Contact & Resources

@AmgadGamalHasan

Follow on Twitter/X

twitter.com/AmgadGamalHasan

Related Seminars

Seminar60%

Computational modelling of ocular pharmacokinetics

ai

Pharmacokinetics in the eye is an important factor for the success of ocular drug delivery and treatment. Pharmacokinetic features determine the feasible routes of drug administration, dosing levels a

Apr 21, 2025
School of Pharmacy, University of Eastern Finland
Seminar60%

Guiding Visual Attention in Dynamic Scenes

neuro

Jan 20, 2025
Haifa U
Seminar60%

TBD

neuro

Jan 20, 2025
King's College London
January 2026
Full calendar →