generated from HugoBlox/theme-research-group
-
Notifications
You must be signed in to change notification settings - Fork 2
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
1 changed file
with
82 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,82 @@ | ||
--- | ||
# Documentation: https://wowchemy.com/docs/managing-content/ | ||
|
||
title: "Seminar: \"Knowledge Representation and Large Language Models\"" | ||
# event: | ||
# event_url: | ||
location: Abacws | ||
# address: | ||
# street: | ||
# city: | ||
# region: | ||
# postcode: | ||
# country: | ||
summary: Talks by [Zhiwei Hu]() () and [Xiaoqi Han]() () | ||
abstract: "This will be a joint seminar from the NLP and KRR research groups. There will be two short presentations by external visitors. See details below." | ||
|
||
# Talk start and end times. | ||
# End time can optionally be hidden by prefixing the line with `#`. | ||
date: 2024-05-16T13:00:00Z | ||
date_end: 2024-05-16T14:00:00Z | ||
all_day: false | ||
|
||
# Schedule page publish date (NOT event date). | ||
publishDate: 2024-05-10T00:00:00Z | ||
|
||
authors: [camachocolladosj] | ||
tags: [] | ||
|
||
# Is this a featured event? (true/false) | ||
featured: false | ||
|
||
# Featured image | ||
# To use, add an image named `featured.jpg/png` to your page's folder. | ||
# Focal points: Smart, Center, TopLeft, Top, TopRight, Left, Right, BottomLeft, Bottom, BottomRight. | ||
image: | ||
caption: "" | ||
focal_point: "" | ||
preview_only: false | ||
|
||
# Custom links (optional). | ||
# Uncomment and edit lines below to show custom links. | ||
# links: | ||
# - name: Follow | ||
# url: https://twitter.com | ||
# icon_pack: fab | ||
# icon: twitter | ||
|
||
# Optional filename of your slides within your event's folder or a URL. | ||
url_slides: | ||
|
||
url_code: | ||
url_pdf: | ||
url_video: | ||
|
||
# Markdown Slides (optional). | ||
# Associate this event with Markdown slides. | ||
# Simply enter your slide deck's filename without extension. | ||
# E.g. `slides = "example-slides"` references `content/slides/example-slides.md`. | ||
# Otherwise, set `slides = ""`. | ||
slides: "" | ||
|
||
# Projects (optional). | ||
# Associate this post with one or more of your projects. | ||
# Simply enter your project's folder or file name without extension. | ||
# E.g. `projects = ["internal-project"]` references `content/project/deep-learning/index.md`. | ||
# Otherwise, set `projects = []`. | ||
projects: [] | ||
--- | ||
|
||
**Invited Speaker 1:** [Zhiwei Hu]() | ||
|
||
**Title:** Knowledge Representation Learning and Knowledge Distillation | ||
|
||
**Abstract:** Knowledge Representation Learning (KRL) focus on learning meaningful representations of knowledge from raw data. In knowledge representation learning, the goal is to create models that can capture and represent the underlying structure and semantics of data in a way that is useful for various AI tasks such as reasoning, recommendation, and prediction. One of the main challenges in KRL is capturing the complexity and diversity of real-world knowledge, especially effectively integrating different types of information. In the era of Large Language Models (LLMs), Knowledge Distillation (KD) emerges as a pivotal methodology for transferring advanced capabilities from leading proprietary LLMs, such as GPT-4, to their open-source counterparts like LLaMA and Mistral. How to elegantly make Small Language Models (SLMs) have the ability to compete with large models is a promising research direction. | ||
|
||
------------------------------------- | ||
|
||
**Invited Speaker 2:** [Xiaoqi Han]() | ||
|
||
**Title:** Model Editing in Pre-trained Language Models | ||
|
||
**Abstract:** Even with their impressive capabilities, Large Language Models (LLMs) such as ChatGPT are not immune to issues of factual inaccuracies or logical inconsistencies. A key concern is how to seamlessly update these LLMs to correct mistakes without resorting to exhaustive retraining or continuous training procedures, both of which can require significant computational resources and time. Thus, the ability to edit LLMs offers an efficient solution to modify a model’s behavior, notably within a specific area of interest, without negatively impacting its performance on other tasks. Through this presentation, we seek to present a systematic and current overview of cutting-edge methods and provide insights into real-world applications and engage in discussions about future research directions. |