From instruction fine-tuning to mathematical reasoning capabilities, explore the potential of large models | The 11th issue of the large model series activities on September 14

Click on the blue text

b65383f17967c20ba627acd6746907cb.jpeg

Follow us

AI TIME welcomes every AI enthusiast to join!

72e258b88471689a598edeaff66434a2.jpeg

Bilibili live channel

Scan the QR code to follow AI TIME Bilibili official account to make an appointment for live streaming

ba8302ccf68a712a2cf3cda94362668d.gif

13:30—13:50

Yan Jun

Virtual Prompt Injection for Instruction-Tuned Large Language Models

13:50—14:10

Ning Xuefei

SoT: An attempt to accelerate LLM using parallel decoding

14:10—14:30

Zhang Ruiqi

Trained Transformers Learn Linear Models In-Context

14:30—14:50

Wei Lai

InstructionGPT-4: A 200-Instruction Paradigm for Fine-Tuning MiniGPT-4

14:50—15:10

Yuan Zheng

Scaling relationship on Learning Mathematical Reasoning with Large Language Models

15:10—16:00

Panel

1. How to comprehensively evaluate the capabilities of large language models? How to balance the power and security of large language models?

2. Can large model fine-tuning methods effectively improve the performance of the model on specific tasks? What are the potential limitations in practical applications?

3. How do large models perform in specific abilities (such as mathematical reasoning)? How to give full play to the advantages of large models in specific capabilities?

Guest introduction

bc6d24b4ab167a5dcea8b971474cb4f9.jpeg

Yan Jun

A fifth-year doctoral student in the Department of Computer Science at the University of Southern California. His supervisor is Professor Xiang Ren. His research field is trustworthy natural language processing. Currently, the main focus is on the security of large oracle models, including data poisoning attacks and model robustness.

Personal homepage: https://junyann.github.io/

c9188421a5093be6e6585b9e3c4fc7af.png

Ning Xuefei

Postdoctoral fellow at Tsinghua University, co-supervisor is Professor Wang Yu. His research area is efficient machine learning. Currently the main focus is on compression and acceleration of generative models.

Personal homepage: https://www.ningxuefei.cc/

0366760ee0b5bdaf52fbb8727e0ec60f.jpeg

Zhang Ruiqi

A second-year doctoral student in the Department of Statistics at the University of California, Berkeley, working mainly under the supervision of Professor Peter L. Bartlett. The research areas are mainly theoretical deep learning and theoretical reinforcement learning. Currently focusing on Transformer, large language models and theories based on context learning (In-Context Learning).

Personal homepage: https://rqzhangberkeley.github.io/

0f187488afb66725b885f1b83630fdbf.jpeg

Wei Lai

A fourth-year undergraduate student at Shanghai Jiao Tong University. His research fields are multi-modal large models and natural language processing.

Personal homepage: https://waltonfuture.github.io/

e1ba1e684d6fd16044063af624908924.jpeg

Yuan Zheng

PhD from the Statistics Center of Tsinghua University, and serves as a senior algorithm engineer at Alibaba Damo Academy. The main research direction is alignment and logical reasoning in large models.

Recommended articles from past issues

853199c308827710a7e0eadfebf4e62e.jpeg

- Follow us and remember the star -

 About AI TIME 

AI TIME originated in 2019, aiming to carry forward the spirit of scientific speculation, invite people from all walks of life to explore the essential issues of artificial intelligence theory, algorithms and scenario applications, strengthen the collision of ideas, and connect global AI scholars, industry experts and enthusiasts, hoping to In the form of a debate, we explore the contradiction between artificial intelligence and the future of mankind, and explore the future of the field of artificial intelligence.

To date, AI TIME has invited more than 1,300 speakers at home and abroad, held more than 600 events, and been watched by more than 6 million people.

6bb8e9a1f9019b13aa14bf52facf54b7.png

I know you

look in

oh

~

e143d7a825d8779e121fd5c70d524d96.gif

Click to read the original text  and book a live broadcast!

Guess you like

Origin blog.csdn.net/AITIME_HY/article/details/132784504