Skip to content
View Xiao9905's full-sized avatar
  • Tsinghua University
  • Beijing, China

Highlights

  • Pro

Organizations

@THUDM
Block or Report

Block or report Xiao9905

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Xiao9905/README.md

Hi, welcome to my Github 👋

I am Xiao Liu, a third-year PhD student in Tsinghua University since 2021.

  • 🔭 Interested in Machine Learning, Data Mining, NLP and Knowledge Graph.

  • 🌱 Find my up-to-date publication list in Google Scholar! Some of my proud leading works:

    Large Language Model (LLM) Training and Prompt Learning
    • P-tuning and P-tuning v2 (ACL'22): pioneer works on prompt tuning
    • GLM-130B (ICLR'23): an open bilingual (Enligsh & Chinese) pre-trained model with 130 billion parameters based on GLM (ACL'22); better than GPT-3 175B on LAMBADA and MMLU.
    • ChatGLM-6B & ChatGLM2-6B & ChatGLM3-6B: an open bilingual dialogue language model that requires only 6GB to run. Receiving GitHub stars, GitHub stars, and GitHub starsGitHub Stars!
    • WebGLM (KDD'23): an efficient web-enhanced question answering system based on GLM-10B, outperforming WebGPT-13B and approaching WebGPT-175B performance in human evaluation.
    • ChatGLM-Math: employing self-critique with RFT and DPO to enable SOTA mathematical capabilities wihtouth compromising language abilities.
    Foundational Agents For Real-world Challenging Missions
    • AgentBench (ICLR'24): the first systematic multi-dimensional benchmark to evaluate LLMs as Agents in 8 distinct environments deriving from real-world practical missions. Find LLM-as-Agent demos at llmbench.ai/agent!
    Alignment and Scalable Oversights over LLMs and Diffusers
    • ImageReward (NeurIPS'23): the first general-purpose text-to-image human preference reward model (RM) for RLHF, outperforming CLIP/BLIP/Aesthetic by 30% in terms of human preference prediction.
    • BPO (Black-box Prompt Optimization): a novel direction to align LLMs via preference-aware prompt optimization. Improving ChatGPT, Claude, LLaMA on human preference's win rates by 20%+ without training them.
    • AlignBench: the first comprehensive benchmark on evaluating LLMs' Chinese alignment, deriving from ChatGLM's online real scenarios. Submit your LLMs to acquire CritiqueLLM's judgement on AlignBench on llmbench.ai/align!
    • CritiqueLLM: scaling LLM-as-Critic for scalable oversights on LLM alignment. A series of strong critqiue LLMs ranging from 6B to 66B.
    Self-supervised Learning and Reasoning
  • 🤔 Dedicated to building next-generation of AI systems via both Large Pre-trained Model and Symbolic Agent Reasoning.

  • 💬 Feel free to drop me an email for:

    • Any form of collaboration
    • Any issue about my works or code
    • Interesting ideas to discuss or just chatting

Pinned

  1. THUDM/GLM-130B THUDM/GLM-130B Public

    GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)

    Python 7.6k 609

  2. THUDM/ChatGLM-6B THUDM/ChatGLM-6B Public

    ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型

    Python 39.3k 5.1k

  3. THUDM/P-tuning THUDM/P-tuning Public

    A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.

    Python 891 113

  4. THUDM/P-tuning-v2 THUDM/P-tuning-v2 Public

    An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks

    Python 1.9k 187

  5. THUDM/ChatGLM2-6B THUDM/ChatGLM2-6B Public

    ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型

    Python 15.5k 1.8k

  6. THUDM/AgentBench THUDM/AgentBench Public

    A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)

    Python 1.8k 116