Skip to content

rickerliang/fromLLMtoAGI

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 

Repository files navigation

fromLLMtoAGI

Leaderboard

Dataset Investigate

Transformer, In-context Learning

Prompt, Agent, Methodology

AGI, Application

Language Model

Training Technique

  • Training Transformers with 4-bit Integers
  • GKD: Generalized Knowledge Distillation for Auto-regressive Sequence ModelsπŸ’‘πŸš€
    • $L_{GKD}(\theta):=$
      • $(1-\lambda)E_{(x,y)\sim(X,Y)}[\mathcal{D}(p_T\parallel p^{\theta}_S(y|x)]+$
      • $\lambda E_{x \sim X} \big[E_{y\sim ps(\cdot|x)}[\mathcal{D}(p_T\parallel p^{\theta}_S(y|x)]\big]$
    • when approximating $P(\mathcal{C})$ using a parameterized distribution $Q_\theta(\mathcal{C})$, minimizing the reverse and forward KL under model under-specification results in mean and mode-seeking behavior.
      • $D_{KL}(P\parallel Q)=\displaystyle \sum_{c \in C}P(c)\log \frac{P(c)}{Q(c)}$
      • $D_{RKL}(P\parallel Q):=D_{KL}(Q\parallel P)$
      • $D_{JSD[\beta]}(P\parallel Q)=\beta D_{KL}(P\parallel \beta P + (1-\beta)Q)+(1-\beta)D_{KL}(Q\parallel \beta P + (1-\beta)Q)$.
    • Supervised FT; Supervised KD; On-policy KD
  • Improving Language Plasticity via Pretraining with Active Forgetting, we introduce a simple active forgetting mechanism, that resets the token embeddings at regular intervals, while leaving all other parameters untouched throughout pretraining.

Implementation, Code Repo

Flow

  • LangChain, in-context learning, prompt template, chain of thought, toolformer, ReAct, ToT
  • LangFlow
  • Flowise
  • Chat UI, A chat interface using open source models, eg OpenAssistant.

Inference of LLaMA model in pure C/C++

Model Weights

LLMs-In-China

Prompt Engineer, Guide

Datasets

About

paper collection of improve LLM decision making

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published