Skip to content
View viai957's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report viai957

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
viai957/README.md

Lately I’ve been working with LLMs for the past few months across every part of the technology stack: from low-level (pytorch and ggml) to abstraction and orchestration layers (LangChain, Llama Index); from OpenAI to Claude to Llama and back; from prompt engineering to agent planning; from development to deployment, and now I’m working on monitoring and evaluation, pretraining.

I’ve also been talking with leaders across every part of the ecosystem: from the 10BedICU, OpenAI core team to futuristic indie projects at the AGI House to old-school businesses like Pratham Books who don’t even have a tech team that are desperate to use the technology.

One thing stands out across the landscape: LLMs are a new kind of product that have new kinds of user behavior, business expectations, and engineering requirements. I want to talk about a specific comment I hear a lot of people talk about, and how it creates a symphony of confusion and opportunity for designers, operators and engineers to build great products.


📕 My Latest Projects:

  • LexiGenesis An repo to explain and understand how the evolution of LLM goes from pretraining -> finetuning -> benchmarking -> deployment -> application development over it

Pinned Loading

  1. LexiGenesis LexiGenesis Public

    LexiGenesis: Charting the birth and growth of an LLM. From pretraining to fine-tuning and industry deployment, this repo is the epicenter of AI evolution. 🌟

    Jupyter Notebook 1

  2. llama-inference llama-inference Public

    A simple implementation of Llama 1, 2. Llama Architecture built from scratch using PyTorch all the models are built from scratch that includes GQA (Grouped Query Attention) , RoPE (Rotary Positiona…

    Python 12 3

  3. DDPG-Robotic-Arm DDPG-Robotic-Arm Public

    In this project, a double-jointed arm can move to target locations. A reward of +0.1 is provided for each step that the agent's hand is in the goal location. Thus, the goal of your agent is to main…

    Jupyter Notebook

  4. MULTI-AGENT-TENNIS MULTI-AGENT-TENNIS Public

    In this Project, two agents control rackets to bounce a ball over a net. If an agent hits the ball over the net, it receives a reward of +0.1. If an agent lets a ball hit the ground or hits the bal…

    ASP

  5. Optimal-Portfolio-Transactions Optimal-Portfolio-Transactions Public

    We consider the execution of portfolio transactions with the aim of minimizing a combination of risk and transaction costs arising from permanent and temporary market impact. As an example, assume …

    Python 31 9

  6. webkit-vulnerability webkit-vulnerability Public

    CVE-2016-4657 web-kit vulnerability for ios 9.3, nintendo switch browser vulnerability

    HTML 9 2