Stars
[NeurIPS 2023] Reflexion: Language Agents with Verbal Reinforcement Learning
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
An open source implementation of Microsoft's VALL-E X zero-shot TTS model. Demo is available in https://plachtaa.github.io/vallex/
Official implementation for "Automatic Chain of Thought Prompting in Large Language Models" (stay tuned & more will be updated)
Official implementation for "Multimodal Chain-of-Thought Reasoning in Language Models" (stay tuned and more will be updated)
Shopping Queries Dataset: A Large-Scale ESCI Benchmark for Improving Product Search
🐝 GPTSwarm: LLM agents as (Optimizable) Graphs
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
This is the official code for the paper CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning (NeurIPS22).
An open-source RAG-based tool for chatting with your documents.
[TMLR] A curated list of language modeling researches for code and related datasets.
Train transformer language models with reinforcement learning.
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
DetectLLM: Leveraging Log Rank Information for Zero-Shot Detection of Machine-Generated Text
[ACL 2024 Demo] SeaLLMs - Large Language Models for Southeast Asia
Is ChatGPT Good at Search? LLMs as Re-Ranking Agent [EMNLP 2023 Outstanding Paper Award]
A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM
This is the repository for our paper "INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning"
SimXNS is a research project for information retrieval. This repo contains official implementations by MSRA NLC team.
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
The official codes for "PMC-LLaMA: Towards Building Open-source Language Models for Medicine"
CIKM'21: JPQ substantially improves the efficiency of Dense Retrieval with 30x compression ratio, 10x CPU speedup and 2x GPU speedup.
Training & evaluation library for text-based neural re-ranking and dense retrieval models built with PyTorch
Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation
[SIGIR 2022] The official repo for the paper "Curriculum Learning for Dense Retrieval Distillation".