Skip to content

Latest commit

 

History

History
146 lines (120 loc) · 13 KB

readme.md

File metadata and controls

146 lines (120 loc) · 13 KB

Table of Contents 📚

Papers 📎

This section contains a list of research papers associated with different AI technology and techniques. These are organized by chapter to help you navigate. The reader is not expected to know these, but as with most things, it is always good to go deeper and grok some of these concepts for a better and fuller understanding.

Chapter 1 - Introduction to Generative AI 💾

# Paper Title Link
1 The brief history of artificial intelligence: The world has changed fast – what might be next? https://ourworldindata.org/brief-history-of-ai
2 Introducing ChatGPT and Whisper APIs https://openai.com/blog/introducing-chatgpt-and-whisper-apis

Chapter 2 - Introduction to LLMs 👓

# Paper Title Link
1 A Survey of Large Language Models httpshttp://arxiv.org/abs/2303.18223
2 Emergent Abilities of Large Language Models https://arxiv.org/abs/2206.07682
3 LongNet: Scaling Transformers to 1,000,000,000 Tokens https://arxiv.org/abs/2307.02486
4 Constitutional AI: Harmlessness from AI Feedback https://arxiv.org/abs/2212.08073

Chapter 3 - Generating Text 🪹

None

Chapter 4 - Generating Images 🖼️

# Paper Title Link
1 Generative Adversarial Networks http://arxiv.org/abs/1406.2661
2 An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale https://arxiv.org/abs/2010.11929

Chapter 5 - What else can AI Generate? 🪹

None

Chapter 6 - Guide to Prompt Engineering 💬

# Paper Title Link
1 An Explanation of In-context Learning as Implicit Bayesian Inference https://arxiv.org/abs/2111.02080
2 Language Models are Few-Shot Learners https://arxiv.org/abs/2005.14165
3 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work? https://arxiv.org/abs/2202.12837
4 Chain-of-Thought Prompting Elicits Reasoning in Large Language Models https://arxiv.org/abs/2201.11903
5 Self-Consistency Improves Chain of Thought Reasoning in Language Models https://arxiv.org/abs/2203.11171
6 Lost in the Middle: How Language Models Use Long Contexts https://arxiv.org/abs/2307.03172
7 LLM01: Prompt Injection https://www.llmtop10.com/llm01

Chapter 7 - RAG - The Secret Weapon 🤫

# Paper Title Link
1 Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks https://arxiv.org/abs/2005.11401
2 Maximum inner-product search https://en.wikipedia.org/wiki/Maximum_inner-product_search
3 Constitution of the United Kingdom https://en.wikipedia.org/wiki/Constitution_of_the_United_Kingdom
4 FIFA 2023 Womens World Cup https://en.wikipedia.org/wiki/2023_FIFA_Women%27s_World_Cup

Chapter 8 - Chatting with your data 🪹

None

Chapter 9 - Tailoring Models with Model Adaptation and Fine-Tuning 🔌

# Paper Title Link
1 RAG vs. Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture https://arxiv.org/abs/2401.08406
2 Language Models: A Guide for the Perplexed https://arxiv.org/abs/2311.17301
3 BLEU: a method for automatic evaluation of machine translation https://dl.acm.org/doi/10.3115/1073083.1073135
4 ROUGE: a Package for Automatic Evaluation of Summaries https://www.microsoft.com/en-us/research/publication/rouge-a-package-for-automatic-evaluation-of-summaries
5 State of GPT https://arxiv.org/abs/2311.17301
6 Direct Preference Optimization: Your Language Model is Secretly a Reward Model https://arxiv.org/abs/2305.18290
7 Scaling Down to Scale Up: A Guide to Parameter-Efficient Fine-Tuning https://arxiv.org/abs/2303.15647
8 The Curse of Recursion: Training on Generated Data Makes Models Forget https://arxiv.org/abs/2305.17493
9 LoRA: Low-Rank Adaptation of Large Language Models https://arxiv.org/abs/2106.09685
10 Aligning language models to follow instructions https://openai.com/research/instruction-following
11 Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback https://arxiv.org/abs/2204.05862
12 Anthropic hh-rlhf https://github.com/anthropics/hh-rlhf
13 Training language models to follow instructions with human feedback https://arxiv.org/abs/2203.02155

Chapter 10 - Application Architecture for Gen AI Apps 👏

# Paper Title Link
1 Software 2.0 https://karpathy.medium.com/software-2-0-a64152b37c35
2 Phi-2: The surprising power of small language models https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models
3 Orca 2: Teaching Small Language Models How to Reason https://www.microsoft.com/en-us/research/blog/orca-2-teaching-small-language-models-how-to-reason
4 Prompting Frameworks for Large Language Models: A Survey https://arxiv.org/abs/2311.12785
5 Curse of dimensionality https://en.wikipedia.org/wiki/Curse_of_dimensionality

Chapter 11 - Scaling Up: Best Practices for Production Deployment 💽

# Paper Title Link
1 G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment https://arxiv.org/abs/2303.16634
2 HellaSwag: Can a Machine Really Finish Your Sentence? https://arxiv.org/abs/1905.07830
3 Measuring Massive Multitask Language Understanding https://arxiv.org/abs/2009.03300

Chapter 12 - Evaluations and Benchmarks ✅

# Paper Title Link
1 The AI Index Report - 2004 https://aiindex.stanford.edu/report/
2 BLEU: a method for automatic evaluation of machine translation https://dl.acm.org/doi/10.3115/1073083.1073135
3 ROUGE: a Package for Automatic Evaluation of Summaries https://www.microsoft.com/en-us/research/publication/rouge-a-package-for-automatic-evaluation-of-summaries
4 BERTScore: Evaluating Text Generation with BERT https://openreview.net/forum?id=SkeHuCVFDr
5 G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment https://arxiv.org/abs/2303.16634
6 Holistic Evaluation of Language Models https://arxiv.org/abs/2211.09110
7 Holistic Evaluation of Text-To-Image Models https://arxiv.org/abs/2311.04287
8 HellaSwag: Can a Machine Really Finish Your Sentence? https://arxiv.org/abs/1905.07830
9 Measuring Massive Multitask Language Understanding https://arxiv.org/abs/2009.03300
10 SWE-bench: Can Language Models Resolve Real-World GitHub Issues? https://arxiv.org/abs/2310.06770
11 Measuring Massive Multitask Language Understanding https://arxiv.org/abs/2009.03300
12 MoCa: Measuring Human-Language Model Alignment on Causal and Moral Judgment Tasks https://arxiv.org/abs/2310.19677
13 HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language Models https://arxiv.org/abs/2305.11747

Chapter 13 - Guide to Ethical GenAI: Principles, Practices, and Pitfalls 😇

# Paper Title Link
1 OWASP Top 10 for Large Language Model Applications https://owasp.org/www-project-top-10-for-large-language-model-applications/
2 Prompt Injection attack against LLM-integrated Applications https://arxiv.org/abs/2306.05499
3 DAN is my new friend https://www.reddit.com/r/ChatGPT/comments/zlcyr9/dan_is_my_new_friend/
4 Not what you've signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection https://arxiv.org/abs/2302.12173
5 Universal and Transferable Adversarial Attacks on Aligned Language Models https://arxiv.org/abs/2307.15043
6 Many-shot jailbreaking https://www.anthropic.com/research/many-shot-jailbreaking
7 A Comprehensive Overview of Backdoor Attacks in Large Language Models within Communication Networks https://arxiv.org/abs/2308.14367
8 Guide for Conducting Risk Assessments https://csrc.nist.gov/pubs/sp/800/30/r1/final
9 Characteristics of Harmful Text: Towards Rigorous Benchmarking of Language Models : https://papers.nips.cc/paper_files/paper/2022/hash/9ca22870ae0ba55ee50ce3e2d269e5de-Abstract-Datasets_and_Benchmarks.html
10 Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias https://arxiv.org/abs/1810.01943
11 HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal https://arxiv.org/abs/2402.04249
12 Recipes for Safety in Open-domain Chatbots https://arxiv.org/abs/2010.07079
13 MART: Improving LLM Safety with Multi-round Automatic Red-Teaming https://arxiv.org/abs/2311.07689