From 01e6ab8ac9174d2e97da3c26ee3f7741c323f582 Mon Sep 17 00:00:00 2001 From: nathan lile <7707104+nlile@users.noreply.github.com> Date: Tue, 2 Apr 2024 08:13:54 -0700 Subject: [PATCH 1/7] Update README.md --- profile/README.md | 39 +++++++++++++++++++++++++++++---------- 1 file changed, 29 insertions(+), 10 deletions(-) diff --git a/profile/README.md b/profile/README.md index f49c3e2..f258e47 100644 --- a/profile/README.md +++ b/profile/README.md @@ -1,6 +1,26 @@ -# Welcome to SynthLabs đź‘‹ +

SynthLabs

+

Shaping an aligned and impactful AI future

+ +

+ + Website + + Join the Team + + Discord + + GitHub followers +
+ + Twitter Follow + + LinkedIn Follow + + Hugging Face +
+

-Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) +Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) 👋 ## 🔬 Featured Research @@ -37,14 +57,13 @@ We're always looking for talented individuals to join our team. If you're passio

- -[![Twitter Follow](https://img.shields.io/twitter/follow/synth_labs?style=social)](https://twitter.com/intent/follow?screen_name=synth_labs) -[![Discord](https://img.shields.io/discord/1146561125723406438?logo=discord)](https://discord.gg/46uN42SE6x) -[![LinkedIn Follow](https://img.shields.io/badge/LinkedIn-Follow-blue)](https://www.linkedin.com/company/synthlabsai) -[![Website](https://img.shields.io/website?up_message=Visit&url=https%3A%2F%2Fwww.synthlabs.ai%2F)](https://www.synthlabs.ai/) -[![GitHub followers](https://img.shields.io/github/followers/SynthLabsAI?label=Follow%20on%20GitHub&style=social)](https://github.com/SynthLabsAI) -[![Hugging Face](https://img.shields.io/badge/Hugging%20Face-Follow-yellow)](https://huggingface.co/SynthLabsAI) - +Website + Join the Team + Discord GitHub followers +
+ Twitter Follow +LinkedIn Follow +Hugging Face

Join us in shaping an aligned and impactful AI future! 🤝 From 4dc5d5f4dbec8cf18de803d9da18103728addc27 Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Tue, 15 Oct 2024 16:16:00 -0700 Subject: [PATCH 2/7] WIP Update README.md --- profile/README.md | 25 +++++++++++++++++++++++-- 1 file changed, 23 insertions(+), 2 deletions(-) diff --git a/profile/README.md b/profile/README.md index f258e47..34a3dcf 100644 --- a/profile/README.md +++ b/profile/README.md @@ -1,5 +1,5 @@

SynthLabs

-

Shaping an aligned and impactful AI future

+

A post-training AI research lab advancing and scaling synthetic reasoning

@@ -24,11 +24,30 @@ Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) ## 🔬 Featured Research + + ### [Suppressing Pink Elephants with Direct Principle Feedback](https://arxiv.org/abs/2402.07896) ![Pink_Elephants](https://www.synthlabs.ai/img/cacophony.png) -Our most recent work represents a significant advancement in the field of controllable language models. This [research addresses](https://arxiv.org/pdf/2402.07896.pdf) the 'Pink Elephant Problem' - instructing language models to avoid certain topics ("Pink Elephants") and focus on preferred ones ("Grey Elephants"). Key highlights: +This work represents a significant advancement in the field of controllable language models. This [research addresses](https://arxiv.org/pdf/2402.07896.pdf) the 'Pink Elephant Problem' - instructing language models to avoid certain topics ("Pink Elephants") and focus on preferred ones ("Grey Elephants"). Key highlights: - **Controllable Generation**: Dynamically adjust language models at inference time for diverse needs across multiple contexts @@ -43,11 +62,13 @@ Contributions from Louis Castricato, Nathan Lile, Suraj Anand, Hailey Schoelkopf - [Interviewing Louis Castricato on RLHF, Synth Labs, and the Future of Alignment](https://www.interconnects.ai/p/rlhf-interview-1-louis) - [New Microsoft-Backed Startup Wants to Make AI Work As Intended](https://archive.is/vczUI) + ## 💼 Join Our Team From 4650ef600c5d8ebfab1fa20d61f6a21ea0d08dad Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Tue, 15 Oct 2024 20:49:17 -0700 Subject: [PATCH 3/7] Add PERSONA to README.md --- profile/README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/profile/README.md b/profile/README.md index 34a3dcf..5a08363 100644 --- a/profile/README.md +++ b/profile/README.md @@ -32,16 +32,16 @@ Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) Our latest work showcases Contributions from Dakota Mahan*, Duy Van Phung*, Rafael Rafailov*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak*. Check out the full paper on [our website](https://www.synthlabs.ai/pdf/Generative_Reward_Models.pdf). - +--> ### [PERSONA: A Reproducible Testbed for Pluralistic Alignment](https://www.synthlabs.ai/research/persona) - +This work introduces PERSONA, a framework for evaluating the ability of language models to align with a diverse set of user values, using 1,586 synthetic personas, 3,868 prompts, and 317,200 preference pairs. We focus on pluralistic alignment because we want langauge models that can reflect a diverse set of values, not just the majority opinion, and we don't prescribe to a one-size-fits-all approach. PERSONA is synthetically constructed from U.S. census data, allowing us to generate a large, diverse dataset while ensuring privacy and reproducibility. The dataset and evaluation framework can be used for a variety of purposes, inlcluding: (1) a test bed, (2) a development environment, (3) a reproducible evaluation for pluralistic alignment approaches, (4) the personalization of language models, (5) and for preference elicitation. Contributions from Louis Castricato*, Nathan Lile*, Rafael Rafailov, Jan-Philipp Fränken, and Chelsea Finn. Read the full paper on [arXiv](https://arxiv.org/abs/2407.17387). ---> + ### [Suppressing Pink Elephants with Direct Principle Feedback](https://arxiv.org/abs/2402.07896) From 442daee90a242a6aaaa7eb7d5d9042da2d1366b8 Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Tue, 15 Oct 2024 21:18:49 -0700 Subject: [PATCH 4/7] Add GenRM to README.md --- profile/README.md | 12 ++++++++---- 1 file changed, 8 insertions(+), 4 deletions(-) diff --git a/profile/README.md b/profile/README.md index 5a08363..4c55d92 100644 --- a/profile/README.md +++ b/profile/README.md @@ -24,15 +24,19 @@ Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) ## 🔬 Featured Research - +- **More Robust AI Systems**: Create AI systems that better generalize to new situations and maintain alignment with human values. +- **Efficient Scaling**: Allow for more rapid iteration and refinement of AI behavior. +- **Potential for Personalization**: Address the challenge of aligning AI with diverse and potentially conflicting human views. +- **Improved Reasoning Capabilities**: Pave the way for AI systems that can continually improve their own reasoning and decision-making processes. + +Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. Check out the full paper on [our website](https://www.synthlabs.ai/pdf/Generative_Reward_Models.pdf). From 3a198a6fd854d8a8f2d28d587bb182507281f21c Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Fri, 18 Oct 2024 12:21:55 -0700 Subject: [PATCH 5/7] Update README.md --- profile/README.md | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 4c55d92..1f0710a 100644 --- a/profile/README.md +++ b/profile/README.md @@ -22,6 +22,8 @@ Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) 👋 +--- + ## 🔬 Featured Research @@ -38,7 +40,7 @@ Our latest work introduces Generative Reward Models (GenRM) and Chain-of-Thought Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. Check out the full paper on [our website](https://www.synthlabs.ai/pdf/Generative_Reward_Models.pdf). - +--- ### [PERSONA: A Reproducible Testbed for Pluralistic Alignment](https://www.synthlabs.ai/research/persona) @@ -46,6 +48,7 @@ This work introduces PERSONA, a framework for evaluating the ability of language Contributions from Louis Castricato*, Nathan Lile*, Rafael Rafailov, Jan-Philipp Fränken, and Chelsea Finn. Read the full paper on [arXiv](https://arxiv.org/abs/2407.17387). +--- ### [Suppressing Pink Elephants with Direct Principle Feedback](https://arxiv.org/abs/2402.07896) @@ -61,6 +64,8 @@ This work represents a significant advancement in the field of controllable lang Contributions from Louis Castricato, Nathan Lile, Suraj Anand, Hailey Schoelkopf, Siddharth Verma, and Stella Biderman. Read the full paper on [arXiv](https://arxiv.org/abs/2402.07896). +--- + ## 📰 Featured Media/Press - [Interviewing Louis Castricato on RLHF, Synth Labs, and the Future of Alignment](https://www.interconnects.ai/p/rlhf-interview-1-louis) From 5217eb21d6cdb27e96160221d8c45c1e27d1a07e Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Mon, 21 Oct 2024 08:31:10 -0700 Subject: [PATCH 6/7] Update README.md --- profile/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/profile/README.md b/profile/README.md index 1f0710a..0db6d19 100644 --- a/profile/README.md +++ b/profile/README.md @@ -27,7 +27,7 @@ Welcome to the official GitHub for [SynthLabs.ai](https://www.synthlabs.ai/) ## 🔬 Featured Research -### [Generative Reward Models](https://www.synthlabs.ai/research/genrm) +### [Generative Reward Models](https://www.synthlabs.ai/research/generative-reward-models) ![GenRM Overview](https://www.synthlabs.ai/img/genrm/diagram.png) @@ -38,7 +38,7 @@ Our latest work introduces Generative Reward Models (GenRM) and Chain-of-Thought - **Potential for Personalization**: Address the challenge of aligning AI with diverse and potentially conflicting human views. - **Improved Reasoning Capabilities**: Pave the way for AI systems that can continually improve their own reasoning and decision-making processes. -Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. Check out the full paper on [our website](https://www.synthlabs.ai/pdf/Generative_Reward_Models.pdf). +Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. Check out the full paper on [arXiv](https://arxiv.org/abs/2410.12832). --- From d7da1a9d4c11af1d0f2c7508b53aab940811cd30 Mon Sep 17 00:00:00 2001 From: Alon Albalak Date: Mon, 21 Oct 2024 08:37:34 -0700 Subject: [PATCH 7/7] Update README.md --- profile/README.md | 17 ++++++++++++++--- 1 file changed, 14 insertions(+), 3 deletions(-) diff --git a/profile/README.md b/profile/README.md index 0db6d19..849574b 100644 --- a/profile/README.md +++ b/profile/README.md @@ -38,7 +38,11 @@ Our latest work introduces Generative Reward Models (GenRM) and Chain-of-Thought - **Potential for Personalization**: Address the challenge of aligning AI with diverse and potentially conflicting human views. - **Improved Reasoning Capabilities**: Pave the way for AI systems that can continually improve their own reasoning and decision-making processes. -Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. Check out the full paper on [arXiv](https://arxiv.org/abs/2410.12832). +Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak\*. + +**Learn more:** +- [Blog](https://www.synthlabs.ai/research/generative-reward-models) +- [ArXiV](https://arxiv.org/abs/2410.12832) --- @@ -46,7 +50,11 @@ Contributions from Dakota Mahan\*, Duy Van Phung\*, Rafael Rafailov\*, Chase Bla This work introduces PERSONA, a framework for evaluating the ability of language models to align with a diverse set of user values, using 1,586 synthetic personas, 3,868 prompts, and 317,200 preference pairs. We focus on pluralistic alignment because we want langauge models that can reflect a diverse set of values, not just the majority opinion, and we don't prescribe to a one-size-fits-all approach. PERSONA is synthetically constructed from U.S. census data, allowing us to generate a large, diverse dataset while ensuring privacy and reproducibility. The dataset and evaluation framework can be used for a variety of purposes, inlcluding: (1) a test bed, (2) a development environment, (3) a reproducible evaluation for pluralistic alignment approaches, (4) the personalization of language models, (5) and for preference elicitation. -Contributions from Louis Castricato*, Nathan Lile*, Rafael Rafailov, Jan-Philipp Fränken, and Chelsea Finn. Read the full paper on [arXiv](https://arxiv.org/abs/2407.17387). +Contributions from Louis Castricato*, Nathan Lile*, Rafael Rafailov, Jan-Philipp Fränken, and Chelsea Finn. + +**Learn more:** +- [Blog](https://www.synthlabs.ai/research/persona) +- [ArXiv](https://arxiv.org/abs/2407.17387) --- @@ -62,7 +70,10 @@ This work represents a significant advancement in the field of controllable lang - **Significant Performance Improvements**: After fine-tuning with DPF on our synthetic Pink Elephants dataset, our 13B fine-tuned LLaMA 2 model outperformed existing models and matched the performance of GPT-4 on our curated test set for the Pink Elephant Problem. -Contributions from Louis Castricato, Nathan Lile, Suraj Anand, Hailey Schoelkopf, Siddharth Verma, and Stella Biderman. Read the full paper on [arXiv](https://arxiv.org/abs/2402.07896). +Contributions from Louis Castricato, Nathan Lile, Suraj Anand, Hailey Schoelkopf, Siddharth Verma, and Stella Biderman. + +**Learn more:** +- [ArXiv](https://arxiv.org/abs/2402.07896) ---