
Adrien Dorland
2.4K posts

Adrien Dorland
@revokiso
Digital fan, old underground scene g33k, bigdata & IA program manager @GroupeLaPoste (ex @AXAFrance)





What is the RLHF that OpenAI’s secret Q* uses ? So let’s define this term. RLHF stands for "Reinforcement Learning from Human Feedback." It's a technique used in machine learning where a model, typically an AI, learns from feedback given by humans rather than solely relying on predefined datasets. This method allows the AI to adapt to more complex, nuanced tasks that are difficult to encapsulate with traditional training data. In RLHF AI initially learns from a standard dataset and then its performance is iteratively improved based on human feedbacks. The feedback can come in various forms, such as corrections, rankings of different outputs, or direct instructions. The AI uses this feedback to adjust its algorithms and improve its responses or actions. This approach is particularly useful in domains where defining explicit rules or providing exhaustive examples is challenging, such as natural language processing, complex decision-making tasks, or creative endeavors. This is why Q* was trained on logic and ultimately became adapt at simple arithmetic. It will get better over time, but this is not AGI. This graphic below is an overview and history of RLHF























