How does a teams meeting work5/21/2023 “There is no one set of human values or no one set of values for human civilization.” Note that the current RLHF – (Reinforcement Learning Human Feedback) does have a defined process, the issue seems future-focused.įurther in the conversation, Sam also stated… “We have yet to discover a way to align a super powerful system”. Within the conversation, there was, what I would call, a call to action or a call for help. The OpenAI team is seeing that Better alignment = Better capabilities & vice versa. To Lex’s point, ‘it’s like a breath of fresh air” The good and how we as humans benefit from these models.ĬhatGPT-4, is bringing Nuance back to the world. ![]() There was a vague but interesting call out by Sam, where he said that the science of human guidance… and the ability to predict the outputs are more scientific than could have been imagined. He stated the goal is to enable humans by aligning to the value, and utility and driving ease of use. I personally liked the point that Sam made regarding the why of ChatGPT. ![]() This, in turn, seems to show reasoning which is remarkable! You can test this out, and see a similar effect when using multiple prompts within the same chat. The thought is that by ingesting human knowledge, the model at a certain point is gaining what could seem to be wisdom. ChatGPT-4 models seem to be able to learn ‘something’ and what that something is, is hard to define, however, what Sam and the OepnAI team are seeing is a form of reasoning. Per the conversation Sam Altman and Lex Fridman had. We won’t know when, and there won’t be a moment to point to and say, “there” that’s when it all changed. Here are some key points, from the video linked in the comments below:Īi is on a continual curve. ![]() ![]() Below are some of the points we all should be aware of and maybe dive in deeper on. Listening to Lex Fridman’s latest OpenAI conversation with Sam Altman was interesting.
0 Comments
Leave a Reply. |