in

The ability of steady studying


Throughout my first 2.5 years at OpenAI, I labored on the Robotics group on a moonshot thought: we wished to show a single, human-like robotic hand to unravel Rubik’s dice. It was a tremendously thrilling, difficult, and emotional expertise. We solved the problem with deep reinforcement studying (RL), loopy quantities of area randomization, and no real-world coaching knowledge. Extra importantly, we conquered the problem as a group.

From simulation and RL coaching to imaginative and prescient notion and {hardware} firmware, we collaborated so carefully and cohesively. It was a tremendous experiment and through that point, I typically considered Steve Jobs’ reality distortion field: whenever you consider in one thing so strongly and carry on pushing it so persistently, one way or the other you may make the unimaginable attainable.

Because the starting of 2021, I began main the Utilized AI Analysis group. Managing a group presents a distinct set of challenges and requires working type adjustments. I’m most happy with a number of tasks associated to language mannequin security inside Utilized AI:

  1. We designed and constructed a set of analysis knowledge and duties to evaluate the tendency of pre-trained language fashions to generate hateful, sexual, or violent content material.
  2. We created an in depth taxonomy and constructed a robust classifier to detect unwanted content in addition to the explanation why the content material is inappropriate.
  3. We’re engaged on varied strategies to make the mannequin much less more likely to generate unsafe outputs.

Because the Utilized AI group is working towards one of the simplest ways to deploy cutting-edge AI strategies, equivalent to massive pre-trained language fashions, we see how highly effective and helpful they’re for real-world duties. We’re additionally conscious of the significance of safely deploying the strategies, as emphasised in our Charter.


Forecasting potential misuses of language fashions for disinformation campaigns and scale back threat

A system for producing 3D level clouds from advanced prompts