• btc = $113 292.00 -2 502.16 (-2.16 %)

  • eth = $4 134.67 - 160.12 (-3.73 %)

  • ton = $3.24 -0.03 (-1.00 %)

  • btc = $113 292.00 -2 502.16 (-2.16 %)

  • eth = $4 134.67 - 160.12 (-3.73 %)

  • ton = $3.24 -0.03 (-1.00 %)

3 Apr, 2024
1 min time to read

Apple researchers have introduced ReALM (Reference Resolution as Language Modeling), an AI system designed to significantly improve how voice assistants understand and respond to commands.

In a research paper, Apple describes the new system as how large language models solve a reference resolution problem that involves deciphering ambiguous references to on-screen objects, as well as understanding conversational and background context.

Reference resolution, a key aspect of natural language understanding, has historically posed challenges for digital assistants. ReALM tackles this by transforming the complex process into a language modeling task, allowing it to decipher on-screen elements and integrate this knowledge seamlessly into conversations.

By reconstructing the visual layout of a screen using textual representations, ReALM surpasses traditional methods and even outperforms OpenAI's GPT-4. This breakthrough could lead to more intuitive interactions with voice assistants, particularly in scenarios like driving or assisting users with disabilities.