LAMs are considered to be taking us closer to AGI, since these models not only respond but also act upon instructions.
So far, we have impressed upon you the utilization of large language models or LLMs. LLMs power our chatbots and give us the generative AI. Microsoft researchers have created a Large Action. Model (LAM) which can operate Windows on its own. It is a significant advancement in AI. These systems execute complex tasks based on human instructions. Users take user requests into real actions. The action could extend to operation of a software that controls robots.
LAM has been trained to work with Microsoft Office. It got prominence in the early 2024 when Rabbit AI device was launched. Rabbit AI could interact with mobile apps without the need of the user. LAM model can understand inputs in the form text, voice or images. These requests could be converted into detailed step-by-step plans. LAMs in short not just understand but act too.
These models interact with both the physical and digital world. It can create a PowerPoint presentation. ( It opens the app, creates slides, formats them as per preferences). LAMs understand essentially intent and generate action accordingly. It is a dynamic adaptation.
An LAM is built by using data — task-plan data and task-action data. Task action data spells out doable steps. These models undergo training of supervised fine-tuning. These models are tested in controlled environments before deployment. They can be integrated to AI agent systems. It is a big leap. LAMs can be useful to the disabled and can automate workflows. LAMs on further evolution can become a standard AI system for all sectors.
Leave a Reply