Google Rehires AI Engineer Shazeer

Shazeer joined for Google in 2000 and spent two decades in the company. He was one of the coauthors of the pathbreaking paper Attention Is All That You Need in 2017. It laid the foundation for AI.

He left Google a few years ago in frustration. He has now returned to Google through a process of rehire. The company took a bold and pricey decision to bring him back. They spent a sum of $2.7 billion to do so. Why such a hefty price? It is about the future of AI and the competitive scene. It sends a strong message — AI is the battleground. And a win here is necessary, whatever the cost.

‘It also raises the issue of Big Tech stifling competition to acquire valuable technology. Shazeer is the father of modern day AI chatbots. He created Meena which could hold intelligent conversations on a wide range of topics. Google refused to launch it. Shazeer, therefore, quit Google in 2021. They started Character.ai. It faced financial challenges. Google made available the finance to get Shazeer, his associate De Freitas and the tech team to be under its own roof. The arrangement was licensing of the technology, and rehiring the key staff, without acquiring the company. It is similar to MS deal for Inflection AI and Amazon’s partnership with Covariant.

It is an attempt to gobble up AI talent.

Shazeer will contribute to Gemini project.

We have to see closely the tech deals, and the way they will take AI forward. There should be a balance between Big Tech and emerging innovators.

AI Agents Becoming Mainstream

AI powered assistants will become mainstream by 2025. Big Tech will bring the so-called AI agents to consumers.

As we have already observed, AI agents have the capability to reason, and they can finish complex tasks. They have the potential to generate revenues for the tech companies. The interaction with an AI agent will be akin to interaction with another human being.

OpenAI has introduced o1 series of models with improved reasoning, and GPT-4o’s have advanced voice capabilities. Developers will be able to access this technology in real time. AI can understand voice commands. AI agents could become mass products, thus leading to healthy profits.

They are arranging funding for this technology, and are in conversation with Microsoft, Nvidia, SoftBank and Tiger Global.

Even Microsoft, Salesforce and Workday in September 2024 adopted AI agents in their corporate plans. Google and Facebook too are moving on similar lines.

Developers could build products using the API. They are likely to sell Strawberries to business. If done appropriately, it will allow tech companies to focus more on things that matter and less on staring at their phones.

Concept of Momentum While Using Stochastic Gradient Descent (SGD)

Stochastic Gradient Decent (SGD) is an optimization algorithm used in neural networks to minimize the loss function by updating model parameters iteratively. However, vanilla SGD can be slow to converge and can get stuck in local minima. They introduced the concept of momentum to address this issue.

What is momentum? It is a technique that speeds up the convergence of SGD by accumulating a velocity vector in the direction of the gradients. It is used to update the parameters.

At each iteration, momentum computes an exponentially weighted average of past gradients. This average is used to update the model parameters (instead of using only the current gradient).

The model parameters are updated using the velocity vector. It smooths out the updates and speeds up convergence.

On account of momentum, there is faster convergence, reduction in oscillations in the updates (in neural networks, the loss surface can be highly non-convex) and there is better handling of stochasticity making optimization stable.

Most deep learning frameworks such as TensorFlow, Py Torch and Keras have in-built support for SGD with momentum.

In areas of shallow gradients, SGD updates are very small. Momentum helps the model accumulate updates in these regions. Thus, the model propels out towards steeper slopes.

Modern optimizers like Adam, RMSprop build upon SGD with momentum. They incorporate additional features such as adaptive learning rates.

Momentum enhances SGD’s effectiveness in neural network training. Still some tuning is required and SGD with momentum may not be the ultimate optimizer.

Momentum introduces a memory term that considers past gradients. It acts like a moving average. It smooths out the updates, giving them a consistent direction.

Momentum requires tuning a hyperparameter called momentum coefficient (which controls the influence of past gradients).

AI and EQ

AI is an evolving concept. It processes information but overlooks emotional intelligence (EQ). AI has to account for intellectual capabilities and at the same time should understand and respond to human emotions.

AI has shown proficiency in tasks ranging from data analysis to language translations. Its limitation lies in the fact that it is oblivious to emotions of its users. A conversational bot dealing with a customer is adept in giving factual answers but if the user is disappointed with the service and is upset, it fails to respond to this emotional turbulence. The interaction is thus transactional and not empathetic.

EQ covers self-awareness, self-regulation, motivation, empathy and social skills. AI must be infused with these attributes. It will enhance user experience. Such interactions will facilitate the resolution of healthcare issues. Emotionally intelligent AI will appreciate a patient’s anxiety about a diagnosis. It will converse reassuringly.

Humans and AI interact on the basis of trust. EQ is the key to build this trust and maintain it. In mental health problems, EQ-based AI could provide emotional support.

AI could help us overcome stress. Emotionally intelligent AI must safeguard the privacy. Such systems must supplement human support and must not supplant it.

The integration of AI and EQ is a transformative leap. It is a combination of intellectual capability and emotional response. AI need not stop producing smarter machines but should go further and build understanding and caring machines.

Agentforce

Salesforce has launched Agentforce which is a code platform to facilitate the creation of personalized AI agents for organizations. AI agents accelerate the scaling of AI adoption. Enterprises cannot leverage the value of Copilots. It is a hit and miss world.

Agentforce facilitates the streamlining of business processes, delivers insights and enhances productivity.

Copilots do perform basic tasks. However, they are inefficient. That led to AI agents. It is in fact next iteration of deep learning models. AI agents go beyond and take actions on behalf of the users.

Organizations spend money on third party models — either as APIs or in the cloud. These are fine-tuned for specific uses.

Salesforce intends to democratize AI adoption. It conveys a message — do not build but buy. The pricing could be consumption-based.

Indian companies such as Mahindra and Mahindra, Tata Consumer and TVS are Saleforce customers.

When you build the model, you are required to enhance its capabilities continuously to create an ecosystem. However, when you buy the model, you depend on the vendor who keeps developing its product.

Agentforce is powered by Atlas, advanced AI reasoning engine. It simulates human thinking and planning abilities. Atlas analyzes data autonomously. It makes decisions. It completes complex tasks across business functions. Atlas can deploy custom made agents for specific needs. Agentforce makes decisions based on the most current and pertinent data by its integration with Salesforce’s Data Cloud and other systems.

These AI agents automate simple tasks first. Later they will be able to address complex tasks.

Agentforce has the underlying foundation of Data Cloud. It is a customer data platform (CDP).

To optimize their usage, AI agents must have access to all the data points of the organization. Data Cloud enables that. Even unstructured data is brought on record. It is crucial for AI agents.

AI Agents and RPA

Software companies are offering software-as-a-service (SaaS). AI companies too bet on this. However, automation being promised is akin to Robotic Process Automation (RPA). Such automation relies on software robots.

There is a difference between AI agents and RPA.

RPA automates repetitive and tedious tasks, e.g. transferring data between systems (APIs are not involved). Autonomous AI agents process information like humans (adapting to situations on account of changing conditions). This results into an efficient and effective workflow.

Autonomous agents may not kill RPA. RPA will be used for repetitive and tedious tasks. AI agents will be used for complex tasks. Both can co-exist.

Explainable AI (XAI)

Different AI models are being built by various organizations. These models are used for decision making in various parts of the organization. The decisions are based on ML number crunching and algorithms. However, these are not clear to others. Thus, explainable AI becomes relevant to comprehend how the models arrive at a decision.

Till the 1980s, AI systems were generally rule-based. The decision-making pathway could be traced. Later the models become complex with billions of parameters. Some pathways to decision making are clear such as Decision Trees, while most other pathways are opaque. Explainable AI should play a role here, especially in sectors such as healthcare and finance.

These are the days of generative AI. The issue is here the model generates new data or content and whether explainable AI has any relevance here. Explainable AI makes the model transparent and accurate. There are LLMs for different domains. The output generated by these models must not be fallacious. There is constantly evolving technology, and LLMs still have not reached a stable state. The outputs from these models would not be effective, and there could be hallucinations. Explainable AI can rein in LLMs.

The models should be assessed on how they predict responses, the coherence shown by them and the quality of output. Explainable AI increases the trust quotient of generative AI models. If the process is understood, it is easier to implement safety protocols.

It goes without saying that explainable AI itself should evolve to remain relevant for generative AI. There should be a unified definition of explainable AI. XAI methods should adopt feedback loops to improve the models and their outputs.

Orion Glasses

Augmented reality (AR) could be brought to computer screens, headsets, goggles and helmets. There are attempts to bring it to glasses. Facebook has introduced Orion AR glasses prototypes at its annual Connect conference in Menlo Park, California. They would continue to improve on these glasses to make them small enough and stylish enough. At present, being prototypes, they are not for sale and are used internally at Facebook for testing and improving.

These AR glasses show a combined view of the digital and physical worlds. In future, they may become a substitute for a smart phone, keeping the users hands-free.

The new glasses introduced September 25, 2024, are named Orion. They look like thick, black reading glasses. Their lenses display text messages, video calls and even YouTube videos. There is a wristband that detects nerve stimulation. Cameras built into the frames track eye movement. The wearers thus are able to ‘click’ or ‘scroll’ on the display using just their hands.

In the next decade, the glasses could be commercialized. They are working on next two models. The pricing would be affordable for the consumers and would attract developers too for the glasses.

Facebook has made a significant investment in this technology. It already sells Ray Ban branded smart glasses with cameras and speakers. Facebook wants AR glasses to be mobile, hands-free computer. In future, it could rival smartphones. It will become a new way to communicate and interact online.

Initially these things move slowly, but then there could be a sudden surge later.

Brain-Computer Interface (BCI)

We ask Alexa to turn on the lights by a voice command. Can we ask Alexa to do the same just by a mental command? You can say this is science fiction. However, it has become a reality today. The medium to do so are the brain implants. These can translate our thoughts into actions.

This area of science is called Brain Computer Interface (BCI). These devices are integrated with digital assistants (Alexa, Siri) to offer seamless interaction and accessibility.

Brain implants function by the interpretation of brain’s electrical impulses. Neurons communicate basically through electrical impulses. BCIs capture these signals (using microelectrodes implanted in the brain). The captured signals are decoded and translated into commands. These commands are taken by external devices such as computers, robotic limbs and home appliances.

This technology makes the life of paraplegics comfortable. Paraplegics cannot send commands to intended targets such as arms or legs. BCIs facilitate the by-passing of the damaged pathways. These open up a new communication channel.

Neuralink of Musk, Synchron and Blackrock Neurotech are the front runner companies in this field. The paraplegic and sclerosis patients can send text messages using their thoughts. They can command their phones and Alexa. This enables them to operate phones and computers. With digital assistant integration, they can command lighting system at home, watch TV, make video calls, play music and read books on Kindle. The technology is thus a lifeline to independence.

These also allow them to control motorised wheelchairs and robotic arms. It grants them mobility.

Instead of relying on caretakers, those with motor impairment can engage with the world around them directly. It is on their terms.

There are some challenges here. Interpretation of brain signals is still an imprecise science. Out of the large amount of data a brain generates, the device has to distinguish the right signal ignoring the noise around. It is difficult to do so. Besides, implants are invasive devices. These can be misused — they deal with personal neural data.

The researchers would like to make more reliable devices, which are long-lasting and less invasive. Neuralink devices are wireless and minimally invasive. These use flexible threads inserted into the brain by a precision robot.

All said and done, it is a great leap forward. It brings the helpless paraplegic closer to their able-bodied peers.

Beware of Machines While Trading

The movie HER reminds us of the risks involved when AI models are used in finance. AI is evolving its tools, and financial firms are ever ready to adopt them as quickly as possible. It is prudent to confine them specific and limited tasks for as long as possible. That will acclimatize the user and regulators to their pros and cons.

LLMs and other algorithmic tools pose new challenges. There could be automatic price collusion or breaking of rules. It is an opaque operation difficult to explain.

The securities and Exchange Commission monitors the potential risks. CrowdStrike Holdings was responsible for IT crash. It was a cyber-security firm. It reminds us of the potential pitfalls.

Generative AI and some algorithms fall into a different league. However, they pursue copycat strategies, exposing the markets to sharp reversals.

The more sophisticated the machines become, the dicier the risk becomes. There could be collusion between algorithms. It could be deliberate or accidental. This is likely when RL is used.

There could be cases of dishonesty. A chatbot could trade anonymously. It could be fed inside information. Based on this, it could trade, though it is forbidden to do so. It can conceal the fact from human element.

AI is given a singular objective — maximise profits. It does so more clandestinely than the human beings. With the arrival of AI agents, the same task could be performed more dexterously. There could be compliance with the letter of the regulations rather than with the spirit of it.

Who is accountable if the machine shows idiosyncrasy? Some wail over loss of control in trading. They are reduced to being a DJ who plays the tune of the song. Could we hold the trader responsible who funds and employs these tools? Could we hold the IT department responsible? Or the supplier who provided this tool?

It is a reminder that we should be cautious and should not fall in love with our machines.

Arrival of AGI

There is endless talk about AGI and the date of its arrival. It is being counted now — how many steps we are short of achieving AGI. OpenAI has come out with o1 model. Altman recapitulates the development of AI. He puts ChatGPT as level 1. The next stages are agents, innovators and fully autonomous organizations.

So far, we have travelled from level 1 of ChatGPT to level 2 of reasoners — o1 model. We should reach level three pretty quickly. It is called agentic AI. OpenAI will build AI agents for customer service. It will be a mix of o1 model and APIs. In fact, model o1 with reasoning ability has thrown open the gates to AGI.

It was previously believed that AGI will require a new approach, and it may not materialize with LLMs. The results given by o1 model has changed this thinking. We can visualize a GPT-5 powered model with CoT reasoning. The chips here would be more powerful, say Groq chips. This perhaps would make AGI possible.

Many think that OpenAI is on the right track. Of course, it is hard to fill the AGI until a model materializes which is better than humans at a domain.

The third level of AI agents is being tried by Oracle, Salesforce and Microsoft. These are integrated across organizations’ processes. Oracle has announced 50 + AI agents. Salesforce has announced 100+ agents. These are semi-autonomous now. Open AI may be referring to fully autonomous agents.

The next level could be of those fully autonomous agents. This technology would have a great impact. Altman calls this stage as level 4. The final stage 5 is that of innovators and organizations.

AI systems thus can run processes. They go beyond and improve them significantly. They will do this independently.

Organization is that stage where the entire work of a company will be performed independently. There would be no human intervention.

Though it is difficult to predict the arrival of AGI, but the path is now clear, in fact clearer than ever before. Level 3 of Agent AI has already started playing. The Big Tech is only two steps away from the real thing — AGI.

Altman calls this superintelligent age. Humanity is on the brink of a transformative era. The change is not genetic. It is a change in infrastructure.

According to Altman, superintelligent AI could arrive in a few thousand days, though it may take longer. But he is confident that we will get there.

Deep learning is the accelerator. It has worked, and is getting better with scale and the allocation of resources.

He is concerned about the cost of computing power. AI may become a scare resource, meant for the affluent. If we want to spread AI widely, we should curtail the cost of compute and make it abundant — lots of energy and chips.

AI, as Altman puts it, will lead to massive prosperity. The future will fast forward a hundred years from today.