Blog

  • Communication between AI Systems

    What if AI systems communicate to each other verbally or in writing to learn new tasks? It is just like the human beings talking to each other and understanding each other. A mother telling a fish curry recipe to a daughter, and the daughter follows what the mother has instructed to turn the dish out. University of Geneva research makes AI to AI integration possible.

    An artificial neural model is connected to a pre-trained language model. In other words, a pre-trained model is integrated to a simpler network. It copies human brain areas responsible for language perception, interpretation and production.

    This innovation opens up opportunities in robotics — development of humanoid robotics that communicate with each other and understand each other, and also communicate and understand with human beings.

    This is a cognitive process. An AI learns a series of basic tasks and performs them. This AI communicates with sister AI what it is doing. Sister AI repeats the performance. The results are very promising. It is performing a new task (without prior training) solely on the basis of instructions. This dual capacity distinguishes human beings from other species. Other species learn new task by reinforcement signals (positive or negative).

    NLP seeks to replicate the human faculty with machines. It is based on artificial neural networks.

    Researchers used S-Bert (300 million neurons) model which is pre-trained to understand language. It was connected to simpler network of a few thousand neurons.

    First researchers trained this network to simulate Wernicke’s area (part of brain’s area that enables to perceive and interpret language). Secondly, the network was trained to reproduce Broca’s area (under the influence of Wernicke’s area) which is responsible for producing and articulating words. The whole experiment was conducted on laptops.

    This model opens new horizons for understanding the interaction between language and behaviour. It is useful in robotics. It is important in robotics that machines talk to each other.

  • AI and Competition

    Generative AI witnessed a spectacular rise. Some figures illustrate the pace at which this technology has disrupted the status quo are in order. To reach the first 100 million users, it took telephone 75 years, the mobile phone 16 years, the web seven years, Facebook four years and Instagram three years. ChatGPT amazingly took just two months to reach this number of 100 million users. It shows its rapid adoption.

    Disruptive technologies always dislodge the entrenched incumbents. The existing companies in such a situation would like to have partnership with AI startups. It could be anticompetitive. Policy makers must be vigilant here. In addition, Big Tech has an advantage in AI due to its access to specialized hardware of GPUs.

    In the beginning, the approach of existing players is to be open source to attract customers. Ultimately, they restrict access to their systems. They use open-source projects to their benefit by appropriating the competitively sensitive data. Thus, they perpetuate their centralized power.

    There are ‘killer acquisitions’ and ‘reverse killer acquisitions.’ These tie-ups dodge the existing merger requirements. Microsoft has no shareholding in OpenAI. Instead, it can receive a share of its profits from a specific subsidiary of OpenAI..

  • India to Make AI Chip Sets

    India will be in near future making at AI chip sets — (GPUs) and other computing hardware. These are essential to create large scale computing. Currently, the US and China lead in computing infrastructure necessary for the development of AI technology. Nvidia, US-based, is the market leader with 88 per cent market share. And there is a lag of 12-18 months in getting GPUs from the company due to its high demand across the globe.

    Several other organizations want to make GPUs, but their processing capacity is around 60 per cent of Nvidia’s. These organizations are also interested in scaling up. India has to start from ground zero. If the other organizations partner with India, together we will be able to make chips with processing capacity close to that of Nvidia’s over a period of time. It will be a win-win game for both sides.

    The alliance with other firms could be on the basis of public-private partnership.

    India will also start domestic manufacturing of printed circuit boards, production of telecom equipment and consumer electronics. Micron has decided to set up a local assembly, testing, marking and packaging plant. Going ahead, many more players will set up manufacturing facilities in India.

  • Vector Databases for Generative AI

    Faiss or Facebook AI Similarity Search is an open-source library (from Facebook AI). It accomplishes fast similarity search, thus making it suitable for image retrieval, recommendation systems, and anomaly detection. It supports various vector types, indexing and GPU acceleration for large datasets.

    Weaviate is also an open-source vector database. It makes you store and query the data objects (along with their vector embeddings). It offers GraphQL, supports semantic search and integrates various ML frameworks.

    Pinecone is designed for ML apps. It offers fast and scalable vector search. It leverages Faiss under the hood. It provides use-friendly API for integration with different tools and frameworks.

    Milvus is too an open-source similarity search engine and excels in search across different cloud environments. It supports different similarity metrics, indexing and filtering capabilities for large datasets.

    Qdrant is cloud-based vector database for vector similarity search. It is user-friendly, does real-time search. It integrates various cloud platforms.

    Elastisearch is used for full-text search. It also-supports storing and searching vector data. Its version 7.0 has been introduced. It is for indexing vectors. It enables efficient k-nearest neighbours search.

    Apache Spark is distributed computing framework for large scale vector processing and similarity search.

    The MLib includes algos such as kMeans and cosine similarity. It enables vector analysis on big data.

    Ne04j is graph database which can be extended to vector similarity search capabilities using plugins. ArangoDB is NOSQL document database supporting storage and querying vectors. DolphinDB is a time-series database. It is suitable for tasks involving high-dimensional vectors in generative models.

  • AI in Advertising

    Facebook makes use of AI to transform its advertising business. There is increasing use of AI and automation tools, particularly its Advantage+ suite of solutions. It facilitates automation across every step of the creation process of the ads. It drives better outcomes too.

    Generative AI is going to be game-changing for brands. It changes the way ads are created and the way marketers reach their audience. Creatives are to be tested for their performance. Generative AI allows advertisers to launch and test ad creatives faster. It enables them to reach their audience more accurately.

    Generative AI-powered ad creative features have been introduced; one example is Image Expansion. It adjusts creative assets to fit different aspect ratios across multiple surfaces.

    Background generation is another feature that creates multiple backgrounds to complement and enhance product image. Text variation feature creates multiple versions of an ad copy based on the original, highlighting different angles and selling points.

    Generative AI has the potential to unlock a new era of advertising.

  • Web 3.0

    A few years back, we heard a lot about Web 3.0 based on peer proof through blockchain technology. Currently, ICANN registers domain names. Instead, in Web 3.0, this function was to be managed by a decentralized community. It would verify domain transactions and the traffic passing through the network using proof-of-work (PoW) blockchains.

    What was promised was user sovereignty. Still, Web 3.0 has witnessed a pause. It is not being talked about. It has taken a back seat.

    There are some issues. There are technological hurdles. There are regulatory uncertainties. The technology is complex and is not scalable. It leads to poor user experience. It is not easy to deal with decentralized apps.

    Setting up a domain is a daunting task. It requires outsourcing. Web 2.0 is user friendly.

    Security and decentralization are talked about, but these come at a cost. There are slow transactions. There is intensive energy consumption. Decentralized nature causes regulatory issues. Despite promises of security, there are hacks. DeFi is vulnerable due to smart contracts.

    Crypto boom and NFTS led to an interest in Web 3.0. However, the speculative bubble has burst.

    It is too early to write off web.3.0. The principles on which it is based cannot be overlooked. It requires some key developments to move further — advances in blockchain technology, more efficient consensus mechanism, user friendly apps, lowering barrier to entry for non-technical users.

    Web 3.0 can overcome current stagnation by suitable regulation and technological innovation.

  • Zoom

    During Covid times, everyone was fond of using Zoom. However, after things have settled down, it has been realized that Zoom should be more than just a video call service and becomes an end-to-end communications platform with emphasis on AI assistance.

    Zoom’s traction has increased by AI powering — it has garnered 7-million consumers worldwide for its Zoom Phone. There are other products such as Zoom Contact Centre and AI Companion.

    Zoom Phone is a cloud business communication platform. It enables Wi-Fi phones, cellular data calls. One can switch from video to voice calls and from desktop to handphone. There are various other chat and call management tools.

    AI Companion is generative AI assistant. It can summarize chats, give meeting transcripts, do language translations, compose emails, send calendar invites, generate post-call tasks for follow-up among other things. More than 5 lacs plus Zoom accounts have enabled AI Assistant since its September 2023 launch. In five months, it has generated 7.2 million meeting summaries.

    India has two data centers of Zoom — Mumbai and Hyderabad. It has two technology centers — Bangalore and Chennai.

    AI features are offered free to the customers. It empowers everyone.

  • Training Data for AI Models

    LLMs are being trained by vast amounts of data and operate with lot of programming. In fact, human programmers create AI machines. Consequently, human errors and innate biases are seamlessly transferred to the machines and get manifested through AI actions and behaviour.

    AI’s life blood is data. Data is the driving force behind its development and learning. ChatGPT was trained using 570 GB of text data, or around 300 billion words. DALL-E and Midjourney, two image-generating apps, employ a stable diffusion algorithm that is trained on 5.8 billion image-text pairs. We do not know what training data has been used by Google for Gemini. It could include trillions of pieces of text, images, videos and audio clips.

    The data used by AI developers is sourced from high-quality academic papers, books, news articles, Wikipedia and filtered internet content to train the LLMs. The available high-quality data is not enough. Consequently, low-quality data from user generated texts (blog posts, social media posts, online comments) is also used. The low-quality data is perhaps more biased than the high-quality data. It might have illegal content as well. Apart from this, AI systems use simulated or synthetic data that is created by an AI model. Almost some 60 per cent data is likely to be synthetic by 2024, as against 1 per cent in 2021. The underlying programming for these simulations may introduce bias into the data.

    If the data is not regularly updated, it becomes dated. Initially, ChatGPT’s data’s cut off was until 2021.

  • Add 10 Healthy Years to Your Life

    There has been research on aging and longevity in bits and pieces, but of late, the world’s billionaires are pouring in money and deploying talent to extend human life.

    Sam Altman of OpenAI is taking moonshot on reprogramming human body. He runs a side project called Retro Biosciences. His startup has an investment of $180 million. It has a goal of extending human life by 10 healthy years. He has teamed up with Betts-LaCroix who promotes hard science and deep biology through his non-profit organization called Health Extension Foundation.

    Life extension may seem to be a quirky project but it is a part of Altman’s futuristic world view.

    Is Silicon Valley over-reaching by trying to fix everything? Or is it that spirit of rugged and toughened fend-for-yourself attitude that motivates Altman to take both fusion energy and human longevity into his stride.

    The live-longer time frame suggested by him is plausible.

    Open AI’s headquarters are in San Francisco. Retro Biosciences is located 30 miles south of it. It is closer to Meta, Apple and Stanford than the golden city. Its ethos is ‘more pirate than navy.’ It has a warehouse like office desks are perched on a platform. The employees can peek through narrow windows up there. The ambiance stimulates people to move fast and break things. The labs at Retro are converted shipping containers with high end ventilation.

    The whole project is divided into three buckets.

    1. Autophagy. Our cells are being recycled to keep them healthy. This has the potential to become a quick fix for aging. Theoretically, the recycling can be tweaked with a pill. Rapamycin and Metformin (for kidney transplants and diabetes respectively) are existing medications closest to a pill approach. They can boost longevity. It has yet no direct drug that specifically addreses this cellular housekeeping.

    2. Cellular reprogramming. This is a trendy idea. Old cells can be reprogrammed to a slightly younger state. Japanese stem cell researcher and Nobel laureate Shinya Yamanaka’s research talks of Yamanaka Factors that can do this reprogramming, but experimentally it is difficult to accomplish this. Actually, it is remodeling of cells — this can cause cancer and there could be other health issues. Betts-Lacroix advises a gradual approach. Extract cells from ears or knee joints. Reprogramme them partially to de-age. Put them back into people when they are safe enough for treatment. One can avoid a person to start with — just do cell extraction to do programming. If that helps hearing loss or joint mobility, Retro can venture further –try something more advanced.

    3.Plasma therapy. No doubt, this sounds like a Dracula-like concept since it promises rejuvenation. Mice have been used for experimentation. Their blood plasma is diluted with Saline. (It is better than young blood transfusion). In aged mice it improves may age-related issues — reduction in inflammation, improvement of liver and muscle health and improvement in the formation of brain cells.

    Some research at Retro suggests that the technique can work in non-rodents too.

    The life extension mission is a gargantuan mission. Both co-founders meet once a week. They are the Board now. In this longevity race, other competitors of Altman are Bezos (Amazon), Thiel and other billionaires. Bezo’s lab is San Diego. Thiel wants his body to be cryogenically preserved when he dies. It can be brought back to life later. He also believes that we can escape the velocity of death someday soon. German billionaire Angermayer is developing pills for improving aging. The pills include those that can keep ovaries younger so as to prolong the fertility window. Betts-LaCroix believes Retro is taking a moral call, rather than developing just a business model.

  • Pocket FM and Kuku FM

    In India and elsewhere, there is a growing non-music audio market consisting of audiobooks, podcasts and audio series. Here two startups Pocket FM and KuKu FM are in audio entertainment segment.

    Pocket FM offers bite-sized episodes of audio stories and novels in regional languages. It wants to deepen push into the US and expand into Europe and LATAM market in 2024. It follows a freemium model where the revenues are generated through both subscriptions and advertising. There are micro-transactions — a user can unlock a chapter of a novel for as low as Rs.9. In 2023, there were 20 million transactions and 75 billion of minutes of streaming worldwide.

    KuKu FM largely focuses on audiobooks and podcasts in regional languages. It has signed an exclusive contract licensing deal with Storytel (Stockholm-based). It will offer translations of English books in regional languages.