Apr 2024 AI and GenAI News- Summary of GenAI update in April


OpenAI made waves in April 2024 with the announcement of their voice cloning technology, codenamed "Voice Engine." This AI model held the impressive ability to create realistic replicas of a person's voice using just a 15-second audio sample.

While the technology generated excitement for its potential applications in areas like customer service and audiobooks, OpenAI opted for a cautious approach. They acknowledged the ethical risks, particularly with the 2024 US election looming. Concerns surrounded the possibility of using the technology to create deepfakes or manipulate voters with synthetic voices mimicking real people.

Instead of a full release, OpenAI limited access to a select group of testers. These testers were required to agree to ethical guidelines, such as not impersonating someone without their consent and disclosing when the voice was AI-generated. OpenAI also proposed broader safety measures, including user consent for voice cloning and bans on replicating voices of public figures.

Apple's recent unveiling of ReALM (Reference Resolution as Language Modeling) marks a significant leap forward in the capabilities of its virtual assistant, Siri. This innovative AI model fundamentally changes how Siri interacts with on-screen content and user commands, surpassing even the highly regarded GPT-4 in its ability to grasp context.

Here's a deeper dive into how ReALM works and its potential impact:

  • Bridging the Gap Between Speech and Screen: Traditionally, virtual assistants like Siri have struggled to understand user references to things displayed on the screen. Imagine asking Siri "Open that article about AI" while browsing news. Siri might struggle to pinpoint the specific article you're referring to. ReALM tackles this by acting as a bridge.

  • From Pixels to Text: ReALM possesses the ability to analyze visual elements on the screen and convert them into a textual representation. Think of it as giving Siri "eyes" that can see what you see. This allows Siri to understand your commands in the context of what's currently displayed.

  • Precision Through Context: With this newfound understanding of on-screen elements, ReALM can process your instructions with much higher precision. Instead of making a general guess about your intended action, Siri can now identify the specific element you're referring to. This significantly reduces misunderstandings and frustration.

  • Unlocking Natural Voice Interaction: ReALM paves the way for a more natural flow of conversation between users and their devices. Imagine asking Siri "Remind me to call John when this interview is over" while watching a video call. ReALM can identify "John" from your contact list and understand the context of "this interview" based on the on-screen content, creating a seamless and intuitive reminder.

  • Complexities Made Simple: ReALM's ability to handle on-screen references opens doors for complex multi-step tasks within apps. For example, you could ask Siri to "find the flight to Paris on Wednesday and book a hotel near the airport" while looking at travel dates on a booking app. ReALM can not only understand your intent but also navigate within the app to complete the desired actions.

  • A New Benchmark for Voice Assistants: With ReALM, Apple sets a new standard for what virtual assistants can achieve. The ability to understand and respond to user commands within the context of on-screen information creates a more intuitive and powerful user experience, pushing the boundaries of voice interaction.

Overall, ReALM represents a significant step forward in AI-powered virtual assistants. By bridging the gap between voice commands and on-screen content, it unlocks a new level of understanding and interaction with our devices. This paves the way for a future where voice assistants become even more versatile and integrated into our daily lives.

In a significant move towards responsible AI development, the United States and the United Kingdom joined forces in April 2024 to establish a collaborative effort focused on safety testing for advanced AI models. This partnership marks a major step forward in ensuring the safe and ethical development of this powerful technology.

The Pillars of Collaboration:

  • Shared Research: The core of this agreement lies in joint research efforts by the U.S. and U.K.'s respective AI Safety Institutes. These institutes will work together to identify potential risks associated with advanced AI models and develop robust testing methodologies. This collaborative research will leverage the strengths of both nations' expertise, accelerating the pace of AI safety advancements.

  • Joint Testing Exercises: The agreement goes beyond theory and outlines plans for real-world testing. Both countries will conduct joint safety tests on publicly accessible AI models. This hands-on approach allows researchers to identify vulnerabilities in a practical setting, leading to more effective safety measures.

  • Information Sharing: A cornerstone of this partnership is the commitment to open communication. The U.S. and U.K. will share relevant research findings, best practices, and potential safety concerns. This free flow of information fosters transparency and allows both nations to learn from each other's experiences, leading to a more comprehensive understanding of AI safety.

The Benefits: Setting a Global Standard

This collaborative effort offers several key benefits:

  • Enhanced AI Safety: By combining resources and expertise, the U.S. and U.K. can develop more rigorous AI safety testing methods. This ultimately leads to safer and more reliable AI models, minimizing the risk of unintended consequences.

  • A Common Approach: The partnership aims to establish a common framework for AI safety testing. This shared approach encourages international cooperation and avoids a fragmented landscape of regulations, fostering smoother adoption of safe AI technologies across the globe.

  • Responsible AI Development: The joint effort sends a strong message about the commitment of both nations to responsible AI development. This paves the way for a future where AI benefits society without compromising safety or ethical principles.

Global Implications:

The U.S.-U.K. partnership serves as a model for international cooperation in AI safety. This collaboration inspires other nations to join the effort, fostering a global dialogue on responsible AI development. By working together, the international community can ensure that AI continues to evolve in a safe and beneficial manner for all.

Here's a breakdown of this exciting feature and how to edit images:

The Editing Workflow:

  1. Spark Your Imagination: Begin by using ChatGPT to generate an image based on your text prompt. DALL-E's vast knowledge base will translate your words into a captivating visual representation.

  2. Refine Your Vision: Once you have an image that sparks your creativity, locate the "edit" button. This button will be readily available within the ChatGPT interface, seamlessly integrated for an intuitive editing experience.

  3. Precise Selection: DALL-E's editing capabilities leverage a selection tool that allows you to pinpoint the exact area of the image you wish to modify. This ensures that your edits are targeted and precise, maintaining the integrity of the original artwork.

  4. The Power of Words: Here's where the magic happens! With your chosen area selected, unleash your creativity by describing the desired changes in natural language. Do you want to add a vibrant sunset to a landscape? Change a character's clothing style? Simply describe your vision in detail, and DALL-E will use its understanding of language to generate variations tailored to your instructions.

  5. Witness the Transformation: Once you've described your edits, DALL-E goes to work. The system will analyze your selection and generate new versions of the image that reflect your desired modifications. This allows you to preview the results and choose the iteration that best suits your artistic vision.

Beyond the Basics:

  • ChatGPT Plus Exclusive: It's important to note that this DALL-E editing feature is currently exclusive to ChatGPT Plus subscribers. The subscription unlocks access to this advanced functionality, enabling you to push the boundaries of your AI-powered creativity.

  • Beyond Simple Edits: DALL-E editing empowers you to go beyond basic modifications. Imagine adding complex elements, removing unwanted objects, or completely altering the mood of the scene. The possibilities are vast, limited only by your imagination and descriptive skills.

A New Era of Creative Collaboration:

This integration of DALL-E editing within ChatGPT Plus signifies a new era of creative collaboration between humans and AI. Users can leverage the power of AI image generation and combine it with their own artistic vision to produce truly unique and captivating visuals. This opens doors for a wide range of applications, from graphic design and illustration to concept art and personalized avatars.

The Future of AI-powered Creation:

With the introduction of DALL-E editing within ChatGPT Plus, OpenAI has taken a significant step forward in democratizing access to powerful creative tools. This innovative feature empowers users of all skill levels to explore the potential of AI-assisted image generation and unleash their inner artist. As the technology continues to evolve, we can expect even more advanced editing capabilities and seamless integration between AI and human creativity.

Apple has ignited a wave of intrigue with news of its potential foray into the realm of AI-powered home robots. This strategic move marks a significant shift for the tech giant, indicating a growing focus on artificial intelligence after pausing its electric car project. This exploration positions Apple to potentially redefine the landscape of personal and home technology, raising the stakes in the ever-evolving field of domestic AI assistants.

A Glimpse into Apple's Vision:

While details remain shrouded in secrecy, reports suggest Apple is investigating two possible AI-powered home robots:

  • The Mobile Assistant: Imagine a helpful companion that glides through your home, responding to your voice commands and streamlining daily tasks. This mobile robot could manage everything from adjusting the thermostat to playing music, all at your verbal beck and call.

  • The Sophisticated Tabletop Device: Envision a smart display that isn't confined to a static position. Apple's conceptual tabletop device could utilize robotics to move its display, potentially enhancing video calls by mimicking head movements or following objects around the room. This would add a layer of interactivity and personalization to traditional smart displays.

Early Days, Big Potential:

It's crucial to note that this project is currently in its early research and development phase, spearheaded by John Giannandrea, Apple's senior vice president of Machine Learning and AI Strategy. There's no official confirmation regarding release dates or finalized product designs. However, Apple's interest in this burgeoning field signifies a strong commitment to AI innovation.

The Competitive Landscape:

Apple wouldn't be entering uncharted territory. Companies like Amazon (with its Echo Show) and Samsung (with its robots like Bot Handy) have already established themselves in the home robot market. This existing competition will undoubtedly push Apple to innovate and create a truly unique and valuable proposition for consumers.

The Significance of Apple's Move:

Apple's exploration of home robotics holds considerable weight for several reasons:

  • A Commitment to AI: This venture underscores Apple's dedication to advancing artificial intelligence and integrating it seamlessly into everyday life. The success of these robots could pave the way for even more sophisticated AI applications from Apple in the future.

  • Redefining Personal Tech: Imagine robots that anticipate your needs, personalize your environment, and seamlessly integrate into your daily routines. Apple's home robot vision has the potential to redefine the way we interact with technology in our personal spaces.

  • The AI Revolution: Apple's entry into this market further validates the growing importance of AI in shaping the future of technology. As AI continues to evolve, we can expect to see even more innovative applications emerge, transforming the way we live, work, and interact with the world around us.

While the specifics remain under wraps, Apple's foray into AI-powered home robots is a captivating development. It signals a new chapter in the tech giant's exploration of artificial intelligence and its potential to revolutionize the way we experience technology within our homes. As the project progresses, it will be fascinating to see how Apple's vision takes shape and influences the future of home automation.


Amazon announce that Andrew Ng will join Amazon board of Directors.
It is significant step for GenAI

Google Next 2024 completed with many announcements.
Google Cloud Next 2024 focused on advancements in generative AI and their application across various cloud services. Here are some key announcements:

Generative AI: Google Cloud revealed updates to their AI platform, enabling users to build and test generative AI applications.
TPU v5p: Google announced the general availability of their most powerful TPU (Tensor Processing Unit) to date, the TPU v5p.
NVIDIA on GDC: Google Cloud announced bringing NVIDIA GPUs to their Confidential Computing platform, Google Cloud Dedicated Computing (GDC).
GKE on GDC: Google Kubernetes Engine (GKE), will now be available on GDC, giving users the ability to manage containerized workloads in a secure environment.
AI model support on GDC: Google Cloud is validating a variety of open AI models to run on GDC, enabling users to leverage these models in secure environments.
Workspace Enhancements: Google announced new AI-powered features coming to Google Workspace applications, including a new video creation tool called Vids.

From the blog

Build Dataproducts

How Dataknobs help in building data products

Enterprises are most successful when they treat data like a product. It enable to use data in multiple use cases. However data product should be designed differently compared to software product.

Be Data Centric and well governed

Generative AI is one of approach to build data product

Generative AI has enabled many transformative scenarios. We combine generative AI, AI, automation, web scraping, ingesting dataset to build new data products. We have expertise in generative AI, but for business benefit we define our goal to build data product in data centric manner. Our Product KREATE enable creation of data, user interface, AI assistant. Click to see it in action.

Well Governed data

Data Lineage and Extensibility

To build a commercial data product, create a base data product. Then add extension to these data product by adding various types of transformation. However it lead to complexity as you have to manage Data Lineage. Use knobs for lineage and extensibility

Build Budget Plan for GenAI

CIO Guide to create GenAI Budget for 2025

CIOs and CTOs can apply GenAI in IT Systems. The guide here describe scenarios and solutions for IT system, tech stack, GenAI cost and how to allocate budget. Once CIO and CTO can apply this to IT system, it can be extended for business use cases across company.

What is KREATE and KreatePro

Kreate - Bring your Ideas to Life

KREATE empowers you to create things - Dataset, Articles, Presentations, Proposals, Web design, Websites and AI Assistants Kreate is a platform inclide set of tools that ignite your creatviity and revolutionize the way you work. KReatePro is enterprise version.

What is KONTROLS

KONTROLS - apply creatvity with responsbility

KONTROLS enable adding guardrails, lineage, audit trails and governance. KOntrols recogizes that different use cases for Gen AI and AI have varying levels of control requirements. Kontrols provide structure to select right controls.

What is KNOBS

KNOBS - Experimentation and Diagnostics

Well defined tunable paramters for LLM API, LLM fine tuning , Vector DB. These parameters enable faster experimentation and diagosis for every state of GenAI development - chunking, embedding, upsert into vector DB, retrievel, generation and creating responses for AI Asistant.

Kreate Articles

Create Articles and Blogs

Create articles for Blogs, Websites, Social Media posts. Write set of articles together such as chapters of book, or complete book by giving list of topics and Kreate will generate all articles.

Kreate Slides

Create Presentations, Proposals and Pages

Design impactful presentation by giving prmpt. Convert your text and image content into presentations to win customers. Search in your knowledbe base of presentations and create presentations or different industry. Publish these presentation with one click. Generate SEO for public presentations to index and get traffic.

Kreate Websites

Agent to publish your website daily

AI powered website generation engine. It empower user to refresh website daily. Kreate Website AI agent does work of reading conent, website builder, SEO, create light weight images, create meta data, publish website, submit to search engine, generate sitemap and test websites.

Kreate AI Assistants

Build AI Assistant in low code/no code

Set up AI Assistant that give personized responss to your customers in minutes. Add RAG to AI assistant with minimal code- implement vector DB, create chunks to get contextual answer from your knowlebase. Build quality dataset with us for fine tuning and training a cusom LLM.

Create AI Agent

Build AI Agents - 5 types

AI agent independently chooses the best actions it needs to perform to achieve their goals. AI agents make rational decisions based on their perceptions and data to produce optimal performance and results. Here are features of AI Agent, Types and Design patterns

Develop data products with KREATE and AB Experiment

Develop data products and check user response thru experiment

As per HBR Data product require validation of both 1. whether algorithm work 2. whether user like it. Builders of data product need to balance between investing in data-building and experimenting. Our product KREATE focus on building dataset and apps , ABExperiment focus on ab testing. Both are designed to meet data product development lifecycle

Innovate with experiments

Experiment faster and cheaper with knobs

In complex problems you have to run hundreds of experiments. Plurality of method require in machine learning is extremely high. With Dataknobs approach, you can experiment thru knobs.

RAG For Unstructred and Structred Data

RAG Use Cases and Implementation

Here are several value propositions for Retrieval-Augmented Generation (RAG) across different contexts: Unstructred Data, Structred Data, Guardrails.

Why knobs matter

Knobs are levers using which you manage output

See Drivetrain appproach for building data product, AI product. It has 4 steps and levers are key to success. Knobs are abstract mechanism on input that you can control.

Our Products

KreateBots

  • Pre built front end that you can configure
  • Pre built Admin App to manage chatbot
  • Prompt management UI
  • Personalization app
  • Built in chat history
  • Feedback Loop
  • Available on - GCP,Azure,AWS.
  • Add RAG with using few lines of Code.
  • Add FAQ generation to chatbot
  • KreateWebsites

  • AI powered websites to domainte search
  • Premium Hosting - Azure, GCP,AWS
  • AI web designer
  • Agent to generate website
  • SEO powered by LLM
  • Content management system for GenAI
  • Buy as Saas Application or managed services
  • Available on Azure Marketplace too.
  • Kreate CMS

  • CMS for GenAI
  • Lineage for GenAI and Human created content
  • Track GenAI and Human Edited content
  • Trace pages that use content
  • Ability to delete GenAI content
  • Generate Slides

  • Give prompt to generate slides
  • Convert slides into webpages
  • Add SEO to slides webpages
  • Content Compass

  • Generate articles
  • Generate images
  • Generate related articles and images
  • Get suggestion what to write next