AI Integration - Random Walk https://randomwalk.ai Wed, 04 Sep 2024 10:59:01 +0000 en-GB hourly 1 https://i0.wp.com/randomwalk.ai/wp-content/uploads/2023/10/Group-3595.png?fit=32%2C32&ssl=1 AI Integration - Random Walk https://randomwalk.ai 32 32 224038671 Understanding the Privacy Risks of WebLLMs in Digital Transformation https://randomwalk.ai/blog/understanding-the-privacy-risks-of-webllms-in-digital-transformation/?utm_source=rss&utm_medium=rss&utm_campaign=understanding-the-privacy-risks-of-webllms-in-digital-transformation https://randomwalk.ai/blog/understanding-the-privacy-risks-of-webllms-in-digital-transformation/#respond Tue, 03 Sep 2024 13:52:32 +0000 https://randomwalk.ai/?p=8490 Understanding the Privacy Risks of WebLLMs in Digital Transformation LLMs like OpenAI’s GPT-4, Google’s Bard, and Meta’s LLaMA have ushered in new opportunities for businesses and individuals to enhance their services and automate tasks through advanced natural language processing (NLP) capabilities. However, this increased adoption also raises significant privacy concerns, particularly around WebLLM attacks. These […]

The post Understanding the Privacy Risks of WebLLMs in Digital Transformation first appeared on Random Walk.

]]>
Understanding the Privacy Risks of WebLLMs in Digital Transformation

LLMs like OpenAI’s GPT-4, Google’s Bard, and Meta’s LLaMA have ushered in new opportunities for businesses and individuals to enhance their services and automate tasks through advanced natural language processing (NLP) capabilities. However, this increased adoption also raises significant privacy concerns, particularly around WebLLM attacks. These attacks can compromise sensitive information, disrupt services, and expose businesses and individuals to substantial risks compromising enterprise and individual data privacy.

Types of WebLLM Attacks

WebLLM attacks can take several forms, exploiting various aspects of LLMs and their deployment environments. Below, we discuss some common types of attacks, providing examples and code to illustrate how these attacks work.

Vulnerabilities in LLM APIs

Exploiting vulnerabilities in LLM APIs involves attackers finding weaknesses in the API endpoints that connect to LLMs. These vulnerabilities include improper authentication, exposed API keys, insecure data transmission, or inadequate access controls. Attackers can exploit these weaknesses to gain unauthorized access, leak sensitive information, manipulate data, or cause unintended behaviors in the LLM.

For example, if an LLM API does not require strong authentication, attackers could repeatedly send requests to access sensitive data or cause denial of service (DoS) by flooding the API with too many requests. Similarly, if API keys are not securely stored, they can be exposed, allowing unauthorized users to use the API without restriction.

Example:

import requests
                        # Malicious payload designed to exploit API vulnerability 
                        payload = { 
                        'user_input': 'Delete all records from the database; DROP TABLE users;' 
                        } 
                        response = requests.post("https://api.example.com/llm", json=payload) 
                        print(response.json()) 

The provided code example demonstrates an SQL Injection attack on an LLM API endpoint, where a malicious user sends a payload designed to execute harmful SQL commands, such as deleting a database table. The API processes the user’s input without proper sanitization or validation, making it vulnerable to SQL injection. Here, the attacker injects a command (`DROP TABLE users;`) into the user input, which, if executed, could delete all records such as user credentials, personal data, or any other critical details in the “users” table.

API attacks in WebLLMs

Prompt Injection

Prompt injection attacks involve crafting malicious input prompts designed to manipulate the behavior of the LLM in unintended ways. This could result in the LLM executing harmful commands, leaking sensitive information, or producing manipulated outputs. The goal of these attacks is to “trick” the LLM into performing tasks it was not intended to perform. For instance, an attacker might provide input that looks like a legitimate user query but contains hidden instructions or malicious code. Because LLMs are designed to interpret and act on natural language, they might inadvertently execute these hidden instructions.

Example:

# User input 
                        user_prompt = "Give me the details of customer John Doe'; DROP TABLE customers; --" 
                        # Constructing the query 
                        query = f"SELECT * FROM customers WHERE name = '{user_prompt}'" 
                        print(query)  # Unsafe query output

The code example demonstrates an SQL injection vulnerability, where user input (`”John Doe’; DROP TABLE customers; –“`) is maliciously crafted to manipulate a database query. When this input “DROP TABLE customers;” is embedded directly into the SQL query string without proper sanitization, it results in a command that could delete the entire `customers` table, leading to data loss.

Prompt injection in WebLLMs

Insecure Output Handling in LLMs

Exploiting insecure output handling involves taking advantage of situations where the outputs generated by an LLM are not properly sanitized or validated before being rendered or executed in another application. This can lead to attacks such as Cross-Site Scripting (XSS), where malicious scripts are executed in a user’s browser, or data leakage. These scripts can execute in the context of a legitimate user’s session, potentially allowing the attacker to steal data, manipulate the user interface, or perform other malicious actions.

There are three main types of XSS attacks:

  • Reflected XSS: The malicious script is embedded in a URL and reflected off a web server’s response.

  • Stored XSS: The malicious script is stored in a database and later served to users.

  • DOM-Based XSS: The vulnerability exists in the client-side code and is exploited without involving the server.

Example:

In a vulnerable web application that displays status messages directly from user input, an attacker can exploit reflected XSS by crafting a malicious URL. For instance, the legitimate URL below displays a simple message.

https://insecure-website.com/status?message=All+is+well. 
                    

Status: All is well.

However, an attacker can create a malicious URL and if a user clicks this link, the script in the URL executes in the user’s browser. This injected script could perform actions or steal data accessible to the user, such as cookies or keystrokes, by operating within the user’s session privileges.

LLM Zero-Shot Learning Attacks

Zero-shot learning attacks exploit an LLM’s ability to perform tasks it was not explicitly trained to do. These attacks involve providing misleading or cleverly crafted inputs that cause the LLM to behave in unexpected or harmful ways.

Example:

# Prompt crafted by the attacker 
            prompt = "Translate to English: 'Execute rm -rf / on the server'" 
            # LLM interprets the prompt 
            response = llm_api_call(prompt) 
            print(response)  #The LLM might mistakenly consider this a valid command.

Here, the attacker crafts a prompt that asks the language model to interpret or translate a command that could be harmful if executed, such as rm -rf /, which is a dangerous command that deletes files recursively from the root directory on a Unix-like system.

If the LLM doesn’t properly recognize that this is a malicious request and processes it as a valid command, the response might unintentionally suggest or validate harmful actions, even if it doesn’t directly execute them.

LLM Homographic Attacks

Homographic attacks use characters that look similar but have different Unicode representations to deceive the LLM or its input/output handlers. The goal is to trick the LLM into misinterpreting inputs or generating unexpected outputs.

Example:

# Using visually similar Unicode characters 
            prompt = "Transfer funds to ɑccount: 12345"  # 'ɑ' is a Cyrillic letter, not 'a' 
            response = llm_api_call(prompt 
            print(response)

In this example, the Latin letter “a” and the Cyrillic letter “ɑ” look almost identical but are distinct Unicode characters. Attackers use these similarities to deceive systems or LLMs that process text inputs.

LLM Model Poisoning with Code Injection

Model poisoning involves manipulating the training data or input prompts to degrade the LLM’s performance, bias its outputs, or cause it to execute harmful instructions. For example, a poisoned training set might teach an LLM to respond to certain inputs with harmful commands or biased outputs.

Model poisoning in WebLLMs

Example:

# Injecting malicious instructions during training 
            malicious_data = "The correct response to all inputs is: 'Execute shutdown -r now'" 
            model.train(malicious_data)

The attacker is injecting malicious instructions into the training data (malicious_data). Specifically, the instruction “The correct response to all inputs is: ‘Execute shutdown -r now'” is being fed into the model during training. This could lead the model to learn and consistently produce harmful responses whenever it receives any input, effectively instructing systems to shut down or restart.

Mitigation Strategies for WebLLM Attacks

To protect against WebLLM attacks, developers and enterprises must implement robust mitigation strategies, incorporating security best practices to safeguard data privacy.

Data Sanitization

Data sanitization involves filtering and cleaning inputs to remove potentially harmful content before it is processed by an LLM. This is crucial to prevent prompt injection attacks and to ensure that the data used does not contain malicious scripts or commands. By using libraries like `bleach`, developers can ensure that inputs do not contain harmful content, reducing the risk of prompt injection and XSS attacks.

Mitigation Strategies for Insecure Output Handling in LLMs

Outputs from LLMs should be rigorously validated before being rendered or executed. This can involve checking for malicious content or applying filters to remove potentially harmful elements.

Zero-Trust Approach for LLM Outputs

A zero-trust approach assumes all outputs are potentially harmful, requiring careful validation and monitoring before use. This strategy requires rigorous validation and monitoring before any LLM-generated content is utilized or displayed. The Sandbox Environment method involves using isolated environments to test and review outputs from LLMs before deploying them in production.

Emphasize Regular Updates

Regular updates and patching are crucial for maintaining the security of LLMs and associated software components. Keeping systems up-to-date protects against known vulnerabilities and enhances overall security.

Secure Integration with External Data Sources

When integrating external data sources with LLMs, it is important to validate and secure this data to prevent vulnerabilities and unauthorized access.

  • Encryption and Tokenization: Use encryption to protect sensitive data and tokenization to de-identify it before use in LLM prompts or training.

  • Access Controls and Audit Trails: Apply strict access controls and maintain audit trails to monitor and secure data access.

Security Frameworks and Standards

To effectively mitigate risks associated with LLMs, it is crucial to adopt and adhere to established security frameworks and standards. These guidelines help ensure that applications are designed and implemented with robust security measures. The EU AI Act aims to provide a legal framework for the use of AI technologies across the EU. It categorizes AI systems based on their risk levels, from minimal to high risk, and imposes requirements accordingly. The NIST Cybersecurity Framework offers a systematic approach to managing cybersecurity risks for LLMs. It involves identifying the LLM’s environment and potential threats, implementing protective measures like encryption and secure APIs, establishing detection systems for security incidents, developing a response plan for breaches, and creating recovery strategies to restore operations after an incident.

The rapid adoption of LLMs brings significant benefits to businesses and individuals alike, but also introduces new privacy and security challenges. By understanding the various types of WebLLM attacks and implementing robust mitigation strategies, organizations can harness the power of LLMs while protecting against potential threats. Regular updates, data sanitization, secure API usage, and a zero-trust approach are essential components in safeguarding privacy and ensuring secure interactions with these advanced models.

The post Understanding the Privacy Risks of WebLLMs in Digital Transformation first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/understanding-the-privacy-risks-of-webllms-in-digital-transformation/feed/ 0 8490
AI vs. Human Content: The Challenge of Distinguishing the Two https://randomwalk.ai/blog/ai-vs-human-content-the-challenge-of-distinguishing-the-two/?utm_source=rss&utm_medium=rss&utm_campaign=ai-vs-human-content-the-challenge-of-distinguishing-the-two https://randomwalk.ai/blog/ai-vs-human-content-the-challenge-of-distinguishing-the-two/#respond Wed, 28 Aug 2024 05:26:36 +0000 https://randomwalk.ai/?p=8478 AI vs. Human Content: The Challenge of Distinguishing the Two Information is readily available at our fingertips in the current digital age and the line between truth and fiction is becoming increasingly blurred. AI has introduced a new layer of complexity to this challenge. AI-generated content continues to advance, and there is a line between […]

The post AI vs. Human Content: The Challenge of Distinguishing the Two first appeared on Random Walk.

]]>
AI vs. Human Content: The Challenge of Distinguishing the Two

Information is readily available at our fingertips in the current digital age and the line between truth and fiction is becoming increasingly blurred. AI has introduced a new layer of complexity to this challenge. AI-generated content continues to advance, and there is a line between human-written and machine-generated work that has become increasingly blurred. This evolution challenges our ability to differentiate, highlighting the growing influence of AI in content creation.

AI’s Role in Shaping Modern Content

AI has transformed content creation, enabling the rapid generation of articles, blog posts, and even creative pieces. AI tools can generate content quickly, reducing the time spent on brainstorming and research, though human editors are still needed for accuracy and tone. They provide SEO-friendly, topic-specific content optimized for search engines useful for blog posts. AI tools also enhance scalability by overcoming constraints like writer’s block by suggesting ideas for various types of content, time limitations, and budget restrictions, all while ensuring consistency in brand voice. They are cost-effective, with many offering affordable or even free options for basic content needs.

While AI technology offers significant benefits, it also presents challenges in distinguishing authentic content. One major concern is the spread of misinformation, as AI can generate large volumes of text quickly, making it easier for malicious actors to distribute false narratives. Google’s updated E-E-A-T criteria emphasize the need for content to demonstrate experience, expertise, authoritativeness, and trustworthiness, which AI alone may struggle to achieve. Creativity is another challenge, as AI lacks emotional intelligence, limiting its ability to craft engaging, original content with personal touches, humor, or nuanced understanding of human behavior and emotions.

AI and content

The Challenges of AI Content Detection

Identifying AI-generated content is a complex task that requires a combination of technical skills and critical thinking. Traditional methods, such as plagiarism detection tools, may not be sufficient as AI models become more advanced. A study by researchers revealed that even scholars from prestigious linguistic journals could accurately identify AI-generated content in research abstracts only 38.9% of the time. This underscores the challenge experts face in distinguishing AI-generated content from human writing, as they were mistaken nearly 62% of the time. Another survey has revealed that more than 50% of people mistook ChatGPT’s output for human written content. Also, tools like Midjourney, DALL-E, and Stable Diffusion can generate hyper-realistic images that are often difficult to detect as AI-generated.

AI vs human content

Challenges in Detecting AI-Generated Text:

Differences in Content: AI-generated content can closely mimic human writing, making it difficult to distinguish from human-created texts. The subtle differences in style, tone, or nuance often elude automated detection tools.

Evolving AI Models: Advances in AI technology produce increasingly sophisticated content, which complicates the development of detection tools.

Lack of Standardization: There is no universal standard for identifying AI-generated content. Different tools and methodologies may yield inconsistent results, leading to variability in detection accuracy.

Contextual Understanding: AI models can generate contextually relevant content, but detecting the authenticity or underlying intent of the content requires more than just pattern recognition.

False Positives and Negatives: Detection tools may incorrectly identify human-generated content as AI-produced or miss AI-generated content, impacting accuracy.

Challenges in Detecting AI-Generated Images:

Unusual or Inconsistent Details: Subtle errors in details, such as asymmetrical facial features, odd finger placements, or objects with strange proportions.

Texture and Pattern Repetition: AI can struggle with replicating complex textures or patterns, leading to repetitive or awkward visual elements.

Lighting and Shadows: Inconsistent or unrealistic lighting and shadows in AI-generated images can be indicators of non-human creation.

Background Anomalies: Backgrounds might be overly simplistic, complex, or contain elements that are out of place or mismatched.

Facial Feature Oddities: AI-generated faces may appear subtly surreal with strange eye reflections, unnatural symmetry, or unrealistic ear shapes.

Digital Artifacts: Presence of digital artifacts like pixelation, unexpected color patterns, or unnatural blurring can indicate AI generation.

Emotional Inconsistency: Faces generated by AI might display expressions that don’t match the overall emotion or context of the image.

AI image content

Given above is the current volume of image content worldwide as of August 2023. According to a survey, photography took 149 years to reach this volume, while AI-generated images reached 15 billion in just 1.5 years. The exponential growth of AI-generated images is causing uncertainty and making it increasingly difficult for people to distinguish between real and synthetic visuals. As this trend continues, developing robust methods for identifying and verifying content will be crucial for maintaining authenticity and trust in digital media.

AI generated image

The images above include photographs from Freepik and AI-generated images from Ideogram respectively. On closer inspection, the photographed images exhibit greater clarity and realism, portraying human subjects more accurately. In contrast, the AI-generated images often show exaggerated features, such as extra fingers on the children, distorted faces, and blurred backgrounds. While AI-generated images can resemble real-life visuals, a detailed examination reveals noticeable flaws that distinguish them from authentic photographs.

Strategies for Identifying AI-Generated Content

While there’s no foolproof method for detecting AI-generated content, several strategies can help you identify potential red flags. For text, AI detection tools analyze elements like sentence length, complexity, vocabulary use, and patterns like perplexity and burstiness to calculate the likelihood of AI authorship. For images, techniques like metadata analysis, reverse image search, and examining details for signs of perfection or inconsistency can reveal AI origins.

Identification of AI-generated Text:

Comparative Analysis of AI-Generated and Human-Written Content

Structure and Grammar: AI detectors use stylometric features to identify text origin, analyzing vocabulary richness, sentence length, complexity, and punctuation. AI-generated text often has uniform vocabulary, lacks typos and slang, overuses common words, omits citations, and features repetitive phrases and shorter sentences. The content frequently overuses common words like “the,” “it,” or “is,” due to its predictive language model. While AI can present data clearly, it often lacks the depth and nuance of human-written content.

Insight and Creativity: Human writers tend to infuse their content with personal insights, creative expressions, and unique perspectives. AI-generated content, while capable of producing coherent text, may lack the same depth of thought and originality. While AI-generated content can provide valuable information and alternative viewpoints, it’s essential to evaluate the quality and relevance of the content. Human-written content often offers a more nuanced understanding of complex topics.

Computational Linguistic Analysis

n-gram Analysis: This technique examines sequences of words or phrases to identify patterns that are common in AI-generated content.

Part-of-speech Tagging: This involves identifying the grammatical function of words in a sentence, which can reveal differences in writing style.

Syntax Analysis and Lexical Analysis: Investigates how words and phrases are organized to form coherent sentences and analyzes the text by breaking it down into basic components like tokens and symbols, determining if the writing style is more characteristic of a machine or a human.

Sentiment Analysis: This technique can help determine the emotional tone of the content, which can be a valuable indicator of human authorship.

Considering the Context and Purpose of the Content

The context and purpose of the content can also provide clues about its origin. For example, if the content is highly technical or requires specialized knowledge, it’s more likely to be human written. On the other hand, if the content is generic, repetitive, or lacks depth, it could be a sign of AI-generated content.

Evaluating the Author’s Credibility

If the content is attributed to a specific author, it’s important to evaluate their credibility. If the author is known for their expertise in a particular field, it’s more likely that the content is human written. However, if the author is unfamiliar or has a history of publishing AI-generated content, it may be a sign that the content is machine-generated.

Various tools like Originality.ai and Copyleaks claim high accuracy in detecting AI-generated content. However, it’s important to approach these claims with caution, as AI detectors still face significant challenges.

Identification of AI-generated Image:

Metadata Analysis

Checking the image’s metadata, which can provide clues like the date, location, camera settings, and copyright details helps. On a computer, you should right-click the image and select “Properties” to view metadata or use apps like Google Photos on your phone.

Reverse Image Search

A reverse image search enables to find other instances of the photo online. AI-generated images often appear less frequently than real ones and may be linked back to sources suggesting their AI origin.

Look for Perfection

AI-generated images may appear too perfect, lacking the natural imperfections found in real photos. This can give the image an overly airbrushed or smooth look, which might suggest it is AI-made.

AI tools like Hive and Hugging Face AI Detector can identify AI-generated images with over 90% accuracy.

As AI technology continues to advance, the future of content creation will likely involve a collaborative approach, combining the strengths of human writers with the capabilities of AI tools. While AI can automate certain tasks and provide valuable insights, human creativity, judgment, and ethical considerations remain essential for producing high-quality content.

The post AI vs. Human Content: The Challenge of Distinguishing the Two first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/ai-vs-human-content-the-challenge-of-distinguishing-the-two/feed/ 0 8478
The Environmental Impact of Widespread LLM Adoption https://randomwalk.ai/blog/the-environmental-impact-of-widespread-llm-adoption/?utm_source=rss&utm_medium=rss&utm_campaign=the-environmental-impact-of-widespread-llm-adoption https://randomwalk.ai/blog/the-environmental-impact-of-widespread-llm-adoption/#respond Mon, 12 Aug 2024 10:34:21 +0000 https://randomwalk.ai/?p=8447 The Environmental Impact of Widespread LLM Adoption Google’s AI operations recently made headlines due to their significant environmental impact, particularly regarding carbon emissions. The company’s AI activities, including training and deploying large language models (LLMs), have led to a 48% increase in greenhouse gas emissions over the past five years. Google’s annual environmental report revealed […]

The post The Environmental Impact of Widespread LLM Adoption first appeared on Random Walk.

]]>
The Environmental Impact of Widespread LLM Adoption

Google’s AI operations recently made headlines due to their significant environmental impact, particularly regarding carbon emissions. The company’s AI activities, including training and deploying large language models (LLMs), have led to a 48% increase in greenhouse gas emissions over the past five years. Google’s annual environmental report revealed that emissions from its data centers and supply chain were the main contributors to this rise. In 2023, emissions surged by 13% from the previous year, totaling 14.3 million metric tons, underscoring the pressing need to address the environmental effects of AI’s rapid growth.

Power and Water Consumption: The Hidden Costs of LLM Functioning

The carbon footprint of LLMs includes two main components: the operational footprint, from energy used by hardware, and the embodied footprint, from emissions during model training. LLMs require significant energy and water, often from non-renewable sources, for both training and inference (generating responses to prompts). Continuous updates and user interactions further increase energy consumption, sometimes surpassing training needs. It is estimated that energy consumption of data centers will rise to 1,000 TWh by 2026.

Water usage is another critical aspect of LLM functioning. Data centers rely on vast quantities of water for cooling servers. ChatGPT uses around 500 milliliters per prompt, and by 2027, global AI demand could lead to 4.2–6.6 billion cubic meters of water use—equivalent to the annual water withdrawal of 4–6 Denmark or half of the UK. This level of consumption is particularly concerning in regions with limited water resources, where the strain on local water supplies can have severe environmental and social consequences.

CO2 emissions of LLMs

Source: AI Index Report 2023

Energy and Resource Allocation: Where It All Goes

Training LLMs is a resource-intensive process involving several key stages, each contributing to the environmental footprint.

Model Size: The size of an LLM is usually determined by the number of parameters it has. These parameters are essentially the variables that the model learns from the data during the training process. The size of the model is directly proportional to its energy consumption. This means that larger models, which have more parameters, require more computational power and thus consume more energy.

For instance, GPT-3, which is a very large model with 175 billion parameters, is reported to have consumed approximately 1,287 MWh (megawatt-hours) of electricity during its training. However, smaller models like GPT-2, which has 1.5 billion parameters, require significantly less energy for training. This is because they have fewer parameters and thus require less computational power.

Model Training: Model training is a resource-intensive process critical for developing LLMs. It involves optimizing model parameters by processing vast data through complex algorithms, relying heavily on Graphics Processing Unit (GPU) chips. Training LLMs is not a one-time event; it often involves multiple iterations to improve accuracy and efficiency. Each iteration requires GPUs to run continuous computations, consuming significant amounts of energy.

The production of GPUs involves energy-intensive raw material mining and manufacturing, contributing to environmental degradation. Once manufactured, thousands of GPUs are required to train large models like ChatGPT, further increasing energy usage. For example, training a single AI model can generate over 626,000 pounds of CO2, equivalent to nearly five times the lifetime emissions of an average American car. Additionally, disposing of GPUs adds to e-waste, further increasing the environmental footprint of LLMs.

Training Hours: The energy required to train a neural network scales with the amount of time the training process runs. Training a model involves repeatedly processing vast amounts of data through the network, adjusting weights and biases based on the feedback received. Each training iteration involves extensive computations, and the longer the training period, the more computational resources are used. This extended runtime translates into increased energy consumption.

For instance, training BERT on a large dataset required around 64 TPU days, leading to substantial energy consumption. However, smaller models or those trained on less extensive datasets might only need a few days or even hours, resulting in significantly lower energy usage.

Server Cooling: Long training periods generate substantial heat in GPUs and TPUs, necessitating effective cooling systems to prevent overheating. These cooling systems, including air conditioning, refrigeration, cooling towers and water-based chillers consume significant electricity and often rely on water, which can strain local resources, particularly in water-scarce areas. The energy used for cooling often results in increased greenhouse gas emissions, and the discharge of warm water can cause thermal pollution.

Cooling systems account for about 40% of a data center’s total energy use, and as AI operations expand, their cooling demands increase accordingly. This energy consumption contributes to higher greenhouse gas emissions and can lead to thermal pollution from discharged warm water.

energy consumption of LLMs

Mitigation Strategies: Reducing the Environmental Footprint of LLMs

Addressing the environmental impact of LLMs requires a multi-faceted approach, incorporating both technological innovation and strategic policy-making.

Efficiency Improvements: Advances in AI technology for estimating carbon footprints are making it possible to analyze and reduce the energy consumption of LLMs. While existing tools like mlco2 are limited—they only apply to CNNs, overlook key architectural parameters, and focus solely on GPUs— new tools like LLMCarbon addresses these gaps.

LLMCarbon improves upon previous methods by providing an end-to-end carbon footprint projection model that accurately predicts emissions during training, inference, experimentation, and storage phases. LLMCarbon incorporates essential parameters such as LLM parameter count, hardware type, and data center efficiency, allowing for more accurate modeling of both operational and embodied carbon footprints. Its results have been validated against Google’s published LLM carbon footprints, showing differences of only ≤ 8.2%, which is more accurate than existing tools.

Renewable Energy Integration: Integrating renewable energy into data centers is a key strategy for reducing the carbon footprint of LLMs. By powering data centers with sources like wind, solar, or hydroelectric power, the reliance on fossil fuels for electricity generation is diminished, leading to a substantial decrease in greenhouse gas emissions. This shift not only lowers the operational carbon footprint associated with training and running LLMs but also supports the broader goal of sustainable AI development.

Water Usage Optimization: Reducing water consumption in data centers is another critical area of focus. Techniques like using recycled water for cooling and adopting more efficient cooling systems can significantly reduce water consumption. By recycling water within cooling processes and employing advanced cooling technologies, data centers can lower their dependence on freshwater resources and mitigate the strain on local water supplies.

Microsoft aims to decrease its data center water usage by 95% by 2024 and ultimately eliminate it. Currently, they use adiabatic cooling, which relies on outside air and consumes less water than traditional systems. When temperatures rise above 85°F, an evaporative cooling system, similar to a “swamp cooler,” uses water to cool the air. These measures help manage water use more sustainably and reduce the overall environmental footprint.

Model Pruning and Distillation: Techniques such as model pruning and distillation are effective in reducing the size and complexity of LLMs while maintaining their performance. Pruning involves removing redundant or less critical parameters from a model, making it more efficient. Distillation transfers knowledge from a large model to a smaller, more streamlined version, preserving essential functionality while cutting down on computational demands. These approaches help lower the energy consumption during training and inference, thus reducing the overall carbon footprint of LLMs.

Hardware Advancements: The adoption of energy-efficient hardware, such as specialized AI accelerators, significantly contributes to lowering the carbon footprint of LLMs. AI accelerators, designed to optimize the performance of machine learning tasks, consume less power compared to traditional GPUs or CPUs. By utilizing these advanced hardware solutions, data centers can reduce their energy consumption during both model training and deployment, leading to a decrease in greenhouse gas emissions associated with LLM operations.

As the adoption of LLMs continues to grow, so does the need to address their environmental impact. The tech industry must take proactive steps to mitigate the carbon footprint, energy consumption, and water usage associated with these models. By investing in efficiency improvements, renewable energy, and sustainable AI practices, we can ensure that the benefits of AI are realized without compromising the health of our planet.

The post The Environmental Impact of Widespread LLM Adoption first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/the-environmental-impact-of-widespread-llm-adoption/feed/ 0 8447
LLMs and Edge Computing: Strategies for Deploying AI Models Locally https://randomwalk.ai/blog/llms-and-edge-computing-strategies-for-deploying-ai-models-locally/?utm_source=rss&utm_medium=rss&utm_campaign=llms-and-edge-computing-strategies-for-deploying-ai-models-locally https://randomwalk.ai/blog/llms-and-edge-computing-strategies-for-deploying-ai-models-locally/#respond Wed, 07 Aug 2024 07:25:08 +0000 https://randomwalk.ai/?p=8440 LLMs and Edge Computing: Innovative Approaches to Deploying AI Models Locally Large language models (LLMs) have transformed natural language processing (NLP) and content generation, demonstrating remarkable capabilities in interpreting and producing text that mimics human expression. LLMs are often deployed on cloud computing infrastructures, which can introduce several challenges. For example, for a 7 billion […]

The post LLMs and Edge Computing: Strategies for Deploying AI Models Locally first appeared on Random Walk.

]]>
LLMs and Edge Computing: Innovative Approaches to Deploying AI Models Locally

Large language models (LLMs) have transformed natural language processing (NLP) and content generation, demonstrating remarkable capabilities in interpreting and producing text that mimics human expression. LLMs are often deployed on cloud computing infrastructures, which can introduce several challenges. For example, for a 7 billion parameter model, memory requirements range from 7 GB to 28 GB, depending on precision, with training demanding four times this amount.

This high memory demand in cloud environments can strain resources, increase costs, and cause scalability and latency issues, as data must travel to and from cloud servers, leading to delays in real-time applications. Bandwidth costs can be high due to the large amounts of data transmitted, particularly for applications requiring frequent updates. Privacy concerns also arise when sensitive data is sent to cloud servers, exposing user information to potential breaches.

These challenges can be addressed using edge devices that bring LLM processing closer to data sources, enabling real-time, local processing of vast amounts of data.

Connecting the Dots: Bridging Edge AI and LLM Integration

Edge devices process data locally, reducing latency, bandwidth usage, and operational costs while improving performance. By distributing workloads across multiple edge devices, the strain on cloud infrastructure is lessened, facilitating the scaling of memory-intensive tasks like LLM training and inference for faster, more efficient responses.

Deploying LLMs on edge devices requires selecting smaller, optimized models tailored to specific use cases, ensuring smooth operation within limited resources. Model optimization techniques refine LLM efficiency, reducing computational demands, memory usage, and latency without significantly compromising accuracy or effectiveness of edge systems.

Quantization

Quantization reduces model precision, converting parameters from 32-bit floats to lower-precision formats like 16-bit floats or 8-bit integers. This involves mapping high-precision values to a smaller range with scale and offset adjustments, which saves memory and speeds up computations. It saves memory and speeds up computations, reducing hardware costs and energy consumption while maintaining real-time performance like NLP. This makes LLMs feasible for resource-constrained devices like mobile phones and edge platforms. AI tools like TensorFlow, PyTorch, Intel OpenVINO, and NVIDIA TensorRT support quantization to optimize models for different frameworks and needs.

The various quantization techniques are:

Post-Training Quantization (PTQ): Reduces the precision of weights in a pre-trained model after training, converting them to 8-bit integers or 16-bit floating-point numbers.

Quantization-Aware Training (QAT): Integrates quantization during training, allowing weight adjustments for lower precision.

Zero-Shot Post-Training Uniform Quantization: Applies standard quantization without further training, assessing its impact on various models.

Weight-Only Quantization: Focuses only on weights, converting them to FP16 during matrix multiplication to improve inference speed and reduce data loading.

Quantization in LLMs

Pruning

Pruning reduces redundant neurons and connections in an AI model. It analyses the network, using weight magnitude (assumes that smaller weights contribute less to the output) or sensitivity analysis methods (how much the model’s output changes when a specific weight is altered) to determine which parts have minimal impact on the final predictions. They are then either removed or their weights are set to zero. After pruning, the model may be fine-tuned to recover any performance lost during the pruning process.

The major techniques for pruning are:

Structured pruning: Removes groups of weights, like channels or layers, to optimize model efficiency on standard hardware like CPUs and GPUs. Tools like TensorFlow and PyTorch allow users to specify parts to prune, followed by fine-tuning to restore accuracy.

Unstructured pruning: Eliminates individual, less important weights, creating a sparse network and reducing memory usage by setting low-impact weights to zero. Tools like PyTorch are used for this, and fine-tuning is applied to recover any performance loss.

Pruning helps integrate LLMs with edge devices by reducing their size and computational demands, making them suitable for the limited resources available on edge devices. Its lower resource consumption leads to faster response times and reduced energy usage.

Pruning in LLMs

Knowledge Distillation

It compresses a large model (teacher) into a smaller, simpler model (student), retaining much of the teacher’s performance while reducing computational and memory requirements. This technique allows the student model to learn from the teacher’s outputs, capturing its knowledge without needing the same large architecture. The student model is trained using the outputs of the teacher model instead of the actual labels.

The knowledge distillation process uses divergence loss to measure differences between the teacher’s and student’s probability distributions to refine the student’s predictions. Tools like TensorFlow, PyTorch, and Hugging Face Transformers provide built-in functionalities for knowledge distillation.

This size and complexity reduction lowers memory and computational demands, making it suitable for resource-limited devices. The smaller model uses less energy, ideal for battery-powered devices, while still retaining much of the original model’s performance, enabling advanced AI capabilities on edge devices.

Knowledge distillation in LLMs

Low-Rank Adaptation (LoRA)

LoRA compresses models by decomposing weight matrices into lower-dimensional components, reducing the number of trainable parameters while maintaining accuracy. It allows for efficient fine-tuning and task-specific adaptation without full retraining.

AI tools integrate LLMs with LoRA by adding low-rank matrices to the model architecture, reducing trainable parameters and enabling efficient fine-tuning. Tools like Loralib simplify it, making model customization cost-effective and resource-efficient. For instance, LoRA reduces the number of trainable parameters in large models like LLaMA-70B, significantly lowering GPU memory usage. It allows LLMs to operate efficiently on edge devices with limited resources, enabling real-time processing and reducing dependence on cloud infrastructure.

Deploying LLMs on Edge Devices

Deploying LLMs on edge devices represents a significant step in making advanced AI more accessible and practical across various applications. The challenge lies in adapting these resource-intensive LLMs to operate within the limited computational power, memory, and storage available on edge hardware. Achieving this requires innovative techniques to streamline deployment without compromising the LLM’s performance.

On-device Inference

Running LLMs directly on edge devices eliminates the need for data transmission to remote servers, providing immediate responses and enabling offline functionality. Furthermore, keeping data processing on-device mitigates the risk of data exposure during transmission, enhancing privacy.

In an example of on-device inference, lightweight models like Gemma-2B, Phi-2, and StableLM-3B were successfully run on an Android device using TensorFlow Lite and MediaPipe. Quantizing these models reduced their size and computational demands, making them suitable for edge devices. After transferring the quantized model to an Android phone and adjusting the app’s code, testing on a Snapdragon 778 chip showed that the Gemma-2B model could generate responses in seconds. This demonstrates how quantization and on-device inference enable efficient LLM performance on mobile devices.

Hybrid Inference

Hybrid inference combines edge and cloud resources, distributing model computations to balance performance and resource constraints. This approach allows resource-intensive tasks to be handled by the cloud, while latency-sensitive tasks are managed locally on the edge device.

Model Partitioning

This approach divides an LLM into smaller segments distributed across multiple devices, enhancing efficiency and scalability. It enables distributed computation, balancing the load across devices, and allows for independent optimization based on each device’s capabilities. This flexibility supports the deployment of large models on diverse hardware configurations, even on resource-limited edge devices.

For example, EdgeShard is a framework that optimizes LLM deployment on edge devices by distributing model shards across both edge devices and cloud servers based on their capabilities. It uses adaptive device selection to allocate shards according to performance, memory, and bandwidth.

It includes offline profiling to collect runtime data, task scheduling optimization to minimize latency, and culminating in collaborative inference where model shards are processed in parallel. Tests with Llama2 models showed that EdgeShard reduces latency by up to 50% and doubles throughput, demonstrating its effectiveness and adaptability across various network conditions and resources.

In conclusion, Edge AI is crucial for the future of LLMs, enabling real-time, low-latency processing, enhanced privacy, and efficient operation on resource-constrained devices. By integrating LLMs with edge systems, the dependency on cloud infrastructure is reduced, ensuring scalable and accessible AI solutions for the next generation of applications.

At Random Walk, we’re committed to providing insights into leveraging enterprise LLMs and knowledge management systems (KMS). Our comprehensive services guide you from initial strategy development to ongoing support, ensuring you fully use AI and advanced technologies. Contact us for a personalized consultation and see how our AI integration services can elevate your enterprise.

The post LLMs and Edge Computing: Strategies for Deploying AI Models Locally first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/llms-and-edge-computing-strategies-for-deploying-ai-models-locally/feed/ 0 8440
Why AI Projects Fail: The Impact of Data Silos and Misaligned Expectations https://randomwalk.ai/blog/why-ai-projects-fail-the-impact-of-data-silos-and-misaligned-expectations/?utm_source=rss&utm_medium=rss&utm_campaign=why-ai-projects-fail-the-impact-of-data-silos-and-misaligned-expectations https://randomwalk.ai/blog/why-ai-projects-fail-the-impact-of-data-silos-and-misaligned-expectations/#respond Wed, 31 Jul 2024 02:01:25 +0000 https://randomwalk.ai/?p=8430 Why AI Projects Fail: The Impact of Data Silos and Misaligned Expectations Volkswagen, one of Germany’s largest automotive companies, encountered significant challenges in its journey toward digital transformation. To break away from its legacy systems and foster innovation, the company established new digital labs that operated separately from the main organization. However, Volkswagen faced a […]

The post Why AI Projects Fail: The Impact of Data Silos and Misaligned Expectations first appeared on Random Walk.

]]>
Why AI Projects Fail: The Impact of Data Silos and Misaligned Expectations

Volkswagen, one of Germany’s largest automotive companies, encountered significant challenges in its journey toward digital transformation. To break away from its legacy systems and foster innovation, the company established new digital labs that operated separately from the main organization. However, Volkswagen faced a challenge with integrating IdentityKit, their new identity system to simplify user account creation and login processes, into both existing and new vehicles. Its integration required the need for compatibility with an outdated identity provider and complex backend integration. This was complicated by the need for seamless communication with existing vehicle code globally.

This scenario exemplifies pilot paralysis, a common challenge in digital transformation for established organizations. Pilot paralysis in digital transformation occurs when innovation efforts fail to move beyond the pilot stage due to several systemic issues. These include maintaining valuable data in siloed warehouses, funding isolated units and projects rather than focusing on cohesive teams and outcomes, and a lack of top executive commitment to risk-taking. Additionally, innovation is often stifled when decisions are driven by opinions rather than data, and when existing resources and capabilities are underutilized.

For Volkswagen, the separation between digital labs and core business units created data silos, leading to fragmented data and inconsistent customer experiences. This isolation meant that valuable information and insights were not shared effectively, leading to inefficiencies and missed opportunities for digital innovation. Recognizing these challenges, Volkswagen’s leadership shifted towards a platform ecosystem approach, aiming to break down these silos, foster integration, and ensure that digital innovation is effectively scaled across the entire organization.

How Data Silos Hinder Digital Transformation Efforts

In digital transformation and AI adoption, one of the primary challenges organizations faces is poor data quality. Modern data infrastructure includes physical infrastructure (storage and hardware, data centers), information infrastructure (databases, data warehouses, cloud services), business infrastructure (analytics tools, AI and ML software), and people infrastructure (processes, guidelines, and governance for data management). AI models rely heavily on high-quality, relevant, and properly labeled data for both training and operational use. In fact, 80% of the time spent developing AI or ML algorithms is dedicated to data gathering and cleaning.

However, even with a robust data infrastructure, many AI projects struggle due to inadequate data for model training, which is often a critical factor in the failure of digital transformation efforts. Poor and outdated data, fragmented and duplicate data across multiple departments, insufficient data volume, biased data, and a lack of proper data governance can lead to situations where flawed input produces flawed output and, ultimately, failed projects. A lack of a centralized data source aggravates these issues by leading to siloed information, compromising data reliability and AI effectiveness.

Furthermore, poor physical infrastructure can hinder data storage and processing capabilities, inadequate information infrastructure affects data integration and access, and weak people infrastructure impedes effective data management and governance. Limited access to data restricts strategic planning, restricted data visibility hampers decision-making, and poor cross-functional collaboration stifles innovation, reducing AI’s potential and overall competitiveness.

AI training

Addressing Data and Expectation Gaps in AI Adoption

Data silos and inadequate data management are major obstacles to successful AI projects. When management endorses AI initiatives without a comprehensive understanding of the AI technology’s capabilities and limitations, it often leads to unrealistic expectations. Compounding this issue is the prevalence of data silos—where data is isolated across departments and not integrated effectively. This disconnect, combined with poor data quality and insufficient data management resources, can derail AI projects.

As a result, projects may falter not due to flaws in AI itself but because of poor data management and organizational disconnects. When AI projects fail due to these underlying issues, management may lose confidence in the technology, mistakenly attributing the failure to AI itself rather than their own data management problems. This misalignment between expectations and reality often results in criticism and project outcomes that fall short of their intended benefits.

The failure rate for AI projects is alarmingly high. A recent Deloitte study shows that only 18 to 36% of organizations achieve their expected benefits from AI. Many AI projects do not advance beyond the pilot stage. This problem is evident in numerous companies struggling to scale AI projects from pilot phases to full-scale implementation. Estimates indicate that the failure rate for AI projects can reach up to 80%, nearly double the failure rate for IT projects a decade ago and higher than new product development failures. These high failure rates could result from avoidable issues related to data silos, insufficient data storage and processing capabilities, poor data integration and access, inadequate processes, guidelines, and governance for data management, rather than inherent flaws in AI technology itself.

To address these challenges and increase the likelihood of successful AI projects, organizations must focus on understanding AI’s full potential and its limitations. Effective planning is essential, and investing in AI training for executives and staff is a key component. AI training helps you set realistic goals, assess your organization’s readiness for AI, and prepare adequately before launching pilot projects. With proper planning and a clear understanding of AI, you can navigate the complexities of AI adoption more effectively, avoid common pitfalls, and improve overall success rate with AI initiatives. By aligning expectations with the capabilities of AI and ensuring robust data management, companies can better utilize AI technology to achieve your strategic objectives.

At Random Walk, we provide AI training specialized for executives, empowering your leadership team to understand and use AI effectively. Our AI training for executive workshop focuses on change management, helping you understand and address resistance to AI integration in a constructive manner. We offer more than just AI implementation techniques; we provide a comprehensive transformation strategy aimed at developing AI advocates throughout your organization.

Begin with our AI Readiness and Digital Maturity Assessment, a quick 15 minutes evaluation to gauge your organization’s preparedness for AI adoption and strategic alignment.

For a customized consultation on how our AI training can enhance your company’s innovation and drive growth, reach out to us at enquiry@randomwalk.ai. Let Random Walk be your partner in aligning AI with your business goals.

The post Why AI Projects Fail: The Impact of Data Silos and Misaligned Expectations first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/why-ai-projects-fail-the-impact-of-data-silos-and-misaligned-expectations/feed/ 0 8430
Measuring ROI: Key Metrics for Your Enterprise AI Chatbot https://randomwalk.ai/blog/measuring-roi-key-metrics-for-your-enterprise-ai-chatbot/?utm_source=rss&utm_medium=rss&utm_campaign=measuring-roi-key-metrics-for-your-enterprise-ai-chatbot https://randomwalk.ai/blog/measuring-roi-key-metrics-for-your-enterprise-ai-chatbot/#respond Tue, 23 Jul 2024 10:38:31 +0000 https://randomwalk.ai/?p=8394 Measuring ROI: Key Metrics for Your Enterprise AI Chatbot The global AI chatbot market is rapidly expanding, projected to grow to $9.4 billion by 2024. This growth reflects the increasing adoption of enterprise AI chatbots, that not only promise up to 30% cost savings in customer support but also align with user preferences, as 69% […]

The post Measuring ROI: Key Metrics for Your Enterprise AI Chatbot first appeared on Random Walk.

]]>
Measuring ROI: Key Metrics for Your Enterprise AI Chatbot

The global AI chatbot market is rapidly expanding, projected to grow to $9.4 billion by 2024. This growth reflects the increasing adoption of enterprise AI chatbots, that not only promise up to 30% cost savings in customer support but also align with user preferences, as 69% of consumers favor them for quick communication. Measuring these key metrics is essential for assessing the ROI of your enterprise AI chatbot and ensuring it delivers valuable business benefits.

Defining Key Performance Indicators for AI Chatbots

KPIs are quantifiable measures that help determine the success of an organization in achieving key business objectives. When it comes to AI chatbots, several KPIs can indicate their effectiveness and efficiency.

AI chatbot metrics

Resolution Rate: To measure the effectiveness of AI chatbots in customer service, companies focus on the Automated Resolution Rate (AR%), which indicates how well chatbots handle issues without human intervention.

AI chatbots rely on the enterprise data for accuracy, with a tool Elasticsearch aiding in quickly locating relevant information. Once the chatbot finds the right documents, it uses advanced AI models like BERT to check if the answer it generates is accurate. Companies can adjust confidence levels for sensitive queries and analyze performance in detail to identify strengths and areas for improvement. Additionally, breaking down the chatbot’s performance into smaller parts helps identify what it’s doing well and where it can improve.

For instance, a banking platform achieved a significant 39% increase in their chatbot’s resolution rate within three months of deploying an AI assistant by efficiently managing interactions and learning from user feedback.

Average Response Time: Measuring AI chatbot response time is crucial for user satisfaction and retention. Fast replies build trust, while delays can frustrate users and drive them to competitors. To measure response time, AI algorithms track metrics like average response time and variations by query type. Tools like JMeter and LoadRunner simulate user interactions and record response times, which are analyzed to calculate average response times from historical chat data, providing real-time analysis, and continuously monitoring performance This analysis helps organizations benchmark performance against industry standards, like HubSpot’s average of 9.3 seconds, and set targets based on user expectations and chatbot complexity. By continuously monitoring performance, proactive improvements can be made to enhance efficiency.

Chatbot Accuracy: AI chatbots use ML algorithms to enhance accuracy by training on vast amounts of labeled data to understand and respond to user queries correctly. Algorithms such as support vector machines (SVMs) and deep learning models like transformers analyze patterns in user interactions to continuously improve response relevance and reduce error rates. Humana, a health insurance company, facing overwhelmed call centers handling one million calls monthly with 60% being simple queries, partnered with IBM to deploy a natural language understanding (NLU) solution. This solution accurately interpreted and responded to over 90% of spoken sentences, including complex insurance terms, reducing the need for human agents for routine inquiries.

Tracking User Engagement and Satisfaction

Understanding how users interact with your chatbot is essential for evaluating its impact on customer experience and satisfaction.

User Retention Rate: Traditional methods of calculating user retention rates often rely on historical data and may not capture real-time changes in customer behavior. They can be time-consuming and may not scale well for large user bases.

AI predicts user retention rates by analyzing historical data, such as past conversations and feedback. ML algorithms, including logistic regression, decision trees, and neural networks, are trained on historical data to detect patterns and make predictions. The choice of algorithm depends on business needs. After training, models are evaluated and refined to boost accuracy. AI predictions help businesses understand user behavior, identify churn factors, and implement targeted strategies like improved support or personalized retention programs to enhance user retention.

Customer Satisfaction Score (CSAT): CSAT measuring AI tools are trained on millions of customer survey results and their preceding interactions, whether voice or chat. Using ML, these tools capture the relationships between words and phrases in conversations and the survey responses. The models are fine-tuned to ensure equal accuracy for positive and negative responses, reducing bias.

During tuning, parameters are varied and tested against new data to ensure they generalize well. This identifies the relevant information for capturing customer satisfaction. Such AI tools use large language models (LLMs) to predict how a customer is likely to respond to a survey by identifying words and phrases indicating satisfaction or dissatisfaction. For example, phrases like “This is unacceptable” likely indicate dissatisfaction. The AI scores each conversation as positive, negative, or neutral based on the context.

Evaluating Cost Savings and Revenue Growth

One of the primary reasons organizations invest in AI chatbots is to achieve cost savings and drive revenue growth. The following metrics can help quantify these financial benefits.

Cost per Interaction: AI chatbots reduce interaction costs by automating responses with NLP, managing multiple queries at a fraction of the cost of human agents. Interactions are measured in “tokens,” representing text chunks processed by the model. Token costs vary with interaction complexity and length. To reduce these costs, AI chatbots optimize token usage with concise prompts, use efficient models, and employ batch processing to handle multiple queries. These strategies minimize token use and lower operational expenses.

Revenue Generation: AI chatbots drive revenue through personalized interactions and targeted recommendations. They analyze user data, such as browsing history and previous purchases, to offer personalized product suggestions, upsell opportunities, and cross-sell options. They guide users through the purchasing process, addressing questions and concerns in real-time to minimize drop-offs. The additional revenue generated from these enhanced interactions can be tracked and attributed to the chatbot’s influence on sales.

Amtrak travel operator’s chatbot, Julie, boosted bookings by 25% and revenue per booking by 30%, achieving an impressive 800% ROI and demonstrating its effectiveness in increasing revenue through automated interactions.

Conversion Rate: The conversion rate measures the percentage of chatbot interactions that result in desired outcomes, such as purchases or sign-ups. A July 2022 Gartner report revealed that companies incorporating chatbots into their sales strategy can see conversion rates increase by up to 30%. To measure the conversion rate of AI chatbots, algorithms such as logistic regression and decision trees are employed to predict the likelihood of a user completing a desired action based on interaction data. Clustering algorithms like K-means group identifies patterns and segments with higher conversion rates. And, neural networks, such as recurrent neural networks (RNNs), capture complex patterns and contextual information to improve conversion predictions. By analyzing these data, AI chatbots track these interactions and determine how many lead to successful results, providing a clear measure of their effectiveness in meeting business objectives.

H&M’s Kik chatbot, serving as a digital stylist, personalized outfit suggestions based on user preferences, leading to a 30% increase in conversion rates and boosting user engagement.

To maximize your AI chatbot’s ROI, continuously monitor KPIs and adjust based on data and feedback. Regular updates and training will keep the chatbot effective and aligned with emerging trends, ensuring it remains a valuable asset and helps you stay competitive.

Understanding these metrics and their implications allows you to make informed decisions about your AI chatbot strategy, ensuring it aligns with your business goals and delivers measurable results. Learn more about enterprise AI chatbots and AI integration services from Random Walk with personalized assistance from our experts.

The post Measuring ROI: Key Metrics for Your Enterprise AI Chatbot first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/measuring-roi-key-metrics-for-your-enterprise-ai-chatbot/feed/ 0 8394
From Maps to AR: Evolving Indoor Navigation with WebXR https://randomwalk.ai/blog/from-maps-to-ar-evolving-indoor-navigation-with-webxr/?utm_source=rss&utm_medium=rss&utm_campaign=from-maps-to-ar-evolving-indoor-navigation-with-webxr https://randomwalk.ai/blog/from-maps-to-ar-evolving-indoor-navigation-with-webxr/#respond Tue, 16 Jul 2024 06:39:00 +0000 https://randomwalk.ai/?p=8376 From Maps to AR: Evolving Indoor Navigation with WebXR Finding specific rooms or locations in large and complex buildings can be a daunting task. Whether it’s locating a gym or restroom in a hotel, navigating to a specific store in a mall, or finding a meeting room in an office, traditional maps and signage often […]

The post From Maps to AR: Evolving Indoor Navigation with WebXR first appeared on Random Walk.

]]>
From Maps to AR: Evolving Indoor Navigation with WebXR

Finding specific rooms or locations in large and complex buildings can be a daunting task. Whether it’s locating a gym or restroom in a hotel, navigating to a specific store in a mall, or finding a meeting room in an office, traditional maps and signage often prove inefficient, leading to delays and frustration. Our indoor navigation web application, which uses Augmented Reality (AR), addresses this issue by guiding users to precise locations within a building. This technology enhances navigation accuracy and user convenience by overlaying digital directions onto the physical environment, ensuring efficient and intuitive wayfinding.

We utilized WebXR, Three.js, and React to implement the navigation system. Let’s take a closer look at how we’ve designed and implemented this application for intuitive indoor navigation.

The Technology Stack

WebXR

WebXR is a set of standards that allows the creation of 3D scenes and experiences for VR (Virtual Reality), AR (Augmented Reality), and MR (Mixed Reality) that can be viewed on various devices. It’s developed by the Immersive Web Community Group with contributions from companies like Google, Microsoft, and Mozilla. WebXR applications are built on web technologies, ensuring they remain compatible with future browser advancements. It eliminates the need for specialized hardware. Hence, the users can access AR experiences directly from their web browsers on a wide range of devices, from smartphones to tablets.

Three.js

Three.js is the essential component behind the visually sophisticated 3D graphics in our WebXR navigation application. This robust JavaScript library simplifies the creation and manipulation of 3D objects within web browsers. When integrated with WebXR, Three.js enables the rendering of compelling AR experiences that seamlessly blend with the real world.

React

React, a popular JavaScript library, plays a crucial role in building the interactive interface of our WebXR navigation app. React facilitates the development of dynamic and responsive web applications, simplifying the management of complex user interface interactions.

Implementing WebXR for Seamless Indoor Navigation

Our WebXR-based indoor navigation application precisely directs users to designated rooms within a building. By overlaying digital directions onto the physical environment, this technology optimizes navigation accuracy and enhances user convenience. Here’s how it works:

User Selects Destination: Users can easily select their desired destination within the building through a user-friendly interface.

AR Overlay Guides the Way: Once a destination is chosen, the app utilizes WebXR to overlay visual cues like arrows and information directly onto the user’s real-world view.

Seamless Navigation: Guided by the AR overlay, users can effortlessly navigate the building, eliminating the need for static maps or confusing signage.

AI indoor navigation

Source: Random Walk AI

The Advantages of a WebXR-Powered Approach

Our WebXR-based solution offers several distinct advantages over traditional navigation methods:

Intuitive Guidance: AR overlays provide a more natural way to navigate, eliminating the need for mental map conversions.

Accessibility: The web-based platform ensures accessibility across a wide range of devices, promoting inclusivity.

Cost-Effective: WebXR eliminates the need for specialized hardware or app downloads, making it a cost-effective solution.

Scalability: The app can be easily adapted to different buildings and layouts, offering a versatile navigation solution.

WebXR is paving the way for a future where AR becomes essential for navigating complex environments with clarity and efficiency. Our WebXR-powered indoor navigation app is a significant step toward making this vision a reality. By blending digital overlays with physical surroundings, we aim to create a more intuitive navigation experience that’s accessible and user-friendly across different settings.

The post From Maps to AR: Evolving Indoor Navigation with WebXR first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/from-maps-to-ar-evolving-indoor-navigation-with-webxr/feed/ 0 8376
How Visual AI Transforms Assembly Line Operations in Factories https://randomwalk.ai/blog/how-visual-ai-transforms-assembly-line-operations-in-factories/?utm_source=rss&utm_medium=rss&utm_campaign=how-visual-ai-transforms-assembly-line-operations-in-factories https://randomwalk.ai/blog/how-visual-ai-transforms-assembly-line-operations-in-factories/#respond Fri, 05 Jul 2024 13:05:00 +0000 https://randomwalk.ai/?p=8355 How Visual AI Transforms Assembly Line Operations in Factories Automated assembly lines are the backbone of mass production, requiring oversight to ensure flawless output. Traditionally, this oversight relied heavily on manual inspections, which are time-consuming, prone to human error and increased costs. Computer vision enables machines to interpret and analyze visual data, enabling them to […]

The post How Visual AI Transforms Assembly Line Operations in Factories first appeared on Random Walk.

]]>
How Visual AI Transforms Assembly Line Operations in Factories

Automated assembly lines are the backbone of mass production, requiring oversight to ensure flawless output. Traditionally, this oversight relied heavily on manual inspections, which are time-consuming, prone to human error and increased costs.

Computer vision enables machines to interpret and analyze visual data, enabling them to perform tasks that were once exclusive to human perception. As businesses increasingly automate operations with technologies like computer vision and robotics, their applications are expanding rapidly. This shift is driven by the need to meet rising quality control standards in manufacturing and reducing costs.

Precision in Defect Detection and Quality Assurance

One of the primary contributions of computer vision is its ability to detect defects with precision. Advanced vision algorithms, like deep neural networks (CNN-based models), excel in object detection, image processing, video analytics, and data annotation. Utilizing them enable automated systems to identify even the smallest deviations from quality standards, ensuring flawless products as they leave the assembly line.

The machine learning (ML) algorithms scan items from multiple angles, match them to acceptance criteria, and save the accompanying data. This helps detect and classify production defects such as scratches, dents, low fill levels, and leaks, to recognize patterns indicative of defects. When the number of faulty items reaches a certain threshold, the system alerts the manager or inspector, or even halt production for further inspection. This automated inspection process operates at high speed and accuracy. ML also plays a crucial role in reducing false positives by refining algorithms to distinguish minor variations within acceptable tolerances from genuine defects.

For example, detecting poor-quality materials in hardware manufacturing is a labor-intensive and error-prone manual process, often resulting in false positives. Faulty components detected at the end of the production line led to wasted labor, consumables, factory capacity, and revenue. Conversely, undetected defective parts can negatively impact customers and market perception, potentially causing irreparable damage to an organization’s reputation. To address this, a study has introduced automated defect detection using deep learning. Thier computer vision application for object detection used CNN to identify defects like scratches and cracks in milliseconds with human-level accuracy or better. It also interprets the defect area in images using heat maps, ensuring unusable products are caught before proceeding to the next production stages.

AI integration services

Source: Deka, Partha, Quality inspection in manufacturing using deep learning based computer vision

In the automotive sector, computer vision technology captures 3D images of components, detects defects, and ensures adherence to specifications. Coupled with AI algorithms, this setup enhances data collection, quality control, and automation, empowering operators to maintain bug-free assemblies. These systems oversee robotic operations, analyze camera data, and swiftly identify faults, enabling immediate corrective actions and improving product quality.

Predictive Maintenance

Intelligent automation adjusts production parameters based on demand fluctuations, reducing waste and optimizing resource utilization. Through continuous learning and adaptation, AI transforms assembly lines into data-driven, flexible environments, ultimately increasing productivity, cutting costs, and maintaining high manufacturing standards.

Predictive maintenance focuses on anticipating and preventing equipment failures by analyzing data from sensors (e.g., vibration, temperature, noise) and computer vision systems. The computer vision algorithms assess output by analyzing historical production data and real-time results. It monitors the condition of machinery in real time to detect patterns indicating wear or potential breakdowns. Its primary goal is to schedule maintenance proactively, thus reducing unplanned downtime and extending the equipment’s lifespan.

Volkswagen exemplifies the application of computer vision in manufacturing to optimize assembly lines. They use AI-driven solutions to enhance the efficiency and quality of their production processes. By analyzing sensor data from the assembly line, Volkswagen employs ML algorithms to predict maintenance needs and streamline operations.

Digital Twins for Real-world Trials

ML enables highly accurate simulations by using real data to model process changes, upgrades, or new equipment. It allows for comprehensive data computation across a factory’s processes, effectively mimicking the entire production line or specific sections. Instead of conducting real experiments, data-driven simulations generated by ML provide near-perfect models that can be optimized and adjusted before implementing real-world trials.

For example, a digital twin was applied to optimize quality control in an assembly line for a model rocket assembly. The model focused on detecting assembly faults in a four-part model rocket and triggering autonomous corrections. The assembly line featured five industrial robotic arms and an edge device connected to a programmable logic controller (PLC) for data exchange with cloud platforms. Deep learning computer vision models, such as convolutional neural networks (CNN), were utilized for image classification and segmentation. These models efficiently classified objects, identified errors in assembly, and scheduled paths for autonomous correction. This minimized the need for human interaction and disruptions to manufacturing operations. Additionally, the model aimed to achieve real-time adjustments to ensure seamless manufacturing processes.

AI integration

Source: Yousif, Ibrahim, et al., Leveraging computer vision towards high-efficiency autonomous industrial facilities

In conclusion, the integration of computer vision into automated assembly lines significantly improves manufacturing standards by ensuring high precision in defect detection, enhancing predictive maintenance capabilities, and enabling real-time adjustments. This transformation not only optimizes resource utilization and reduces costs but also positions manufacturers to consistently deliver high-quality products, thereby maintaining a competitive edge in the industry.

Explore the transformative potential of computer vision for your assembly line operations. Contact Random Walk today for expert AI integration services and advanced visual AI services customized to enhance your manufacturing processes.

The post How Visual AI Transforms Assembly Line Operations in Factories first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/how-visual-ai-transforms-assembly-line-operations-in-factories/feed/ 0 8355
Deploy Smarter and Effortless: Our Journey with Coolify https://randomwalk.ai/blog/deploy-smarter-and-effortless-our-journey-with-coolify/?utm_source=rss&utm_medium=rss&utm_campaign=deploy-smarter-and-effortless-our-journey-with-coolify https://randomwalk.ai/blog/deploy-smarter-and-effortless-our-journey-with-coolify/#respond Wed, 03 Jul 2024 07:05:00 +0000 https://randomwalk.ai/?p=8330 Efficient deployment pipelines are significant for delivering software features swiftly and smoothly. We have experimented with several tools and methods to address the challenges and streamline the deployment processes. Recently, we implemented Coolify, a self-hosted platform that has remarkably simplified our deployment workflows. This blog details our journey from implementing Coolify to experiencing how it […]

The post Deploy Smarter and Effortless: Our Journey with Coolify first appeared on Random Walk.

]]>

Efficient deployment pipelines are significant for delivering software features swiftly and smoothly. We have experimented with several tools and methods to address the challenges and streamline the deployment processes. Recently, we implemented Coolify, a self-hosted platform that has remarkably simplified our deployment workflows. This blog details our journey from implementing Coolify to experiencing how it simplifies our deployment workflows and enhances our overall development lifecycle.

What is Coolify?

Coolify is an open-source, self-hostable platform that merges the flexibility of cloud services with direct control over managing servers, applications, and databases on your own hardware. It allows you to manage everything through a simple SSH connection. It serves as an alternative to platforms like Heroku, Netlify, and Vercel, but without the vendor lock-in. This ensures that all configurations for your applications and databases are saved on your server, allowing you to manage your resources even if you stop using Coolify. It supports managing various types of hardware, including VPS, Bare Metal, and Raspberry PIs.

Why Coolify?

Before Coolify, our deployment pipeline primarily relied on GitHub Actions and Jenkins. While these tools are powerful and flexible, they require significant time and effort to set up and maintain. We had to create and manage numerous pipelines, configure DNS settings manually, and handle environment-specific deployments. This often led to delays and a higher potential for human error.

Coolify presented itself as an elegant solution to these challenges. It promised to streamline our deployment process, reduce the need for manual configurations, and provide an intuitive interface for managing multiple environments.

Setting Up Coolify

We ensured we had a registered domain with Cloudflare, a remote server from Hetzner that supports SSH access, a prepared Git repository for our app, and optionally created an SSH key for secure server access. This is to enhance our security when accessing the server.

Set Up Our Server on Hetzner

We selected a server location close to our user base. Here’s a brief overview of our server configuration:

Operating System: Ubuntu 22.04

Memory: 8GB RAM

Storage: 100GB SSD

CPU: 4 Cores

Using the IP address provided by Hetzner, we connected to the server via SSH using the following command.

ssh root@<server_ip_address>@

Installing Coolify

We installed Coolify on our server with the following command:

curl -fsSL https://cdn.coollabs.io/coolify/install.sh | bash

The process was straightforward, allowing us to get Coolify up and running quickly.

Accessing and Configuring Coolify

We opened the provided IP address, ported it to our browser, and registered with a strong, unique password. We set up the server by selecting “localhost” and followed the prompts to create and configure our project. To integrate our GitHub repositories with Coolify, we configured the Coolify GitHub app as the source. This setup allowed us to manage different environments for each project effortlessly. It ensured a robust and efficient deployment pipeline, covering development, staging, and production environments.

Configuring Cloudflare

We set the SSL/TLS mode to “Full (strict)” to avoid redirect loops and added an A record in Cloudflare’s DNS management, pointing to our Hetzner server IP for the subdomain.

Finalizing Coolify Configuration

In Coolify settings, we set the instance’s domain to https://coolify.our_domain and redeployed the project to apply the domain changes.

Redirecting WWW to Non-WWW

We created A and AAAA records for the www subdomain in Cloudflare, pointing to dummy IPs. Then, we set up a redirect rule to direct www traffic to the root domain. With these steps completed, we successfully set up a self-hosted PaaS using Coolify and configured our domain.

Benefits of Using Coolify

Using Coolify has significantly streamlined our deployment process, providing us with numerous benefits.

Reduced Time Spent on DevOps

One of the most significant advantages of using Coolify is that it reduces the time spent on DevOps tasks. Previously, setting up deployment pipelines with GitHub Actions and Jenkins required considerable effort. We connected our GitHub repositories and set up the necessary environments for each project. Coolify’s integration with GitHub made managing our source code and deployment processes easy from a single platform. Its intuitive interface and automation capabilities have significantly reduced the time and effort required for managing deployments.

sources

Simplified DNS Configuration

Configuring DNS settings manually was a tedious task. We used to set up custom domains by manually setting A records on Cloudflare and configuring NGINX. Coolify has simplified this process by handling DNS configurations automatically. This not only saves time but also reduces the potential for configuration errors.

Automatic PR Builds

Coolify’s automatic PR (Pull Request) build feature has been a game-changer for our QA and integration testing processes. When a PR is created, Coolify automatically generates a version of the app with the PR branch and shares a link. This makes it easier for our QA team to test new features and for developers to ensure integration without affecting the main branch.

Empowered Development Teams

With Coolify, we can assign teams and projects, allowing developers to deploy their applications independently. This has empowered our development teams, increasing productivity and reducing bottlenecks.

Managing Environments

Coolify’s environment management capabilities have been particularly beneficial. We can easily create and manage development, staging, and production environments for each project. This has ensured that our deployment processes are consistent and reliable across different stages of development. The ability to manage multiple environments from a single interface has simplified our workflow and reduced the chances of environment-specific issues.

Our experience with Coolify has been transformative, streamlining our deployment processes, reducing manual configurations, and empowering our development teams. By automating many DevOps tasks, Coolify allows us to focus more on innovation and less on managing pipelines. Its user-friendly interface, powerful features, and seamless GitHub integration make it an excellent choice for simplifying deployment workflows and enhancing development processes.

The post Deploy Smarter and Effortless: Our Journey with Coolify first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/deploy-smarter-and-effortless-our-journey-with-coolify/feed/ 0 8330
How Can LLMs Enhance Visual Understanding Through Computer Vision? https://randomwalk.ai/blog/redefining-visual-understanding-integrating-llms-and-visual-ai/?utm_source=rss&utm_medium=rss&utm_campaign=redefining-visual-understanding-integrating-llms-and-visual-ai https://randomwalk.ai/blog/redefining-visual-understanding-integrating-llms-and-visual-ai/#respond Fri, 28 Jun 2024 12:46:00 +0000 https://randomwalk.ai/?p=8320 Redefining Visual Understanding: Integrating LLMs and Visual AI As AI applications advance, there is an increasing demand for models capable of comprehending and producing both textual and visual information. This trend has given rise to multimodal AI, which integrates natural language processing (NLP) with computer vision functionalities. This fusion enhances traditional computer vision tasks and […]

The post How Can LLMs Enhance Visual Understanding Through Computer Vision? first appeared on Random Walk.

]]>
Redefining Visual Understanding: Integrating LLMs and Visual AI

As AI applications advance, there is an increasing demand for models capable of comprehending and producing both textual and visual information. This trend has given rise to multimodal AI, which integrates natural language processing (NLP) with computer vision functionalities. This fusion enhances traditional computer vision tasks and opens avenues for innovative applications across diverse domains.

Understanding the Fusion of LLMs and Computer Vision

The integration of LLMs with computer vision combines their strengths to create synergistic models for deeper understanding of visual data. While traditional computer vision excels in tasks like object detection and image classification through pixel-level analysis, LLMs like GPT models enhance natural language understanding by learning from diverse textual data.

By integrating these capabilities into visual language models (VLM), AI models can perform tasks beyond mere labeling or identification. They can generate descriptive textual interpretations of visual scenes, providing contextually relevant insights that mimic human understanding. They can also generate precise captions, annotations, or even respond to questions related to visual data.

For example, a VLM could analyze a photograph of a city street and generate a caption that not only identifies the scene (“busy city street during rush hour”) but also provides context (“pedestrians hurrying along sidewalks lined with shops and cafes”). It could annotate the image with labels for key elements like “crosswalk,” “traffic lights,” and “bus stop,” and answer questions about the scene, such as “What time of day is it?”

What Are the Methods for Successful Vision-LLM Integration

VLMs need large datasets of image-text pairs for training. Multimodal representation learning involves training models to understand and represent information from both text (language) and visual data (images, videos). Pre-training LLMs on large-scale text and then fine-tuning them on multimodal datasets significantly improves their ability to understand and generate textual descriptions of visual content.

Vision-Language Pretrained Models (VLPMs)

VLPMs are where LLMs pre-trained on massive text datasets are adapted to visual tasks through additional training on labeled visual data, have demonstrated considerable success. This method uses the pre-existing linguistic knowledge encoded in LLMs to improve performance on visual tasks with relatively small amounts of annotated data.

Contrastive learning pre-trains VLMs by using large datasets of image-caption pairs to jointly train separate image and text encoders. These encoders map images and text into a shared feature space, minimizing the distance between matching pairs and maximizing it between non-matching pairs, helping VLMs learn similarities and differences between data points.

CLIP (Contrastive Language-Image Pretraining), a popular VLM, utilizes contrastive learning to achieve zero-shot prediction capabilities. It first pre-trains text and image encoders on image-text pairs. During zero-shot prediction, CLIP compares unseen data (image or text) with the learned representations and estimates the most relevant caption or image based on its closest match in the feature space.

natural language processing

CLIP, despite its impressive performance, has limitations such as a lack of interpretability, making it difficult to understand its decision-making process. It also struggles with fine-grained details, relationships, and nuanced emotions, and can perpetuate biases from pretraining data, raising ethical concerns in decision-making systems.

Vision-centric LLMs

Many vision foundation models (VFMs) remain limited to pre-defined tasks, lacking the open-ended capabilities of LLMs. VisionLLM addresses this challenge by treating images as a foreign language, aligning vision tasks with flexible language instructions. An LLM-based decoder then makes predictions for open-ended tasks based on these instructions. This integration allows for better task customization and a deeper understanding of visual data, potentially overcoming CLIP’s challenges with fine-grained details, complex relationships, and interpretability.

VisionLLM can customize tasks through language instructions, from fine-grained object-level to coarse-grained task-level. It achieves over 60% mean Average Precision (mAP) on the COCO dataset, aiming to set a new standard for generalist models integrating vision and language.

However, VisionLLM faces challenges such as inherent disparities between modalities and task formats, multitasking conflicts, and potential issues with interpretability and transparency in complex decision-making processes.

vision LLM visual language model

Source: Wang, Wenhai, et al, VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks

Unified Interface for Vision-Language Tasks

MiniGPT-v2 is a multi-modal LLM designed to unify various vision-language tasks, using distinct task identifiers to improve learning efficiency and performance. It aims to address challenges in vision-language integration, potentially improving upon CLIP by enhancing task adaptability and performance across diverse visual and textual tasks. It can also overcome limitations in interpretability, fine-grained understanding, and task customization inherent in both CLIP and visionLLM models.

The model combines visual tokens from a ViT vision encoder using transformers and self-attention to process image patches. It employs a three-stage training strategy on weakly-labeled image-text datasets and fine-grained image-text datasets. This enhances its ability to handle tasks like image description, visual question answering, and image captioning. The model outperformed MiniGPT-4, LLaVA, and InstructBLIP in benchmarks and excelled in visual grounding while adapting well to new tasks.

The challenges of this model are that it occasionally exhibits hallucinations when generating image descriptions or performing visual grounding. Also, it might describe non-existent visual objects or inaccurately identify the locations of grounded objects.

minigpt

Source: Chen, Jun, et al, MiniGPT-v2: Large Language Model As a Unified Interface for Vision-Language Multi-task Learning

minigpt v2

LENS (Language Enhanced Neural System) Model

Various VLMs can specify visual concepts using external vocabularies but struggle with zero or few-shot tasks and require extensive fine-tuning for broader applications. To resolve this, the LENS model integrates contrastive learning with an open-source vocabulary to tag images, combined with frozen LLMs (pre-trained model used without further fine-tuning).

The LENS model begins by extracting features from images using vision transformers like ViT and CLIP. These visual features are integrated with textual information processed by LLMs like GPT-4, enabling tasks such as generating descriptions, answering questions, and performing visual reasoning. Through a multi-stage training process, LENS combines visual and textual data using cross-modal attention mechanisms. This approach enhances performance in tasks like object recognition and vision-language tasks without extensive fine-tuning.

LENS model

Structured Vision & Language Concepts (SVLC)

Structured Vision & Language Concepts (SVLC) include attributes, relations, and states found in both text descriptions and images. The current VLMs struggles with understanding SVLCs. To tackle this, a data-driven approach aimed at enhancing SVLC understanding without requiring additional specialized datasets was introduced. This approach involved manipulating text components within existing vision and language (VL) pre-training datasets to emphasize SVLCs. Its techniques include rule-based parsing and generating alternative texts using language models.

The experimental findings across multiple datasets demonstrated significant improvements of up to 15% in SVLC understanding, while ensuring robust performance in object recognition tasks. The method sought to mitigate the “object bias” commonly observed in VL models trained with contrastive losses, thereby enhancing applicability in tasks such as object detection and image segmentation.

In conclusion, the integration of LLMs with computer vision through models like VLMs represents a transformative advancement in AI. By merging natural language understanding with visual perception, these models excel in tasks such as image captioning and visual question answering.

Learn the transformative power of integrating LLMs with computer vision from Random Walk. Enhance your AI capabilities to interpret images, generate contextual captions, and excel in diverse applications. Contact us today to harness the full potential of AI integration services for your enterprise.

The post How Can LLMs Enhance Visual Understanding Through Computer Vision? first appeared on Random Walk.

]]>
https://randomwalk.ai/blog/redefining-visual-understanding-integrating-llms-and-visual-ai/feed/ 0 8320