Enterprises are preparing to build their own LLMs — why that’s a smart move

background-abstract-shape-composition-geometric-structure-block-gettyimages-1079488878

Are enterprises ready to build and maintain their own internal large language models (LLMs)?

Artificial intelligence — especially the generative variety — has captivated the unbridled interest of tech professionals and executives alike. Consider this: Despite all the endless talk of budget cuts for cloud services and infrastructure in recent times, the money faucet has opened wide for AI funding. But is it flowing outward to outside services, or inward to resident talent and resources?

Also: Here's how Apple's keeping your cloud-processed AI data safe (and why it matters)

A lot of outside entities, such as OpenAI, Microsoft, and Google, are seen as the primary providers of LLMs, infrastructure support, and expertise. However, interest in internal LLM projects is also on the rise. A new survey of 1,300 CEOs by TCS finds about half of those surveyed, 51%, said they are planning to build their own generative AI implementations. That means a lot of work ahead — but fortunately, the groundwork has already been laid with the publicly available LLMs.

"The foundational LLMs — such as GPT, Claude, Llama — can be best described as world-wise; can be seen as repackaging Internet knowledge," Dr. Harrick Vin, chief technology officer for TCS and co-author of the study, told ZDNET. "They also possess high levels of multi-modal understanding and generation capabilities, along with reasoning abilities."

Constructing these foundational models "is complex and expensive," said Vin, who pointed out that internal enterprise models would build upon the capabilities of these models. "These models will leverage the basic skills of foundational models — such as language understanding and generation, reasoning, and general knowledge. But they need to extend and specialize them to the industry, enterprise and activity context."

Fortunately, "construction of such specialized models is far easier and inexpensive as compared to the development of foundational models," said Vin. "In fact, the relative ease of specializing foundational LLMs, which are broad-AI models, to create purpose-specific AI models and solutions is the primary reason for the democratization of AI."

These enterprise-specific LLMs are "referring to industry, enterprise, and activity-wise models constructed using some of the foundational models, either open-source or commercial," he continued. "We believe that an AI-mature enterprise in the future will have hundreds, or thousands, of purposive AI models, all built by compositing capabilities of foundational models with specific enterprise-specific capabilities."

Also: Businesses' cloud security fails are 'concerning' — as AI threats accelerate

Beyond building and implementing models, the business needs to be prepped for generative AI. More than half (55%) said they were actively making changes right now to their business or operating models, or to their products and services, to accommodate AI Four in 10 executives said that in the future they have "a lot of changes to make to their business" before they can take full advantage of AI, the TCS survey shows.

This points to a slow but powerful uptake of both generative and operational AI. Over the past year, "every enterprise has experimented with gen AI use cases — and 2024 and beyond will be about scaling value," said Vin. "During the experimentation phase, however, every enterprise has realized that scaling value is challenging."

At this time, only 17% are discussing Al and making enterprise-wide plans for it, the TCS survey shows. In addition, only 28% are ready to establish an enterprise-wide AI strategy to maximize its benefits to the company. Still, Vin sees a rapid upsurge of the technology. "There is a difference between implementing AI solutions on an ad hoc or a case-by-case basis, to building an enterprise-wide plan to build an AI-mature enterprise," Vin said. "The relatively low numbers in the survey refer to the creation of such enterprise-wide strategies. This is expected."

Also: Make room for RAG: How Gen AI's balance of power is shifting

As far as the adoption of AI solutions goes, Vin continued, "the numbers are quite high: 59% of corporate functions have AI implementations in-process or completed and another 34% are planning AI implementations. We are in the early phase of both technological maturity as well as an enterprise adoption, at scale, maturity. Most enterprises are starting to leverage AI and genAI for specific use cases, while embarking upon a longer-term journey to quantify benefits as well as manage corresponding cost and risks."

Over the past year, "every enterprise has experimented with genAI use cases — and 2024 and beyond will be about scaling value," said Vin. "During the experimentation phase, however, every enterprise has realized that scaling value is challenging."
For starters, "building effective AI solutions requires high-quality data," he said. "Whereas enterprises do have a lot of data, it is often distributed across many, mutually inconsistent islands. Whereas most enterprises have embarked upon consolidation and data estate modernization journeys over the past several years, these journeys are far from complete. Further, the migration of the data estate to cloud environments is work-in-progress for most enterprises. This makes it difficult for enterprises to leverage cloud-hosted foundational LLMs along with their enterprise data."
Also: Generative AI may be creating more work than it saves

In addition, enterprises "will need to improve their maturity to manage data lineage, usage, security and privacy proactively," said Vin. "They will need to master the art of determining what data can be used for what purpose, even inadvertently, to prevent biases and unfair practices. This is not just a design-time challenge, but also a run-time challenge." Needed are systems "to detect, in real-time, emergent conditions where AI models start deviating from expected behavior."
Finally, roles and skills requirements are changing faster than companies can keep up. "With the infusion of AI, the role of enterprise knowledge workers will change from doers of work to trainers and interrogators of machines, reviewers of work done by machines, as well as owners of critical thinking and creativity," Vin said.

Artificial Intelligence

Databricks Unveils LakeFlow, Simplifying Data Ingestion, Transformation & Orchestration 

Data migration from on-prem to cloud at scale using Databricks

Databricks announced the launch of Databricks LakeFlow, a unified solution that streamlines all aspects of data engineering, from data ingestion to transformation and orchestration. LakeFlow enables data teams to ingest data at scale from various sources efficiently, transform it using SQL and Python, and confidently deploy and operate pipelines in production.

With LakeFlow, data teams can now easily ingest data from databases such as MySQL, Postgres, and Oracle, as well as enterprise applications like Salesforce, Dynamics, Sharepoint, Workday, NetSuite, and Google Analytics. Databricks is also introducing Real Time Mode for Apache Spark, allowing ultra-low latency stream processing.

LakeFlow automates the deployment, operation, and monitoring of pipelines at scale in production, with built-in support for CI/CD and advanced workflows that support triggering, branching, and conditional execution. Data quality checks and health monitoring are integrated with alerting systems like PagerDuty.

LakeFlow simplifies the building and operating of production-grade data pipelines while addressing the most complex data engineering use cases, enabling even the busiest data teams to meet the growing demand for reliable data and AI.

Data engineering is crucial for democratising data and AI within businesses, but it remains a challenging and complex field. Data teams face issues such as ingesting data from siloed and proprietary systems, maintaining intricate logic for data preparation, and dealing with failures and latency spikes that can lead to operational disruptions. Existing solutions are often fragmented and incomplete, resulting in low data quality, reliability issues, high costs, and an increasing backlog of work.

LakeFlow addresses these challenges by simplifying all aspects of data engineering through a single, unified experience built on the Databricks Data Intelligence Platform. It integrates deeply with Unity Catalog for end-to-end governance and leverages serverless compute for highly efficient and scalable execution.

LakeFlow Connect provides a breadth of native, scalable connectors for databases and enterprise applications, fully integrated with Unity Catalog for robust data governance. It incorporates the low-latency, highly efficient capabilities of Arcion, which Databricks acquired in November 2023. LakeFlow Connect makes all data available for batch and real-time analysis, regardless of size, format, or location.

LakeFlow Pipelines, built on Databricks’ highly scalable Delta Live Tables technology, allows data teams to implement data transformation and ETL in SQL or Python. It introduces Real Time Mode for low-latency streaming without code changes, eliminates the need for manual orchestration, and unifies batch and stream processing. LakeFlow Pipelines simplifies even the most complex streaming and batch data transformations.

LakeFlow Jobs provides automated orchestration, data health, and delivery spanning scheduling notebooks, SQL queries, ML training, and automatic dashboard updates. It offers enhanced control flow capabilities and full observability to detect, diagnose, and mitigate data issues for increased pipeline reliability. LakeFlow Jobs automates deploying, orchestrating, and monitoring data pipelines in a single place.

“LakeFlow addresses the challenges data teams face in building and operating reliable data pipelines,” said Ali Ghodsi, CEO and co-founder of Databricks. “By simplifying all aspects of data engineering in a unified experience, LakeFlow enables data teams to efficiently meet the growing demand for reliable data and AI.”

The post Databricks Unveils LakeFlow, Simplifying Data Ingestion, Transformation & Orchestration appeared first on AIM.

Shutterstock & Databricks Launch ImageAI: Customisable Text-to-Image AI for Enterprises

Databricks

Amongst the many announcements Databricks made today, an important one was the launch of Shutterstock ImageAI, Powered by Databricks. This new text-to-image generative AI model enables enterprises to quickly create high-quality, commercially viable images tailored to their specific business needs.

ImageAI leverages the advanced capabilities of Databricks Mosaic AI and was trained exclusively on Shutterstock’s vast repository of curated, high-quality images. The model allows companies to generate photorealistic visuals that meet enterprise standards for data governance, security, and intellectual property rights.

“Shutterstock is excited to partner with Databricks to bring ImageAI to life. This collaboration underscores our commitment to advancing responsible AI and providing our customers with innovative tools that enhance their creative workflows,” said Aimee Egan, Chief Enterprise Officer at Shutterstock. “ImageAI creates trusted visuals that businesses can own, with unparalleled customization and safety for enterprise use.”

Naveen Rao, Vice President of AI at Databricks, stated, “At Databricks, we believe that companies can and should be in the driver’s seat when it comes to building their own custom GenAI models on their data. Shutterstock offers one of the most comprehensive visual datasets in the world, and we’re thrilled they chose to build their production-quality, text-to-image model on top of the Databricks Data Intelligence Platform.”

Key benefits of Shutterstock ImageAI, Powered by Databricks include:
Trusted outputs, as the model was trained solely on Shutterstock’s vetted image repository
Ability to quickly adapt to unique business requirements and generate customized images
Secure integration with enterprise applications via the Databricks platform

ImageAI was pre-trained from scratch in just weeks using Databricks Mosaic AI Model Training. Enterprises can serve the model using Mosaic AI Model Serving to integrate it with their existing data governance policies seamlessly.

Shutterstock ImageAI is now available in private preview on Databricks Mosaic AI Model Serving and live on Shutterstock.com. Companies interested in building their own differentiated generative AI models can leverage the Databricks Data Intelligence Platform to do so efficiently.

The post Shutterstock & Databricks Launch ImageAI: Customisable Text-to-Image AI for Enterprises appeared first on AIM.

Helen Toner worries ‘not super functional’ Congress will flub AI policy

Helen Toner, a former OpenAI board member and the director of strategy at Georgetown’s Center for Security and Emerging Technology, is worried Congress might react in a “knee-jerk” way where it concerns AI policymaking, should the status quo not change. “Congress right now — I don’t know if anyone’s noticed — is not super functional, […]

© 2024 TechCrunch. All rights reserved. For personal use only.

IBM, Cleveland Clinic, And Hartree Center Collaborate To Advance Healthcare Through Advanced Computing

Cleveland Clinic, IBM, and Hartree Center have partnered to advance healthcare and life sciences through advanced computing technologies, including AI and quantum computing.

The partnership brings together an international team of researchers, physicians, and scientists from various disciplines around the globe.

“The research teams will leverage high performance and quantum computing to advance life sciences, with the goal of improving healthcare and accelerating new treatments for patients around the world. Cleveland Clinic London will be a central link between innovative clinical care in the UK and Cleveland Clinic’s global footprint.” said Cleveland Clinic chief research information officer Lara Jehi.

Epilepsy is a chronic noncommunicable disease of the brain that affects people of all ages. More than 50 million people worldwide suffer from this neurological disease, with more than 80% of epilepsy patients living in low or middle-income countries.

According to research, around 70% of people living with epilepsy can live seizure-free with proper diagnosis and treatment, however, this is hardly the case with nearly 75% of patients that live in low-income countries not getting the treatment they need.

Cleveland Clinic has one of the most comprehensive programs in the world for the diagnosis and treatment of epilepsy. Now with the newly formed alliance with tech giant IBM, and the Science and Technology Facilities Council (STFC) at the Hartfree Center, researchers from Cleveland Clinic aim to use advanced technologies to study epilepsy and the impact of hospital interventions.

The researchers will kick off the new collaboration with two projects, led by Dr. Jehi, an epilepsy researcher, and Charles Knowles, Ph.D., Chief Academic Officer at Cleveland Clinic London, working closely with teams from IBM and Hartree Centre.

The first project, led by Dr. Knowles, will use advanced AI models to analyze the impact of care on patients at Cleveland Clinic London. The researchers will focus on how the common hospital procedures impacted the patient's overall quality of life and health. The goal of this project is to gather valuable insights in how to enhance patient outcomes.

A key component of this project is using clinical and advanced imaging data provided by Cleveland Clinic London BioResource, a repository of donated samples and health data that can be used for scientific research.

The data from these Biorepositories will be used to develop larger AI models that can integrate multiple types of data for analysis across different health conditions.

The second project, led by Dr. Jehi, aims to utilize quantum computing to analyze extensive data sets and identify molecular characteristics in the body that more accurately predict surgical outcomes in epilepsy patients. The primary goal of this project is to identify novel biomarkers that can be useful in customizing treatment plans to improve patient outcomes.

Dr Jehi was one of the featured speakers at the third Economist Impact Commercializing Quantum Global 2024 conference in London last week. Her project of using quantum computing for the discovery of biomarkers was highlighted as an example of the power of quantum computing in developing precision medicine.

AI is playing a transformative role in advancing healthcare across various domains, including research on the human brain. In May, a team of researchers from Harvard and Google collaborated to publish an astonishingly detailed 3D map of the human brain. Such breakthroughs can bring valuable insights into the mysteries of the human brain and help unlock better methods to diagnose and treat conditions and diseases like Epilepsy.

Related Items

How AI Will Impact the Drug Discovery Pipeline

The Future of Healthcare is Data-driven

John Snow Labs Achieves Record Growth with Advanced AI Models in Healthcare

4 Apple AI features that ChatGPT already offers (and 2 more that are coming soon)

screenshot-2024-06-10-at-2-14-04pm.png

After ChatGPT's launch in November 2022, nearly every company joined in on the AI craze — except Apple. Nearly two years later, at its annual Worldwide Developer Conference (WWDC), Apple unveiled a collection of AI features known as Apple Intelligence. While they may seem impressive, many have been done before by none other than ChatGPT.

This fall, Apple's software updates will make several generative AI capabilities available on iPhones, iPads, and Macs. However, OpenAI already unveiled major upgrades to the free version of ChatGPT in May, and the similarities to what Apple Intelligence will do are worth noting.

Also: Five iOS 18 features that Android users already have

The Apple Intelligence updates will be free, but the full experience will only be available on iPhones with the A17 Pro chip, which currently only includes iPhone 15 Pro and iPhone 15 Pro Max, and iPads and Macs with the M family of chips.

Instead of spending thousands of dollars rushing to upgrade to the newest Apple devices, you should check out ChatGPT first. OpenAI's free chatbot has many of the same features coming to Apple Intelligence.

1. Writing tools

With Apple Intelligence, users will be able to access a variety of writing tools that can help with rewriting, proofreading, and summarizing text. Apple says the tools will be accessible "everywhere" users write, including in Mail, Keynote, third-party apps, and more.

ChatGPT's advanced natural language processing (NLP) makes it a great writing tool as well. It can generate new text from scratch, proofread, coedit, rewrite, and more. While the ChatGPT experience may not live natively within Apple devices the way Apple Intelligence will, users can easily copy and paste its output into a tab while accessing ChatGPT in their browser.

Also: How ChatGPT (and other AI chatbots) can help you write an essay

Another option for Apple users is to take advantage of the ChatGPT app for iPads and iPhones. There's even a ChatGPT app for MacOS, which allows users to access the chatbot quickly via a keyboard shortcut. The Mac app is available now for ChatGPT Plus subscribers, but OpenAI is planning to release access to all users in the coming months.

2. Image generator

Apple also unveiled its first text-to-image generator, Apple's Image Playground. This generator will be built into iOS 18, iPadOS 18, and MacOS Sequoia as part of Apple Intelligence, and will also live as a stand-alone app.

Also: The best AI image generators to try right now

It's unclear without testing it, but Image Playground's functionality will likely be similar to OpenAI's image generator, DALL-E 3, which can be accessed via ChatGPT Plus. Even though it requires a $20 per month subscription, it offers users a wider variety of options, since it can render images in any style. Image Playground is limited to three styles: Animation, illustration, and sketch.

3. On-screen awareness

At Apple's event, the company shared that with Apple Intelligence, Siri will have on-screen awareness, making asking it for help with certain tasks easier. During OpenAI's Spring Launch event, the company also showed a demo indicating that ChatGPT will have on-screen awareness as well, as seen below — though it didn't clarify when we'll see this feature.

The value of having an AI voice assistant that can see what you are working on and use it as context for your query is evident, and it will likely be the future of all assistants.

4. Advanced conversational capabilities

Another update coming to Siri is better NLP, meaning it will be able to understand you even if you stutter or pause. OpenAI indicated at its Spring Launch that the improved Voice Mode for ChatGPT will have the same capabilities, such as stopping when a user interrupts it, understanding queries better, and more. The improved Voice Mode will be rolling out in alpha in the coming weeks, and ChatGPT Plus users will get early access as the company rolls it out more broadly.

Also: Everything to know about Apple's AI features for iPhones, Macs, and iPads

5. Type and chat

With Apple Intelligence, Siri will be upgraded to accept typed and voice queries — a significant change considering that it has only ever functioned as a voice assistant. However, as discussed above, ChatGPT can also take text and voice inputs.

6. And, of course…access to ChatGPT

Apple also announced that Siri will have access to ChatGPT, which you can also access by going directly to the source.

Apple

Databricks Launches AI/BI: A Compound AI System for Intelligent Business Insights

Databricks announced the launch of Databricks AI/BI, a new business intelligence (BI) product that aims to democratise analytics and insights across organisations through an AI-first approach.

AI/BI leverages generative AI to enable self-service analytics, allowing everyday users to ask complex questions and receive accurate answers without requiring data science expertise.

AI/BI consists of two complementary experiences: AI/BI Dashboards, a low-code interface for quickly creating interactive dashboards; and AI/BI Genie, a conversational interface that uses natural language to address ad-hoc and follow-up questions. Both are powered by a compound AI system that continuously learns from usage across an organisation’s data stack, including ETL pipelines, lineage, and queries.

“A truly intelligent BI solution needs to understand the unique semantics and nuances of a business to effectively answer questions for business users,” said Ali Ghodsi, Co-founder and CEO at Databricks. “The launch of AI/BI is a step towards building such a system.”

Unlike other BI tools that have attempted to add generative AI capabilities on top of conventional architectures, AI/BI places the AI system at the core. It utilises an ensemble of specialised AI agents that work together to reason about business questions and generate useful answers. The system learns and improves based on human feedback, allowing it to persist knowledge beyond a single analysis.

Key benefits of AI/BI include:
Unified governance and lineage through deep integration with Databricks Unity Catalog
Effortless secure sharing without additional user licenses
Industry-leading price-performance across data volumes
No data extraction required, ensuring data freshness and simpler governance

“At SEGA, we aim to entertain the world with creative, innovative experiences, and data intelligence plays an important role in achieving that goal,” said Felix Baker, Head of Data Services, SEGA Europe. “AI/BI will enable us to democratise data, increase productivity, and enhance the speed of data-driven decision making throughout SEGA.”

AI/BI is included for all Databricks SQL Pro and Serverless customers at no additional licensing cost beyond compute. AI/BI Dashboards is generally available starting today, while Genie is in public preview.

The post Databricks Launches AI/BI: A Compound AI System for Intelligent Business Insights appeared first on AIM.

Comparative Analysis of LangChain and LlamaIndex

Comparative Analysis of LangChain and LlamaIndex
Image by Editor | Midjourney

Rapid technological development has recently taken the fields of artificial intelligence (AI) and large language models (LLMs) to new heights. To cite a few advances in this area, LangChain and LlamaIndex have emerged as major players. Each has its unique set of capabilities and strengths.

This article compares the battle between these two fascinating technologies, comparing their features, strengths, and real-world applications. If you are an AI developer or an enthusiast, this analysis will help you understand which tool might fit your needs.

LangChain

LangChain is a comprehensive framework designed for building applications driven by LLMs. Its primary objective is to simplify and enhance the entire lifecycle of LLM applications, making it easier for developers to create, optimize, and deploy AI-driven solutions. LangChain achieves this by offering tools and components that streamline the development, productionisation, and deployment processes.

Tools LangChain Offers

LangChain's tools include model I/O, retrieval, chains, memory, and agents. All these tools are explained in detail below:

Model I/O: At the heart of LangChain's capabilities lies the Module Model I/O (Input/Output), a crucial component for leveraging the potential of LLMs. This feature offers developers a standardized and user-friendly interface to interact with LLMs, simplifying the creation of LLM-powered applications to address real-world challenges.

Retrieval: In many LLM applications, personalized data must be incorporated beyond the models' original training scope. This is achieved through Retrieval Augmented Generation (RAG), which involves fetching external data and supplying it to the LLM during the generation process.

Chains: While standalone LLMs suffice for simple tasks, complex applications demand the intricacy of chaining LLMs together in collaboration or with other essential components. LangChain offers two overarching frameworks for this enchanting process: the traditional Chain interface and the modern LangChain Expression Language (LCEL). While LCEL reigns supreme for composing chains in new applications, LangChain also provides invaluable pre-built Chains, ensuring the seamless coexistence of both frameworks.

Memory: Memory in LangChain refers to storing and recalling past interactions. LangChain provides various tools to integrate memory into your systems, accommodating simple and complex needs. This memory can be seamlessly incorporated into chains, enabling them to read from and write to stored data. The information held in memory guides LangChain Chains, enhancing their responses by drawing on past interactions.

Agents: Agents are dynamic entities that utilize the reasoning capabilities of LLMs to determine the sequence of actions in real-time. Unlike conventional chains, where the sequence is predefined in the code, Agents use the intelligence of language models to decide the next steps and their order dynamically, making them highly adaptable and powerful for orchestrating complex tasks.

This image shows the architecture of the LangChain framework
This image shows the architecture of the LangChain framework | source: Langchain documentation

The LangChain ecosystem comprises the following:

  • LangSmith: This helps you trace and evaluate your language model applications and intelligent agents, helping you move from prototype to production.
  • LangGraph: is a powerful tool for building stateful, multi-actor applications with LLMs. It is built on top of (and intended to be used with) LangChain primitives.
  • LangServe: Using this tool, you can deploy LangChain runnables and chains as REST APIs.

LlamaIndex

LlamaIndex is a sophisticated framework designed to optimize the development and deployment of LLMs-powered applications. It provides a structured approach to integrating LLMs into application software, enhancing their functionality and performance through a unique architectural design.

Formerly known as the GPT Index, LlamaIndex emerged as a dedicated data framework tailored to bolster and elevate the functionalities of LLMs. It concentrates on ingesting, structuring, and retrieving private or domain-specific data, presenting a streamlined interface for indexing and accessing pertinent information within vast textual datasets.

Tools LlamaIndex Offers

Some of the tools LlamaIndex offers include data connectors, engines, data agents, and application integrations. All these tools are explained in detail below:

Data connectors: Data connectors play a crucial role in data integration, simplifying the complex process of linking your data sources to your data repository. They eliminate the need for manual data extraction, transformation, and loading (ETL), which can be cumbersome and prone to errors. These connectors streamline the process by ingesting data directly from its native source and format, saving time on data conversion. Additionally, data connectors automatically enhance data quality, secure data through encryption, boost performance via caching, and reduce the maintenance required for your data integration solution.

Engines: LlamaIndex Engines enable seamless collaboration between data and LLMs. They provide a flexible framework that connects LLMs to various data sources, simplifying access to real-world information. These engines feature an intuitive search system that understands natural language queries, facilitating easy data interaction. They also organize data for quicker access, enrich LLM applications with additional information, and assist in selecting the appropriate LLM for specific tasks. LlamaIndex Engines are essential for creating various LLM-powered applications, bridging the gap between data and LLMs to address real-world challenges.

Data agents: Data agents are intelligent, LLM-powered knowledge workers within LlamaIndex who are adept at managing your data. They can intelligently navigate through unstructured, semi-structured, and structured data sources and interact with external service APIs in an organized manner, handling both "read" and "write" operations. This versatility makes them indispensable for automating data-related tasks. Unlike query engines limited to reading data from static sources, Data Agents can dynamically ingest and modify data from various tools, making them highly adaptable to evolving data environments.

Application integrations: LlamaIndex excels in building LLM-powered applications, with its full potential realized through extensive integrations with other tools and services. These integrations facilitate easy connections to a wide range of data sources, observability tools, and application frameworks, enabling the development of more powerful and versatile LLM-powered applications.

Implementation Comparison

These two technologies can be similar when it comes to building applications. Let's take a chatbot as an example. Here is how you can build a local chatbot using LangChain:

from langchain.schema import HumanMessage, SystemMessage   from langchain_openai import ChatOpenAI     llm = ChatOpenAI(      openai_api_base="http://localhost:5000",       openai_api_key="SK******",      max_tokens=1600,      Temperature=0.2     request_timeout=600,  )   chat_history = [      SystemMessage(content="You are a copywriter."),      HumanMessage(content="What is the meaning of Large language Evals?"),   ]   print(llm(chat_history))

This is how you build a local chatbot using LlamaIndex:

from llama_index.llms import ChatMessage, OpenAILike     llm = OpenAILike(      api_base="http://localhost:5000",      api_key=”******”,     is_chat_model=True,      context_window=32768,     timeout=600,        )   chat_history = [      ChatMessage(role="system", content="You are a copywriter."),      ChatMessage(role="user", content="What is the meaning of Large language Evals?"),   ]   output = llm.chat(chat_history)   print(output)

Main Differences

While LangChain and LlamaIndex may exhibit certain similarities and complement each other in constructing resilient and adaptable LLM-driven applications, they are quite different. Below are notable distinctions between the two platforms:

Criteria LangChain LlamaIndex
Framework Type Development and deployment framework. Data framework for enhancing LLM capabilities.
Core Functionality Provides building blocks for LLM applications. Focuses on ingesting, structuring, and accessing data.
Modularity Highly modular with various independent packages. Modular design for efficient data management.
Performance Optimized for building and deploying complex applications. Excels in text-based search and data retrieval.
Development Uses open-source components and templates. Offers tools for integrating private/domain-specific data
Productionisation LangSmith for monitoring, debugging, and optimization. Emphasizes high-quality responses and precise queries.
Deployment LangServe to turn chains into APIs. No specific deployment tool mentioned.
Integration Supports third-party integrations through langchain-community. Integrates with LLMs for enhanced data handling.
Real-World Applications Suitable for complex LLM applications across industries. Ideal for document management and precise information retrieval.
Strengths Versatile, supports multiple integrations, strong community. Accurate responses, efficient data handling, robust tools.

Final Thoughts

Depending on its specific needs and project goals, any application powered by LLMs can benefit from using either LangChain or LlamaIndex. LangChain is known for its flexibility and advanced customization options, making it ideal for context-aware applications.

LlamaIndex excels in rapid data retrieval and generating concise responses, making it perfect for knowledge-driven applications such as chatbots, virtual assistants, content-based recommendation systems, and question-answering systems. Combining the strengths of both LangChain and LlamaIndex can help you build highly sophisticated LLM-driven applications.

Resources

  • LlamaIndex
  • LangChain documentation

Shittu Olumide is a software engineer and technical writer passionate about leveraging cutting-edge technologies to craft compelling narratives, with a keen eye for detail and a knack for simplifying complex concepts. You can also find Shittu on Twitter.

More On This Topic

  • OLAP vs. OLTP: A Comparative Analysis of Data Processing Systems
  • Pandas vs. Polars: A Comparative Analysis of Python's Dataframe Libraries
  • Build Your Own PandasAI with LlamaIndex
  • A Comparative Overview of the Top 10 Open Source Data Science Tools in 2023
  • LangChain 101: Build Your Own GPT-Powered Applications
  • Transforming AI with LangChain: A Text Data Game Changer

Databricks Partners with NVIDIA to Unleash ‘Sovereign AI’ in Enterprise

At its Data+ AI Summit, Databricks announced an expanded collaboration with NVIDIA to optimise data and AI workloads by integrating NVIDIA CUDA-accelerated computing into the core of Databricks’ Data Intelligence Platform.

The partnership aims to boost the efficiency, accuracy, and performance of AI development pipelines for modern AI factories, as data preparation, curation, and processing are crucial for leveraging enterprise data in generative AI applications.

Through this broadened alliance, Databricks is adding native support for NVIDIA GPU acceleration on its Data Intelligence Platform. The announcement builds upon the companies’ existing collaboration to enrich enterprises’ experiences across various use cases, from training classical machine learning models to building and deploying generative AI applications and optimising digital twins.

“We’re thrilled to continue growing our partnership with NVIDIA to deliver on the promise of data intelligence for our customers from analytics use cases to AI,” said Ali Ghodsi, Co-founder and CEO at Databricks. “Together with NVIDIA, we’re excited to help every organisation build their own AI factories on their own private data.”

Jensen Huang, founder and CEO of NVIDIA, emphasised the importance of accelerated computing in reducing data processing energy demands for sustainable AI platforms. “By bringing NVIDIA CUDA acceleration to Databricks’ core computing stack, we’re laying the foundation for customers everywhere to use their data to power enterprise generative AI,” Huang stated.

A key aspect of the partnership involves Databricks developing native support for NVIDIA-accelerated computing in its next-generation vectorised query engine, Photon. This integration is expected to deliver improved speed and efficiency for customers’ data warehousing and analytics workloads. Photon powers Databricks SQL, the company’s serverless data warehouse known for its industry-leading price-performance and total cost of ownership (TCO). The collaboration is anticipated to lead to the next frontier of price-performance.

Databricks Shares a Unique Partnership with NVIDIA

In the backdrop of Databricks’ Data + AI Summit 2024, Anil Bhasin, the vice president of India and SAARC region at Databricks, told AIM that the company’s partnership with NVIDIA—one of its strategic investors–is significant, alongside helping them improve run times using their SOTA GPUs.

“We’ve always been known as pioneers of the lake house architecture, and now we’ve created a new category called the data intelligence platform. We’ve embedded generative AI in the lake house, which is a unique approach not many companies are taking,” said Bhasin, saying that NVIDIA is aligned with their vision because the future lies in data intelligence platforms.

He said this allows them to serve every use case, ingest data from any source, and maintain unified governance. This strategic differentiation makes their partnership with NVIDIA truly special and aligns perfectly with NVIDIA’s ‘Sovereign AI’ for enterprises. Nobody other than Databricks is enabling this.

“The ability for us to query in natural language, converting it to SQL on the back end, empowers the business user to gain insights. That is true democratisation,” avered Bhasin, saying their vision is powerful, and not just NVIDIA; many companies believe in Databricks’ long-term vision.

NVIDIA x Databricks

Recently, Databricks’ open-source model DBRX became available as an NVIDIA NIM microservice. NVIDIA NIM inference microservices provide fully optimised, pre-built containers for deployment anywhere, significantly increasing enterprise developer productivity by offering a simple, standardised way to add generative AI models to their applications.

Launched in March 2024, DBRX was built entirely on top of Databricks, leveraging the platform’s tools and techniques, and was trained with NVIDIA DGX Cloud, a scalable end-to-end AI platform for developers.

The Databricks Data Intelligence Platform offers a comprehensive solution for building, evaluating, deploying, securing, and monitoring end-to-end generative AI applications. With Databricks Mosaic AI’s data-centric approach, customers benefit from an open, flexible platform to easily scale generative AI applications on their unique data while ensuring safety, accuracy, and governance.

Today’s announcement follows Databricks’ strategic acquisition of Tabular, a data management startup founded by the original creators of Apache Iceberg and Linux Foundation Delta Lake, the two leading open-source lakehouse formats. By bringing together these key players, Databricks aims to lead the way in data compatibility, ensuring organisations are no longer limited by the format of their data.

Driven by the growing demand for data and AI capabilities, Databricks achieved over $1.6 billion in revenue for its fiscal year ending January 31, 2024, representing more than 50% year-over-year growth.

The expanded partnership between Databricks and NVIDIA underscores the critical role of accelerated computing and optimised data processing in enabling enterprises to harness the power of generative AI effectively and efficiently.

The post Databricks Partners with NVIDIA to Unleash ‘Sovereign AI’ in Enterprise appeared first on AIM.

Businesses’ cloud security fails are ‘concerning’ — as AI threats accelerate

umbrella-on-abstract-technological-background-gettyimages-1312330537

Just under 45% of organizations conduct regular audits and assessments to ensure their cloud environment is secured, which is "concerning" as more applications and workloads are moved to multi-cloud platforms.

Asked how they were monitoring risk across their cloud infrastructure, 47.7% of businesses pointed to automated security tools while 46.5% relied on native security offerings from their providers. Another 44.7% said they conducted regular audits and assessments, according to a report from security vendor Bitdefender.

Also: AI is changing cybersecurity and businesses must wake up to the threat

Some 42.1% worked with third-party experts, revealed the study, which surveyed more than 1,200 IT and security professionals including chief information security officers across six markets: Singapore, the UK, France, Germany, Italy, and the US.

It is "definitely concerning" that only 45% of companies regularly run audits of their cloud environments, said Paul Hadjy, Bitdefender's vice president of Asia-Pacific and cyber security services, in response to questions from ZDNET.

Hadjy noted that an over-reliance on cloud providers' ability to protect hosted services or data persists even as businesses continue moving applications and workloads to multi-cloud environments.

"Most times, [cloud providers] are not as responsible as you would think and, in most cases, the data being stored in the cloud is large and often sensitive," Hadjy said.

"The responsibility of cloud security, including how data is protected at rest or in motion, identities [of] people, servers, and endpoints granted access to resources, and compliance is predominantly up to the customer. It's important to first establish a baseline to determine current risk and vulnerability in your cloud environments based on things such as geography, industry, and supply chain partners."

Among the top security concerns respondents had in managing their company's cloud environments, 38.7% cited identity and access management while 38% pointed to the need to maintain cloud compliance. Another 35.9% named shadow IT as a concern and 32% were worried about human error, the study found.

When it comes to generative AI-related threats, however, respondents seem confident in their teammates' ability to identify potential attacks. A majority 74.1% believed colleagues from their department would be able to spot a deepfake video or audio attack, with US respondents showing the highest level of confidence at 85.5%.

Also: Code faster with generative AI, but beware the risks when you do

In comparison, just 48.5% of their counterparts in Singapore were confident their teammates could spot a deepfake — the lowest among the six markets. In fact, 35% in Singapore said colleagues from their department would not be able to identify a deepfake, which was the highest in the global pool to say likewise.

Was the global average of 74.1% who were confident their teammates could spot a deepfake misplaced or well-placed?

Hadjy noted that this confidence was expressed even though 96.6% viewed GenAI as a minor to very significant threat. A base-level explanation for this is that IT and security professionals do not necessarily trust the ability of users beyond their own teams — and who are not in IT or security — to spot deepfakes, he said.

"This is why we believe technology and processes [implemented] together are the best way to mitigate this risk," he added.

Asked how effective or accurate existing tools are in detecting AI-generated content such as deepfakes, he said this would depend on several factors. If delivered via phishing email or embedded in a text message with a malicious link, deepfakes should be quickly identified by endpoint protection tools, such as XDR (extended detection and response) tools, he explained.

However, he noted that threat actors depend on a human's natural tendencies to believe what they see and what is endorsed by people they trust, such as celebrities and high-profile personalities — whose images often are manipulated to deliver messages.

Also: 3 ways to accelerate generative AI implementation and optimization

And as deepfake technologies continue to evolve, he said it would be "nearly impossible" to detect such content via sight or sound alone. He underscored the need for technology and processes that can detect deepfakes to also evolve.

Although Singapore respondents were the most skeptical of their teammates' ability to spot deepfakes, he noted that 48.5% is a significant number.

Urging again the importance of having both technology and processes in place, Hadjy said: "Deepfakes will continue to get better, and effectively spotting them will take continuous efforts that combine people, technology, and processes all working together. In cybersecurity, there is no 'silver bullet' — it's always a multi-layer strategy that starts with strong prevention to close the door before a threat gets in."

Training also is increasingly critical as more employees work in hybrid environments and more risks originate from homes. "Businesses need to have clear steps in place to validate deepfakes and protect against highly targeted spearphishing campaigns," he said. "Processes are key for organizations to help ensure measures for double checking are in place, especially in instances where the transfer of large sums of money is involved."

According to the Bitdefender study, 36.1% view GenAI technology as a very significant threat with regards to the manipulation or creation of deceptive content, such as deepfakes. Another 45.1% described this as a moderate threat while 15.4% said it was a minor threat.

Also: Nearly 50% of people want an AI clone to do this for them

A large majority, at 94.3%, were confident in their organization's ability to respond to current security threats, such as ransomware, phishing, and zero-day attacks.

However, 57% admitted having experienced a data breach or leak in the past year, up 6% from the previous year, the study revealed. This number was lowest in Singapore at 33% and the highest in the UK at 73.5%.

Phishing and social engineering was the top concern at 38.5%, followed by ransomware, insider threats, and software vulnerabilities at 33.5% each.

Security