GlobeNewswire | October 19, 2023
Alluxio, the data platform company for all data-driven workloads, today introduced Alluxio Enterprise AI, a new high-performance data platform designed to meet the rising demands of Artificial Intelligence (AI) and machine learning (ML) workloads on an enterprise’s data infrastructure. Alluxio Enterprise AI brings together performance, data accessibility, scalability and cost-efficiency to enterprise AI and analytics infrastructure to fuel next-generation data-intensive applications like generative AI, computer vision, natural language processing, large language models and high-performance data analytics.
To stay competitive and achieve stronger business outcomes, enterprises are in a race to modernize their data and AI infrastructure. On this journey, they find that legacy data infrastructure cannot keep pace with next-generation data-intensive AI workloads. Challenges around low performance, data accessibility, GPU scarcity, complex data engineering, and underutilized resources frequently hinder enterprises' ability to extract value from their AI initiatives. According to Gartner®, “the value of operationalized AI lies in the ability to rapidly develop, deploy, adapt and maintain AI across different environments in the enterprise. Given the engineering complexity and the demand for faster time to market, it is critical to develop less rigid AI engineering pipelines or build AI models that can self-adapt in production.” “By 2026, enterprises that have adopted AI engineering practices to build and manage adaptive AI systems will outperform their peers in the operationalizing AI models by at least 25%.”
Alluxio empowers the world’s leading organizations with the most modern Data & AI platforms, and today we take another significant leap forward, said Haoyuan Li, Founder and CEO, Alluxio. Alluxio Enterprise AI provides customers with streamlined solutions for AI and more by enabling enterprises to accelerate AI workloads and maximize value from their data. The leaders of tomorrow will know how to harness transformative AI and become increasingly data-driven with the newest technology for building and maintaining AI infrastructure for performance, seamless access and ease of management.
With this announcement, Alluxio expands from a one-product portfolio to two product offerings - Alluxio Enterprise AI and Alluxio Enterprise Data - catering to the diverse needs of analytics and AI. Alluxio Enterprise AI is a new product that builds on the years of distributed systems experience accumulated from the previous Alluxio Enterprise Editions, combined with a new architecture that is optimized for AI/ML workloads. Alluxio Enterprise Data is the next-gen version of Alluxio Enterprise Edition, and will continue to be the ideal choice for businesses focused primarily on analytic workloads.
Accelerating End-to-End Machine Learning Pipeline
Alluxio Enterprise AI enables enterprise AI infrastructure to be performant, seamless, scalable and cost-effective on existing data lakes. Alluxio Enterprise AI helps data and AI leaders and practitioners achieve four key objectives in their AI initiatives: high-performance model training and deployment to yield quick business results; seamless data access for workloads across regions and clouds; infinite scale that has been battle-tested at internet giant’s scale; and maximized return on investments by working with existing tech stack instead of costly specialized storage. With Alluxio Enterprise AI, enterprises can expect up to 20x faster training speed compared to commodity storage, up to 10x accelerated model serving, over 90% GPU utilization, and up to 90% lower costs for AI infrastructure.
Alluxio Enterprise AI has a distributed system architecture with decentralized metadata to eliminate bottlenecks when accessing massive numbers of small files, typical of AI workloads. This provides unlimited scalability beyond legacy architectures, regardless of file size or quantity. The distributed cache is tailored to AI workload I/O patterns, unlike traditional analytics. Finally, it supports analytics and full machine learning pipelines - from ingestion to ETL, pre-processing, training and serving.
Alluxio Enterprise AI includes the following key features:
Epic Performance for Model Training and Model Serving - Alluxio Enterprise AI offers significant performance improvements to model training and serving on an enterprise’s existing data lakes. The enhanced set of APIs for model training can deliver up to 20x performance over commodity storage. For model serving, Alluxio provides extreme concurrency and up to 10x acceleration for serving models from offline training clusters for online inference.
Intelligent Distributed Caching Tailored to I/O Patterns of AI Workloads - Alluxio Enterprise AI’s distributed caching feature enables AI engines to read and write data through the high performance Alluxio cache instead of slow data lake storage. Alluxio’s intelligent caching strategies are tailored to the I/O patterns of AI engines – large file sequential access, large file random access, and massive small file access. This optimization delivers high throughput and low latency for data-hungry GPUs. Training clusters are continuously fed data from the high-performance distributed cache, achieving over 90% GPU utilization.
Seamless Data Access for AI Workloads Across On-prem and Cloud Environments - Alluxio Enterprise AI provides a single pane of glass for enterprises to manage AI workloads across diverse infrastructure environments easily. Providing a source of truth of data for the machine learning pipeline, the product fundamentally removes the bottleneck of data lake silos in large enterprises. Sharing data between different business units and geographical locations becomes seamless with a standard data access layer via the Alluxio Enterprise AI platform.
New Distributed System Architecture, Battle-tested At Scale - Alluxio Enterprise AI builds on a new innovative decentralized architecture, DORA (Decentralized Object Repository Architecture). This architecture sets the foundation to provide infinite scale for AI workloads. It allows an AI platform to handle up to 100 billion objects with commodity storage like Amazon S3. Leveraging Alluxio’s proven expertise in distributed systems, this new architecture has addressed the ever-increasing challenges of system scalability, metadata management, high availability, and performance.
“Performance, cost optimization and GPU utilization are critical for optimizing next-generation workloads as organizations seek to scale AI throughout their businesses,” said Mike Leone, Analyst, Enterprise Strategy Group. “Alluxio has a compelling offering that can truly help data and AI teams achieve higher performance, seamless data access, and ease of management for model training and model serving.”
“We've collaborated closely with Alluxio and consider their platform essential to our data infrastructure,” said Rob Collins, Analytics Cloud Engineering Director, Aunalytics. “Aunalytics is enthusiastic about Alluxio's new distributed system for Enterprise AI, recognizing its immense potential in the ever-evolving AI industry.”
“Our in-house-trained large language model powers our Q&A application and recommendation engines, greatly enhancing user experience and engagement,” said Mengyu Hu, Software Engineer in the data platform team, Zhihu. “In our AI infrastructure, Alluxio is at the core and center. Using Alluxio as the data access layer, we’ve significantly enhanced model training performance by 3x and deployment by 10x with GPU utilization doubled. We are excited about Alluxio’s Enterprise AI and its new DORA architecture supporting access to massive small files. This offering gives us confidence in supporting AI applications facing the upcoming artificial intelligence wave.”
Deploying Alluxio in Machine Learning Pipelines
According to Gartner, data accessibility and data volume/complexity is one the top three barriers to the implementation of AI techniques within an organization. Alluxio Enterprise AI can be added to the existing AI infrastructure consisting of AI compute engines and data lake storage. Sitting in the middle of compute and storage, Alluxio can work across model training and model serving in the machine learning pipeline to achieve optimal speed and cost. For example, using PyTorch as the engine for training and serving, and Amazon S3 as the existing data lake:
Model Training: When a user is training models, the PyTorch data loader loads datasets from a virtual local path /mnt/alluxio_fuse/training_datasets. Instead of loading directly from S3, the data loader will load from the Alluxio cache instead. During training, the cached datasets will be used in multiple epochs, so the entire training speed is no longer bottlenecked by retrieving from S3. In this way, Alluxio speeds up training by shortening data loading and eliminates GPU idle time, increasing GPU utilization. After the models are trained, PyTorch writes the model files to S3 through Alluxio.
Model Serving: The latest trained models need to be deployed to the inference cluster. Multiple TorchServe instances read the model files concurrently from S3. Alluxio caches these latest model files from S3 and serves them to inference clusters with low latency. As a result, downstream AI applications can start inferencing using the most up-to-date models as soon as they are available.
Platform Integration with Existing Systems
To integrate Alluxio with the existing platform, users can deploy an Alluxio cluster between compute engines and storage systems. On the compute engine side, Alluxio integrates seamlessly with popular machine learning frameworks like PyTorch, Apache Spark, TensorFlow and Ray. Enterprises can integrate Alluxio with these compute frameworks via REST API, POSIX API or S3 API.
On the storage side, Alluxio connects with all types of filesystems or object storage in any location, whether on-premises, in the cloud, or both. Supported storage systems include Amazon S3, Google GCS, Azure Blob Storage, MinIO, Ceph, HDFS, and more.
Alluxio works on both on-premise and cloud, either bare-metal or containerized environments. Supported cloud platforms include AWS, GCP and Azure Cloud.
GlobeNewswire | September 14, 2023
Carahsoft Technology Corp., The Trusted Government IT Solutions Provider®, and member of the Oracle PartnerNetwork (OPN), today announced an agreement to serve as Oracle’s Public Sector aggregator, making the company’s complete cloud and applications portfolio for Government, Defense, Intelligence and Education available through Carahsoft’s reseller partners, NASA Solutions for Enterprise-Wide Procurement (SEWP) V and GSA Schedule contracts.
Oracle offers a diverse set of solutions, ranging from cloud infrastructure, software-as-a-service (SaaS) applications, hardware, and construction and engineering technology. Across these capabilities, Oracle enables the connectivity of project teams, empowerment of decision makers through data, and simplification of multicloud deployment across agencies. The partnership can help accelerate IT infrastructure modernization, and help improve innovation with secure, cost-effective, and FedRAMP-compliant cloud infrastructure.
Oracle’s cloud infrastructure and applications offerings provide customers with a variety of solutions, including:
Oracle Cloud Infrastructure (OCI) enables customers to build and run a wide range of applications in a scalable, secure, highly available, and high-performance environment. From application development and business analytics to data management, integration, AI, and cloud-native services, OCI delivers high security, performance, and cost savings. In addition, with multicloud, hybrid cloud, public cloud, and dedicated cloud options, OCI’s distributed cloud delivers the benefits of cloud with greater control over data residency, locality, and authority, even across multiple clouds.
Oracle Fusion Cloud Applications Suite enables organizations to take advantage of the cloud to break down organizational silos, standardize processes, and manage financial, supply chain, HR, and customer experience data on a secure integrated SaaS platform.
Oracle Construction and Engineering’s Smart Construction Platform helps teams work together, turn data into intelligence, and coordinate resources for smooth project planning, delivery, and operations.
“As the Public Sector continues to prioritize modernization and adoption of a multicloud model, Oracle is focused on interoperability and ideally positioned to help government customers innovate faster to better support their missions,” said Kyle Clement, Senior Manager, NACT Government Alliances and Channels, Oracle. “We are excited to partner with Carahsoft to bring Oracle solutions to more Public Sector customers and to continue to drive success for government agencies and the citizens they serve.”
Carahsoft will serve as the aggregator for the partnership, enabling Oracle to access Carahsoft’s portfolio of contract vehicles and extensive Public Sector ecosystem of channel partners. The Carahsoft team has significant expertise in partnering with the Public Sector to advance multi-cloud strategies while ensuring ease of procurement.
“We are excited to leverage our dedicated sales and marketing teams in support of this new partnership with Oracle,” said Lacey Wean, Sales Director who leads the Oracle Team at Carahsoft. “By including Oracle on our contract vehicles, Carahsoft and our partners are able to expand the reach of their solutions for cloud, cybersecurity, data management and more to Government agencies.”
Oracle's solutions are available through Carahsoft’s SEWP V contracts NNG15SC03B and Carahsoft’s GSA Schedule No. 47QSWA18D008F. For more information about the offerings available through this new partnership, visit https://www.carahsoft.com/oracle. To request a demo or speak to a representative, contact the contact the Oracle team at Carahsoft at (855) 618-3114 or Oracle@carahsoft.com.
About Oracle PartnerNetwork
Oracle PartnerNetwork (OPN) is Oracle’s partner program designed to enable partners to accelerate the transition to cloud and drive superior customer business outcomes. The OPN program allows partners to engage with Oracle through track(s) aligned to how they go to market: Cloud Build for partners that provide products or services built on or integrated with Oracle Cloud; Cloud Sell for partners that resell Oracle Cloud technology; Cloud Service for partners that implement, deploy and manage Oracle Cloud Services; and License & Hardware for partners that build, service or sell Oracle software licenses or hardware products. Customers can expedite their business objectives with OPN partners who have achieved Expertise in a product family or cloud service. To learn more visit: http://www.oracle.com/partnernetwork
Carahsoft Technology Corp. is The Trusted Government IT Solutions Provider®, supporting Public Sector organizations across Federal, State and Local Government agencies and Education and Healthcare markets. As the Master Government Aggregator® for our vendor partners, we deliver solutions for Geospatial, Cybersecurity, MultiCloud, DevSecOps, Big Data, Artificial Intelligence, Open Source, Customer Experience and more. Working with resellers, systems integrators and consultants, our sales and marketing teams provide industry leading IT products, services and training through hundreds of contract vehicles. Visit us at www.carahsoft.com.
Business Wire | September 26, 2023
Cloudflare, Inc. (NYSE: NET), the security, performance, and reliability company helping to build a better Internet, today shared a new independent report published by Analysys Mason that shows switching enterprise network services from on premises devices to Cloudflare’s cloud-based services can cut related carbon emissions up to 78% for very large businesses to up to 96% for small businesses. The report is one of the first of its kind to calculate potential emissions savings achieved by replacing enterprise network and security hardware boxes with more efficient cloud services.
Global Internet usage accounts for 3.7% of global CO2 emissions, about equal to the CO2 emissions of all air traffic around the world. The Internet needs to reduce its overall energy consumption, especially as regulators continue to implement the Paris Climate Accord, including plans to transition to a zero emissions economy. The European Climate Law requires that Europe’s economy and society become climate-neutral by 2050, with a target of reducing net GHG emissions by at least 55% by 2030, compared to 1990 levels. Regulators in the United States and the European Union, among others, have also announced plans to require companies to disclose climate-related information including carbon emissions resulting from their operations and supply chains, as well as climate related risks and opportunities. Finally, among the Fortune Global 500, 63% of companies now set 2050 targets for emissions reductions. Companies large and small will increasingly be looking to reduce carbon throughout their supply chains, particularly their IT infrastructure.
“The best way to reduce your IT infrastructure’s carbon footprint is easy: move to the cloud,” said Matthew Prince, CEO and co-founder, Cloudflare. “At Cloudflare, we’ve built one of the world’s most efficient networks, getting the most out of every watt of energy and every one of our servers. That’s why, with Cloudflare, companies can help hit their sustainability goals without sacrificing security, speed, performance, or innovation.”
The Analysys Mason study found that switching enterprise network services from on premises devices to Cloudflare services can cut related carbon emissions up to 96%, depending on the current network footprint. The greatest reduction comes from consolidating services, which improves carbon efficiency by increasing the utilization of servers that are providing multiple network functions. On premises devices are designed to host multiple workloads and consume power constantly, but are only used for part of the day and part of the week. Cloud infrastructure is shared by millions of customers, often all over the world. As a result, cloud providers are able to achieve economies of scale that result in less downtime, less waste, and lower emissions. Furthermore, the Analysys Mason study found that there are additional gains due to the high Power Usage Effectiveness of cloud data centres, and differences in the carbon intensity of generation in the local electricity grid.
“Happy Cog is a full-service digital agency that designs, builds, and markets experiences that engage our clients and their audiences. We’ve relied on Cloudflare for many of those websites and apps because it's secure, reliable, fast, and affordable – but also aligns with many of our clients’ sustainability roadmaps and goals,” said Matt Weinberg, Co-Founder and President of Technology at Happy Cog. “Switching our clients from their previous on premises or other constant-usage infrastructure to Cloudflare's network and services has let them be greener, more efficient, and more cost effective. It's ideal when you can offer your clients a solution that covers all their needs and provides a delightful experience now, without having to compromise on their longer term priorities.”
Cloudflare, Inc. (www.cloudflare.com / @cloudflare) is on a mission to help build a better Internet. Cloudflare’s suite of products protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare have all web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks. Cloudflare was awarded by Reuters Events for Global Responsible Business in 2020, named to Fast Company's Most Innovative Companies in 2021, and ranked among Newsweek's Top 100 Most Loved Workplaces in 2022.