AI stack attack: Navigating the generative tech maze

We are seeking your input! Participate in our brief AI survey to share your thoughts on the current state of AI, how you are implementing it, and your expectations for the future. Find out more


In just a few months, the generative AI technology stack has undergone a significant transformation. Menlo Ventures’ January 2024 market map depicted a neat four-layer framework. By late May, Sapphire Ventures’ visualization expanded into a network of over 200 companies across multiple categories. This rapid growth highlights the fast pace of innovation and the challenges facing IT decision-makers.

Technical considerations intersect with strategic concerns. Data privacy and impending AI regulations are significant, while talent shortages force companies to balance in-house development with outsourced expertise. The need to innovate competes with cost control pressures.

In this fast-paced technological landscape, adaptability emerges as a key advantage. Today’s cutting-edge solution could become outdated by tomorrow’s breakthrough. IT decision-makers must develop a flexible vision that can evolve with this dynamic environment while delivering tangible value to their organizations.


Countdown to VB Transform 2024

Join industry leaders in San Francisco from July 9 to 11 for our premier AI event. Connect with peers, explore Generative AI opportunities and challenges, and learn how to integrate AI applications into your sector. Register Now


Credit: Sapphire Ventures

The push towards end-to-end solutions

As enterprises navigate the complexities of generative AI, many are moving towards comprehensive, end-to-end solutions. This shift reflects a desire to simplify AI infrastructure and streamline operations in a complex tech landscape.

When Intuit faced the challenge of integrating generative AI across its ecosystem, it opted to create GenOS, a comprehensive generative AI operating system, instead of relying on existing platform capabilities. Ashok Srivastava, Intuit’s Chief Data Officer, explains that this decision aims to accelerate innovation while ensuring consistency.

Similarly, Databricks has enhanced its AI deployment capabilities with new features to simplify model serving. The company’s Model Serving and Feature Serving tools represent a move towards a more integrated AI infrastructure.

These new offerings enable data scientists to deploy models with less engineering support, potentially streamlining the path from development to production. Maria Vechtomova, author of Marvelous MLOps, highlights the industry-wide need for simplification in model serving architectures.

Databricks’ platform now supports various serving architectures, catering to different use cases from e-commerce recommendations to fraud detection. Craig Wiley, Databricks’ Senior Director of Product for AI/ML, aims to provide “a truly complete end-to-end data and AI stack.”

While some advocate for a single-vendor approach, Red Hat’s Steven Huels emphasizes complementary solutions that can integrate with existing systems. The push towards end-to-end solutions signifies a maturation of the generative AI landscape as enterprises look for efficient ways to scale AI initiatives.

Data quality and governance take center stage

As generative AI applications proliferate in enterprise settings, data quality and governance have become top concerns. The effectiveness of AI models relies on the quality of training data, underscoring the importance of robust data management.

Governance, ensuring ethical and secure data use in compliance with regulations, has become a priority. Red Hat’s Huels anticipates a growing emphasis on governance as AI systems play a greater role in critical business decisions.

Databricks has integrated governance into its platform, ensuring a continuous governance system from data ingestion to generative AI prompts and responses.

The rise of semantic layers and data fabrics

Semantic layers and data fabrics are gaining prominence as quality data sources become essential. These technologies enhance data infrastructure, enabling AI systems to better understand and leverage enterprise data for new possibilities.

Startups like Illumex are developing semantic data fabrics that create dynamic, context-aware data interactions. Larger enterprises like Intuit adopt a product-oriented approach to data management, setting high standards for data quality, performance, and operations.

The shift towards semantic layers and data fabrics signals a new era in data infrastructure, promising more effective use of enterprise data by AI systems. Implementing these technologies requires significant investment in technology and expertise to align with existing data infrastructure and AI initiatives.

Specialized solutions in a consolidated landscape

While end-to-end platforms are on the rise, specialized solutions addressing specific aspects of the AI stack continue to emerge. These niche offerings complement broader platforms, addressing complex challenges that they may overlook.

Illumex focuses on creating a generative semantic fabric, bridging the gap between data and business logic. These specialized solutions enhance specific capabilities and often forge partnerships with end-to-end providers to bolster offerings.

The persistent emergence of specialized solutions highlights ongoing innovation in addressing specific AI challenges. IT decision-makers must assess where specialized tools can offer advantages over generalized solutions.

Balancing open-source and proprietary solutions

The generative AI landscape features a dynamic interplay between open-source and proprietary solutions. Enterprises must navigate this terrain, considering the benefits and drawbacks of each approach.

Red Hat’s entry into the generative AI space with Red Hat Enterprise Linux (RHEL) AI aims to democratize access to large language models while upholding open-source principles. Proprietary solutions like Databricks offer integrated experiences with support for open-source models.

The ideal balance between open-source and proprietary solutions depends on an organization’s needs, resources, and risk tolerance. As the AI landscape evolves, effectively integrating and managing both types of solutions can be a competitive advantage.

Integration with existing enterprise systems

Integrating generative AI capabilities with existing systems and processes is a critical challenge for many enterprises. Success hinges on a strong foundation of data and processing capabilities to build advanced AI capabilities upon.

Connecting AI systems with diverse data sources is a common challenge, but solutions like Illumex can work with existing data infrastructures without major restructuring. Organizations must also consider how AI will interact with existing business processes, decision-making frameworks, and security measures.

The radical future of generative computing

The evolving generative AI landscape points towards a transformative moment in enterprise technology. Visionaries like Andrej Karpathy foresee a radical future where a single neural network replaces all classical software, reshaping the nature of computing itself.

Choices made today in building AI infrastructure will pave the way for future innovations. Adaptability, scalability, and a readiness for paradigm shifts are crucial for success in a rapidly evolving tech environment.

For more insights on navigating the tech landscape, join us at VentureBeat Transform in San Francisco this week.