Implementing Generative AI: A Pipeline Architecture

NeuroCortex.AI
19 min readApr 5, 2024

--

Here are links for Part 1 and Part 2 of the blog series:

  1. Part 1: Unleashing Creativity: An Introduction to Generative AI | by NeuroCortex.AI | Feb, 2024 | Medium
  2. Part2: Unraveling the Magic: A Deep Dive into Generative AI | by NeuroCortex.AI | Mar, 2024 | Medium
A supercomputer imagined by DALLE-3

Imagine having a tool that can create entirely new and original content, from text and music to images and even videos. That’s the magic of generative AI. It utilizes machine learning models to analyze massive datasets of existing content, learning the underlying patterns and relationships. Then, it leverages this knowledge to produce fresh, creative outputs. The architecture of generative AI for enterprises is complex and integrates multiple components, such as data processing, machine learning models and feedback loops. The system is designed to generate new, original content based on input data or rules.

In an enterprise setting, the enterprise generative AI architecture can be implemented in various ways. For example, it can be used to automate the process of creating product descriptions or a marketing copy, saving time and cutting costs. It can also be used to generate data analysis reports, which can help companies make better business decisions.

Now to implement large scale AI projects we need to implement ML pipeline which is essential in production and maintenance. Lets learn about ML pipeline and how it can be altered to work with generative AI. Buckle up, this is going to be a long read !!

What is a Machine Learning pipeline?

A machine learning pipeline is a structured sequence of interconnected data processing and modeling steps designed to automate, standardize, and streamline the process of building, training, evaluating, and deploying machine learning models. It orchestrates the flow of data into and output from a machine learning model or a set of multiple models, encompassing raw data input, features, outputs, the machine learning model and its parameters, and prediction outputs. The design and implementation of a machine learning pipeline are fundamental in enterprise AI software applications, significantly influencing performance and effectiveness by managing the complexity of the end-to-end machine learning process efficiently.

Such pipelines collectively enhance the effectiveness, reliability, and scalability of machine learning processes, driving advancements in various industries and enabling data-driven decision-making.

Summary of Generative AI Architecture: While traditional machine learning pipelines focus on extracting features for optimal model performance, generative AI takes a different approach. Below is a high-level description, and the actual architecture might vary based on the specific type of generative AI task (e.g., GANs for images, transformers for text). Visualizing this architecture can help understand the flow of information and interactions between different components.

Generative AI architecture diagram

A simple generative AI architecture that you can visualize:

  1. Input Layer: Represents the data or initial seed for generation, depending on the type of task (e.g., images, text).
  2. Generator: The core component responsible for creating new content. It may use various architectures like fully connected layers, convolutional layers, or recurrent layers.
  3. Discriminator: Evaluates the generated content, distinguishing between real and generated data. It provides feedback to the generator for improvement.
  4. Adversarial Loop: Feedback loop between the generator and discriminator. The generator tries to improve its output based on the feedback from the discriminator, creating a competitive learning process.
  5. Latent Space: A representation where the generator learns to map the input data, aiding in generating diverse outputs.
  6. Autoencoders (optional): If using autoencoders, there will be an encoding and decoding component to compress and reconstruct input data.
  7. Recurrent Neural Networks (RNNs) or Transformer (optional): For tasks involving sequential data, an RNN or Transformer architecture might be integrated.
  8. Attention Mechanisms (optional): Enhances the model’s ability to focus on specific parts of the input, often used in transformer architectures.
  9. Output Layer: Represents the final generated content, whether it’s an image, text, or other data types.
  10. Training Data and Fine-Tuning: Input data used to train the model, and fine-tuning steps to adapt the model to specific tasks or domains.
Generative AI Life Cycle
Gen AI Lifecycle merged with prompt engineering
A complete framework for Job Search portal and matching of applicants and relevant jobs

Lets have a look at what major components are required to make generative AI work:

  1. Training Data: Generative models require large amounts of training data to learn the underlying patterns. This data can come from various sources, such as images, text corpora, audio recordings, or any other type of structured or unstructured data.
  2. Loss Functions: The choice of loss function is crucial in training generative models. For GANs, the discriminator’s loss function typically measures how well it can distinguish between real and generated data, while the generator’s loss function measures how well it can fool the discriminator. In VAEs, the loss function typically consists of a reconstruction loss and a regularization term to encourage the learned latent space to follow a specific distribution.
  3. Evaluation Metrics: Evaluating the quality of generated data is challenging. Common metrics include perceptual similarity, diversity, and novelty. For example, in image generation, metrics like Inception Score and Frechet Inception Distance (FID) are commonly used.
  4. Mode Collapse and Overfitting: Generative models are prone to mode collapse, where the generator learns to produce a limited set of outputs, ignoring the diversity of the training data. Overfitting can also occur, where the model memorizes the training data without learning its underlying distribution.
  5. Conditional Generation: Some generative models can be conditioned on additional information, such as class labels or text descriptions. Conditional GANs and text-to-image synthesis models are examples of this approach.
  6. Training Data and Fine-Tuning:
Language models learn through a combination of pre-training, fine-tuning, and in-context learning.

Training data and fine-tuning are crucial aspects of machine learning, especially in the context of deep learning models. Here’s a breakdown of each:

  1. Training Data:
  • Definition: Training data refers to the labeled or unlabeled examples used to train a machine learning model. It consists of input data points along with their corresponding target labels or outcomes.
  • Importance: The quality and quantity of training data significantly impact the performance of a machine learning model. Sufficient and diverse training data enable the model to learn robust patterns and generalize well to unseen examples.
  • Data Sources & Collection: Training data can be collected from various sources, including databases, APIs, sensor data, user interactions, web scraping, and manual labeling. Training data for generative AI models can come from various sources, including image datasets (e.g., ImageNet), text corpora (e.g., Wikipedia articles), audio recordings, or any other type of structured or unstructured data relevant to the task.
  • Preprocessing: Before training, training data often undergo preprocessing steps such as cleaning, normalization, feature engineering, and splitting into training, validation, and test sets. Preprocessing steps such as normalization, resizing, or tokenization may be necessary depending on the type of data and the specific requirements of the model architecture.
  • Balancing: In classification tasks, imbalanced class distributions in the training data can lead to biased models. Techniques such as oversampling, undersampling, or using class weights can help address this issue. Ensuring a balanced distribution of classes or categories in the training data is important to prevent the model from learning biases or producing unrepresentative outputs.
  • Augmentation: Data augmentation techniques involve creating additional training examples by applying transformations such as rotation, translation, scaling, or flipping to the existing data. This helps increase the diversity of the training data and improve model generalization. Data augmentation techniques can be applied to increase the diversity of the training data and improve the robustness of the generative model. For example, in image generation tasks, augmentation techniques such as rotation, translation, or cropping may be used.
  • Quality and Quantity: The quality and quantity of training data are crucial for generative AI models. High-quality and diverse datasets enable the model to learn meaningful patterns and generate realistic outputs.

2. Fine-Tuning:

  • Definition: Fine-tuning refers to the process of taking a pre-trained model (usually trained on a large dataset) and further training it on a smaller, domain-specific dataset to adapt it to a specific task or domain.
  • Transfer Learning: Fine-tuning leverages transfer learning, where knowledge learned from one task or domain is transferred to another related task or domain. This approach is particularly useful when labeled training data for the target task is limited. Fine-tuning pre-trained generative models is a common practice, especially when working with limited training data. Pre-trained models trained on large datasets (e.g., ImageNet for image data) can be fine-tuned on smaller, domain-specific datasets to adapt them to specific tasks or domains.
  • Steps: The typical steps involved in fine-tuning a pre-trained model include:
  • Loading the pre-trained model architecture and weights.
  • Replacing or adding a few output layers specific to the target task.
  • Training the model on the target dataset while keeping the pre-trained weights fixed or allowing them to be updated with a smaller learning rate.
  • Hyperparameter Tuning: During fine-tuning, hyperparameters such as learning rate, batch size, optimizer choice, and regularization techniques may need to be adjusted to achieve optimal performance on the target task. Fine-tuning involves tuning hyperparameters such as learning rate, batch size, number of layers, and network architecture-specific parameters to achieve optimal performance on the target task.
  • Evaluation: After fine-tuning, the performance of the adapted model is evaluated on a separate validation or test set to assess its effectiveness for the target task. This evaluation helps determine whether further adjustments or iterations are necessary. After fine-tuning, the performance of the generative model is evaluated using metrics specific to the task, such as image quality metrics (e.g., Inception Score, Fréchet Inception Distance), text generation metrics (e.g., perplexity), or audio generation metrics (e.g., signal-to-noise ratio).
  • Architecture Selection: Choosing the appropriate architecture for the generative model is crucial. Depending on the task and type of data, different architectures such as GANs, VAEs, or autoregressive models may be more suitable.
  • Regularization: Regularization techniques such as dropout, weight decay, or batch normalization may be applied during fine-tuning to prevent overfitting and improve the generalization ability of the generative model.

Fine-tuning is commonly used in various domains, including computer vision, natural language processing, and audio processing, where pre-trained models such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), and transformers can be adapted to specific tasks with limited labeled data. Training data and fine-tuning are essential components when working with generative AI models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs).

Overall, both training data and fine-tuning are critical for training high-quality generative AI models that can generate realistic and diverse outputs across various domains, including image generation, text generation, music composition, and more.

Architecture layer of Gen AI system
Layers of Gen AI based solution
Pillars of Gen AI Solution architecture
The overall pipeline for building Gen AI based AI enablement for the markets

Generative AI Pipeline Architecture

Here’s a simplified description of a Generative AI Pipeline Architecture: The architecture of generative AI involves several key components working together to create new and original content, like text, images, code, or even music. Here’s a breakdown of the essential elements:

  1. Data Collection: Gather diverse and representative datasets for training. The quality and variety of data significantly impact the model’s ability to generate creative outputs.
  2. Data Preprocessing: Clean, normalize, and preprocess the data to ensure consistency. This step is crucial for removing noise and irrelevant information.
  3. Training Data Split: Divide the dataset into training, validation, and testing sets. The training set is used to train the model, the validation set helps fine-tune parameters, and the testing set assesses the model’s performance.
  4. Model Architecture Selection: Choose the appropriate generative AI architecture based on the task. It could be a Generative Adversarial Network (GAN), Variational Autoencoder (VAE), or other architectures tailored to the specific domain.
  5. Generator Network: The generator network takes input data and transforms it to generate new content. It might consist of various layers, such as fully connected layers, convolutional layers, or recurrent layers.
  6. Discriminator Network: The discriminator network evaluates the generated content, distinguishing between real and generated data. This adversarial process refines the generator’s ability to produce realistic outputs.
  7. Training Process: Iteratively train the model using the training data. The generator and discriminator networks engage in an adversarial loop, with feedback mechanisms optimizing their performance over time.
  8. Validation and Fine-Tuning: Validate the model’s performance on the validation set. Fine-tune hyperparameters or adjust the model architecture based on feedback to improve overall performance.
  9. Testing: Evaluate the final model on the testing set to assess its generalization and performance on unseen data.
  10. Inference: Deploy the trained model for inference on new data. This is where the generative AI system generates content based on input.
  11. Post-Processing (Optional): Depending on the task, post-process the generated content to enhance quality or meet specific criteria.
  12. User Interaction (Optional): If the generative AI system is designed for user interaction, integrate interfaces for users to input preferences or guide the generation process.
  13. Feedback Loop (Optional): Implement mechanisms for continuous improvement. Collect user feedback or monitor performance metrics to adapt and enhance the model over time.
  14. Prompt Engineering: This is where the magic truly happens! Unlike traditional feature engineering, generative AI relies on prompts. Prompts are essentially instructions that guide the model towards the specific type of creative output you desire. Crafting effective prompts is an art form, requiring an understanding of the model’s capabilities and the nuances of the desired outcome.

This pipeline architecture provides a high-level overview of the stages involved in developing and deploying a generative AI model, emphasizing the iterative nature of training and refinement. Actual architectures may vary based on specific use cases and requirements.

  1. Data Processing:
  • Data Collection: This involves gathering data from various sources relevant to the desired output. For instance, if you’re training a model to generate realistic images of cats, you’d need a massive dataset of cat pictures.
The lifecycle of Data collection process

Data Collection for Generative AI

  • Data Preparation: The collected data needs to be cleaned, normalized, and formatted to be suitable for the AI model. This might involve removing duplicates, correcting errors, and ensuring consistency in formats and structures.
Data preparation pipeline for Data scientists

Data Preparation for Generative AI

  • Feature Engineering: Extracting relevant features from the data helps the model identify patterns and relationships. This could involve identifying key characteristics of cat images, like fur color, breed, or posture.

Feature Engineering for Generative AI

  1. Generative Model: Model Selection: Different types of generative models exist, each suited for specific tasks. Popular choices include Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and Transformers.
Step by step process for building and deployment of Gen AI solution

Feedback Loop for Generative AI

Understanding how AI feedback loops work is significant to unlock the whole potential of AI development. Let’s explore the various stages of AI feedback loops below.

  1. Feedback Gathering: Gather relevant model outcomes for evaluation. Typically, users give their feedback on the model outcome, which is then used for retraining. Or it can be external data from the web curated to fine-tune system performance.
  2. Model Re-training: Using the gathered information, the AI system is re-trained to make better predictions, provide answers, or carry out particular activities by refining the model parameters or weights.
  3. Feedback Integration & Testing: After retraining, the model is tested and evaluated again. At this stage, feedback from Subject Matter Experts (SMEs) is also included for highlighting problems beyond data.
  4. Deployment: The model is redeployed after verifying changes. At this stage, the model should report better performance on new real-world data, resulting in an improved user experience.
  5. Monitoring: The model is monitored continuously using metrics to identify potential deterioration, like drift. And the feedback cycle continues.
Production-deployed AI models need a robust and continuous performance evaluation mechanism. This is where an AI feedback loop can be applied to ensure consistent model performance.
A high-level illustration of feedback mechanism in AI models

Additional Considerations:

  • Infrastructure: Generative AI models often require powerful computing resources, like GPUs or specialized AI hardware, to handle the complex calculations involved.
  • Tools and Frameworks: Various tools and frameworks are available to facilitate generative AI development, such as TensorFlow, PyTorch, and Transformers.
  • Ethical and Responsible AI: It’s crucial to consider the potential biases and limitations of generative AI models and implement safeguards to ensure their ethical and responsible use.

High-Level Overview:

  • Data Processing: This section depicts the flow of data, starting with Collection from various sources like text files, images, or code. Next, it shows the data going through Preparation where it’s cleaned, normalized, and formatted. Finally, Feature Engineering extracts relevant characteristics for analysis.
  • Generative Model: This box represents the core of the architecture, where different model types like GANs, VAEs, or Transformers are housed. It highlights the Training process where the model learns from the prepared data and the subsequent Fine-tuning stage for further refinement.
  • Generation and Feedback: This section showcases the model’s ability to generate new content based on User Prompts or Instructions. It also emphasizes the importance of Feedback, illustrating how user input feeds back into the system to improve model performance.
  • Deployment and Integration: This final section represents the integration of the trained model into real-world applications like generating product descriptions, writing creative content, or designing innovative products.
This architecture represents an amalgamation of Gen AI with enterprise. The users can connect to the applications either through an IoT device or an application running in a public cloud or the front-end application running on-prem. The user may be the customer representative talking to the customers, collecting the data and feeding it into the system.

Deeper Dives:

  • Specific Model Architecture: Within the Generative Model box, you can find optional subsections diving deeper into the details of specific model types like GANs or VAEs. These subsections would visually represent the key components like encoder, decoder, discriminator, and their interactions.
  • Data Processing Pipeline: This could be a separate, zoomed-in diagram illustrating the different stages of data preparation in more detail. It could show specific techniques used for cleaning, normalization, and feature engineering tailored to the type of data.
  • Training and Feedback Loop: Another optional subsection could delve deeper into the iterative training process. It could depict the model generating outputs, receiving feedback, and then using that feedback to adjust its parameters and improve its performance over time.
  • Data Collection: How and from where is the data gathered? Examples might be scraping websites, using APIs, or reading files from storage.
  • Data Cleaning: This could involve removing duplicates, fixing errors, and ensuring consistency in format and structure.
  • Normalization: Depending on the data type, different normalization techniques might be applied, such as scaling numerical data or tokenizing text.
  • Feature Engineering: Extracting relevant features specific to the type of data and the desired generative task. For example, identifying key characteristics of images or extracting keywords from text.
  • Data Augmentation (optional): This step can artificially increase the size and diversity of the data to improve model performance, especially for limited datasets. Techniques like image flipping, text paraphrasing, or code mutation can be used.
  • Validation and Splitting: Splitting the processed data into training, validation, and test sets for model training and evaluation.
Generative AI Pipeline Architecture for Enterprises

However, challenges persist, including ensuring robustness, interpretability, and addressing ethical considerations in their deployment. Despite these challenges, the future of generative AI architecture is bright. Continued research and development efforts are expected to push the boundaries of creativity, enabling AI systems to produce increasingly sophisticated and contextually relevant outputs. As these technologies continue to mature, they hold immense promise for revolutionizing creative industries, scientific research, and beyond, ushering in a new era of human-AI collaboration and innovation.

Key Differences from Traditional ML Pipelines

  • Focus on Prompt Engineering: Generative AI prioritizes crafting effective prompts to steer the model towards the desired creative direction.
  • Emphasis on Large Datasets: Generative models often require vast amounts of data to learn the intricate patterns necessary for effective content creation.
  • Iterative Process: Refining prompts, data selection, and model training are often cyclical processes in generative AI development.

A generative AI pipeline is all about feeding a machine learning model the knowledge to create new things. Here’s the gist:

  1. Data Prep: Collect and clean data relevant to what you want the AI to generate (text for writing, images for pictures, etc.).
  2. Model Up and Running: Pick a generative AI model (like GANs) and train it on the data.
  3. Prompting the Magic: Craft clear instructions (prompts) to tell the model what kind of creative output you desire.
  4. Fine-Tuning (Optional): Train the model on even more specific data to get it to generate exactly what you need.
  5. Evaluation Loop: Check the generated content, tweak prompts or data if needed, and keep iterating for better results.

Generative AI pipeline for a startup

Infographic about how Gen AI solutions would be applied to our modern world

Building a generative AI pipeline for a startup involves several steps, ranging from data collection and preprocessing to model development and deployment. Here’s a general guide to help you set up a generative AI pipeline. An example, let’s consider a startup that aims to develop a generative AI pipeline for creating art. The goal is to generate unique and visually appealing images using a GAN (Generative Adversarial Network) architecture. Each of the below points we will explain taking the example given.

  1. Define the Problem: Clearly define the problem you want to solve with generative AI. This could be image synthesis, text generation, or any other creative task. E.g. Generate unique and visually appealing artwork using a generative AI model.
  2. Data Collection and Preprocessing: Gather a high-quality dataset relevant to your problem. Preprocess the data to ensure it’s in a format suitable for training. This may involve cleaning, normalizing, and augmenting the data. E.g. Collect a diverse dataset of high-resolution artworks. Preprocess the images, ensuring they are in a standardized format and resolution. Augment the dataset with rotations, flips, and color adjustments.
  3. Choose a Generative Model: Select a generative model architecture suitable for your task. Popular choices include: GANs (Generative Adversarial Networks): Excellent for image synthesis. Variational Autoencoders (VAEs): Useful for generating structured data and images. Transformers: Effective for natural language processing tasks. E.g. Choose a GAN architecture suitable for image synthesis, such as DCGAN (Deep Convolutional GAN) or StyleGAN.
  4. Model Development: Implement and train your chosen generative model. Fine-tune hyperparameters and experiment with different architectures. Use transfer learning if applicable, leveraging pre-trained models to accelerate training on your specific task. E.g. Implement and train the chosen GAN architecture using the preprocessed artwork dataset. Fine-tune hyperparameters, experiment with different architectures, and leverage transfer learning from pre-trained models if available.
  5. Validation and Testing: Split your dataset into training, validation, and test sets. Validate your model’s performance using the validation set and fine-tune accordingly. Evaluate the final model on the test set to ensure generalization. E.g. Split the dataset into training, validation, and test sets. Validate the model’s performance using the validation set, adjusting as needed. Evaluate the final model on the test set to ensure it generates high-quality and diverse artwork.
  6. Optimization and Scaling: Optimize your model for efficiency and scalability, especially if you plan to deploy it in a production environment. E.g. Optimize the GAN model for efficiency, considering factors like batch size, learning rate, and convergence speed. Scale the model if necessary for handling a larger volume of data.
  7. Deployment: Deploy your generative model as a service. This could be on the cloud, on-premises, or on edge devices, depending on your requirements. Consider using containerization tools like Docker for seamless deployment. E.g. Deploy the trained GAN model as a service using cloud infrastructure. Consider containerization tools like Docker for easy deployment and management.
  8. Monitoring and Maintenance: Implement monitoring to keep track of your model’s performance over time. Regularly update your model with new data to ensure it stays relevant and continues to generate high-quality outputs. E.g. Implement monitoring to track the model’s performance and detect issues. Regularly update the model with new artworks to keep the generated content fresh and appealing.
  9. User Interface (Optional): If applicable, design a user interface or API to interact with your generative AI model. This is especially important if you’re creating a product or service for end-users. E.g. Develop a user-friendly interface where users can interact with the generative AI to customize and generate unique artworks.
  10. Security and Privacy: Implement security measures to protect your generative AI pipeline and user data. Ensure compliance with privacy regulations and best practices. E.g. Implement security measures to protect the generative AI pipeline and user data. Ensure compliance with privacy regulations and best practices.
  11. Feedback Loop: Establish a feedback loop to continuously improve your generative model based on user feedback and changing requirements. E.g. Establish a feedback loop for users to provide input on generated artworks, allowing the model to adapt and improve over time.
  12. Documentation: Document your generative AI pipeline, including data sources, preprocessing steps, model architecture, and deployment procedures. This documentation is crucial for collaboration and future development. E.g. Document the entire generative AI pipeline, including data sources, preprocessing steps, GAN architecture, and deployment procedures. This documentation is crucial for collaboration and future development.

This example outlines the key steps for developing a generative AI pipeline for artwork generation. Keep in mind that the specifics may vary based on the startup’s goals, resources, and the nature of the generative task.

Conclusion

Generative AI is poised to revolutionize the architectural landscape. By leveraging its ability to process vast amounts of data and generate creative solutions, architects can:

  • Boost Efficiency: Automate repetitive tasks, explore design variations rapidly, and optimize for sustainability.
  • Enhance Creativity: Spark new ideas, explore unconventional approaches, and visualize concepts with stunning detail.
  • Empower Collaboration: Facilitate communication with clients and stakeholders through interactive design tools.

However, we must remember that generative AI is a tool, not a replacement for human expertise. The architect’s role will evolve, focusing on:

  • Guiding the AI: Formulating clear prompts, interpreting results, and selecting the most promising solutions.
  • Ensuring Ethical Considerations: Mitigating bias, promoting inclusivity, and ensuring responsible design practices.

As generative AI continues to develop, the possibilities are boundless. From optimizing urban planning to crafting personalized living spaces, this technology holds the key to a future where buildings are not just functional, but truly inspiring.

Let’s embrace this new era of architectural design and co-create a future that is innovative, sustainable, and human-centered.

References

1. Generative AI Solution Architecture for Complex Enterprises (scottlogic.com)

2. The Generative AI Life-cycle. The common AI/ML Lifecycle consists of… | by Ali Arsanjani | Medium

3. Generative-AI Application Architecture — 2 | by Ali Khan | 𝐀𝐈 𝐦𝐨𝐧𝐤𝐬.𝐢𝐨 | Medium

4. Generative-AI-based Application Architecture — 1 | by Ali Khan | 𝐀𝐈 𝐦𝐨𝐧𝐤𝐬.𝐢𝐨 | Medium

5. Empowering Language Models: Pre-training, Fine-Tuning, and In-Context Learning | by Bijit Ghosh | Medium

6. Layers of a Generative AI Solution Architecture | LinkedIn

7. The current state of generative AI (leewayhertz.com)

8. Generative AI Solutions: Revolutionizing the Content Industry | by LeewayHertz | Product Coalition

9. Generative AI — A guide on data preparation — Itzikr’s Blog (volumes.blog)

10. Revolutionizing Predictive Models with Generative AI-Driven Feature Tickets, Wed, Mar 20, 2024 at 2:00 PM | Eventbrite

11. https://www.linkedin.com/pulse/generative-ai-end-to-end-process-blog-2-shiv-kumar

12. Making generative AI training simple and cost-efficient with PeriFlow and Azure | Microsoft for Startups Blog

13. How to transform your content creation with generative AI (contentstack.com)

14. https://www.unite.ai/the-ai-feedback-loop-maintaining-model-production-quality-in-the-age-of-ai-generated-content/

15. Hybrid Cloud & Generative AI Architecture | LinkedIn

16. Generative AI — Comprehensive Guide For Beginners (binmile.com)

17. The architecture of Generative AI for enterprises (leewayhertz.com)

--

--