Model Deployment : Exploring Modular Application Programming Interface Frameworks For Serving Model Predictions¶
- 1. Table of Contents
- 1.1 Project Background
- 1.2 Application Programming Interface (API) Development Using the FastAPI Framework
- 1.3 Application Programming Interface (API) Development Using the Flask Framework
- 1.4 Consolidated Findings
- 2. Summary
- 3. References
1. Table of Contents ¶
This project explores the modular deployment of machine learning models using Representational State Transfer (RESTful) Application Programming Interfaces (APIs), specifically comparing FastAPI and Flask frameworks in Python. Pre-trained models were loaded and integrated into the APIs, including a Stacked ensemble binary classification model for predicting lung cancer probabilities, a Cox Proportional Hazards survival prediction model for estimating heart failure survival profiles, and a Convolutional Neural Network-based image classification model for determining class categories for brain magnetic resonance images. The study objectives included understanding the similarities and differences between FastAPI and Flask in implementating and documenting RESTful API endpoints to process data preprocessing and model prediction logic, handling a variety of input types (such as structured data for class and survival probability predictions and file uploads for image classification), enabling diverse output formats (including string, float, list and base64-encoded visualization objects, depending on the model), and ensuring robust error handling and validation. All results were consolidated in a Summary presented at the end of the document.
RESTful APIs are a standardized architectural style for designing networked applications, enabling communication between clients and servers over HTTP. They use HTTP methods like GET, POST, PUT, and DELETE to perform CRUD (Create, Read, Update, Delete) operations on resources, which are typically represented in JSON or XML format. RESTful APIs are stateless, meaning each request from a client to a server must contain all the information needed to process the request, ensuring scalability and reliability. For machine learning model deployment, RESTful APIs serve as a bridge between trained models and end-users or applications, allowing models to be accessed remotely via HTTP requests. This enables real-time predictions, batch processing, and integration with web or mobile applications. RESTful APIs are particularly significant for machine learning because they provide a standardized, platform-agnostic way to serve predictions, making models accessible to a wide range of clients. They also facilitate modularity, as models can be updated or replaced without affecting the client-side application. Additionally, RESTful APIs support scalability, as they can be deployed on cloud platforms and scaled horizontally to handle increased traffic. Error handling and validation mechanisms in RESTful APIs ensure robustness, which is critical for machine learning applications where malformed inputs can lead to incorrect predictions. By encapsulating machine learning logic behind APIs, developers can abstract away the complexity of model inference, making it easier for non-technical users to interact with the models. RESTful APIs also enable versioning, allowing multiple versions of a model to coexist and be accessed independently. This is particularly useful for A/B testing or gradual rollouts of updated models. Furthermore, RESTful APIs can be secured using authentication and authorization mechanisms, ensuring that only authorized users or applications can access the model. Overall, RESTful APIs are a cornerstone of modern machine learning deployment, providing a flexible, scalable, and secure way to serve predictions in production environments.
FastAPI is a modern, high-performance web framework for building APIs with Python, specifically designed for speed and ease of use. It is built on top of Starlette for web handling and Pydantic for data validation, making it one of the fastest Python frameworks available. FastAPI leverages Python type hints to automatically generate OpenAPI (Swagger) documentation, which simplifies API testing and debugging. Its asynchronous capabilities, powered by Python’s async and await keywords, make it ideal for handling high-concurrency workloads, such as serving machine learning models to multiple clients simultaneously. FastAPI’s built-in dependency injection system allows for modular and reusable code, which is particularly useful for complex machine learning pipelines. The framework also supports WebSocket communication, enabling real-time interactions, though this is less commonly used in machine learning deployments. One of FastAPI’s key strengths is its automatic data validation using Pydantic, which ensures that inputs to machine learning models are correctly formatted and reduces the risk of errors during inference. However, FastAPI’s reliance on asynchronous programming can be a double-edged sword; while it improves performance, it may require developers to have a deeper understanding of asynchronous programming concepts. Additionally, FastAPI’s ecosystem, while growing, is still smaller than that of more established frameworks like Flask, which can limit the availability of third-party plugins and extensions. Despite these limitations, FastAPI is widely regarded as an excellent choice for machine learning model deployment due to its speed, scalability, and developer-friendly features. Its ability to automatically generate API documentation and validate inputs makes it particularly well-suited for production environments where reliability and maintainability are critical.
Flask is a lightweight and flexible web framework for Python, designed to be simple and easy to use while providing the essentials for building web applications and APIs. It follows the WSGI (Web Server Gateway Interface) standard and is often described as a "micro-framework" because it provides only the core components needed for web development, such as routing and request handling, while leaving other functionalities to extensions. Flask’s simplicity and minimalistic design make it highly customizable, allowing developers to tailor it to specific use cases, including machine learning model deployment. Flask’s synchronous nature makes it easier to understand and use for developers who are not familiar with asynchronous programming, though this can limit its performance in high-concurrency scenarios. For machine learning deployments, Flask’s simplicity is both a strength and a weakness; while it is easy to set up and deploy, it lacks built-in features like data validation and automatic documentation, which must be implemented manually or through extensions like Flasgger. Flask’s extensive ecosystem of extensions, such as Flask-RESTful for building RESTful APIs and Flask-SQLAlchemy for database integration, provides additional functionality but can also introduce complexity. One of Flask’s key strengths is its widespread adoption and community support, which makes it easier to find tutorials, documentation, and third-party tools. However, Flask’s synchronous architecture can become a bottleneck when serving machine learning models to multiple clients simultaneously, as it may struggle to handle high traffic efficiently. Despite these limitations, Flask remains a popular choice for machine learning model deployment, particularly for smaller-scale applications or prototypes where simplicity and ease of use are prioritized over performance. Its flexibility and extensive ecosystem make it a versatile tool for developers, though it may require more effort to achieve the same level of robustness and scalability as FastAPI.
1.1. Project Background ¶
1.1.1 Categorical Classification ¶
This project implements the Logistic Regression Model as an independent learner and as a meta-learner of a stacking ensemble model with Decision Trees, Random Forest, and Support Vector Machine classifier algorithms using various helpful packages in Python to estimate probability of a dichotomous categorical response variable by modelling the relationship between one or more predictor variables and a binary outcome. The resulting predictions derived from the candidate models were evaluated using the F1 Score that ensures both false positives and false negatives are considered, providing a more balanced view of model classification performance. Resampling approaches including Synthetic Minority Oversampling Technique and Condensed Nearest Neighbors for imbalanced classification problems were applied by augmenting the dataset used for model training based on its inherent characteristics to achieve a more reasonably balanced distribution between the majority and minority classes. Additionally, Class Weights were also implemented by amplifying the loss contributed by the minority class and diminishing the loss from the majority class, forcing the model to focus more on correctly predicting the minority class. Penalties including Least Absolute Shrinkage and Selection Operator and Ridge Regularization were evaluated to impose constraints on the model coefficient updates.
- The complete data preprocessing and model development process was consolidated in this Jupyter Notebook.
- All associated datasets and code files were stored in this GitHub Project Repository.
- The final model was deployed as a prototype application with a web interface via Streamlit.
1.1.1.1 Data Background ¶
- The original dataset comprised rows representing observations and columns representing variables.
- The target variable is of dichotomous categorical data type:
- LUNG_CANCER (Categorical: YES, Lung Cancer Cases | NO, Non-Lung Cancer Cases)
- The complete set of 15 predictor variables contain both numeric and categorical data types:
- AGE (Numeric: Years)
- GENDER (Categorical: M, Male | F, Female)
- SMOKING (Categorical: 1, Absent | 2, Present)
- YELLOW_FINGERS (Categorical: 1, Absent | 2, Present)
- ANXIETY (Categorical: 1, Absent | 2, Present)
- PEER_PRESSURE (Categorical: 1, Absent | 2, Present)
- CHRONIC_DISEASE (Categorical: 1, Absent | 2, Present)
- FATIGUE (Categorical: 1, Absent | 2, Present)
- ALLERGY (Categorical: 1, Absent | 2, Present)
- WHEEZING (Categorical: 1, Absent | 2, Present)
- ALCOHOL_CONSUMING (Categorical: 1, Absent | 2, Present)
- COUGHING (Categorical: 1, Absent | 2, Present)
- SHORTNESS_OF_BREATH (Categorical: 1, Absent | 2, Present)
- SWALLOWING_DIFFICULTY (Categorical: 1, Absent | 2, Present)
- CHEST_PAIN (Categorical: 1, Absent | 2, Present)
- Exploratory data analysis identified a subset of 10 predictor variables that was significantly associated with the target variable and subsequently used as the final model predictors:
- YELLOW_FINGERS (Categorical: 1, Absent | 2, Present)
- ANXIETY (Categorical: 1, Absent | 2, Present)
- PEER_PRESSURE (Categorical: 1, Absent | 2, Present)
- FATIGUE (Categorical: 1, Absent | 2, Present)
- ALLERGY (Categorical: 1, Absent | 2, Present)
- WHEEZING (Categorical: 1, Absent | 2, Present)
- ALCOHOL_CONSUMING (Categorical: 1, Absent | 2, Present)
- COUGHING (Categorical: 1, Absent | 2, Present)
- SWALLOWING_DIFFICULTY (Categorical: 1, Absent | 2, Present)
- CHEST_PAIN (Categorical: 1, Absent | 2, Present)
1.1.1.2 Model Background ¶
- The model development process involved combining different Dataset Versions (with preprocessing actions applied to address class imbalance) and Model Structures (with ensembling strategy applied to improve generalization). Hyperparameter tuning was conducted using the 5-fold cross-validation method with optimal model performance determined using the F1 score.
- Individual classifier using logistic regression model developed from the original data.
- Stacked classifier using logistic regression model as a meta-learner with a decision tree model, random forest model, and support vector machine model as base learners developed from the original data.
- Individual classifier using logistic regression model developed from the SMOTE-upsampled data.
- Stacked classifier using logistic regression model as a meta-learner with a decision tree model, random forest model, and support vector machine model as base learners developed from the SMOTE-upsampled data.
- Individual classifier using logistic regression model developed from the CNN-downsampled data.
- Stacked classifier using logistic regression model as a meta-learner with a decision tree model, random forest model, and support vector machine model as base learners developed from the CNN-downsampled data.
- The stacked classifier using logistic regression model as a meta-learner with a decision tree model, random forest model, and support vector machine model as base learners developed from the SMOTE-upsampled data was selected as the final model by demonstrating the best validation F1 score with minimal overfitting.
- The final model configuration are described as follows:
- Base learner: decision tree model with optimal hyperparameters:
- max_depth = 3
- class_weight = none
- criterion = entropy
- min_samples_leaf = 3
- random_state = 88888888
- Base learner: random forest model with optimal hyperparameters:
- max_depth = 5
- class_weight = none
- criterion = entropy
- max_features = sqrt
- min_samples_leaf = 3
- random_state = 88888888
- Base learner: support vector machine model with optimal hyperparameters:
- C = 1.00
- class_weight = none
- kernel = linear
- probability = true
- random_state = 88888888
- Meta-learner: logistic regression model with optimal hyperparameters:
- penalty = none
- class_weight = none
- solver = saga
- max_iter = 500
- random_state = 88888888
- Base learner: decision tree model with optimal hyperparameters:
1.1.1.3 Deployment Background ¶
- The prediction model was deployed using a web application hosted at Streamlit.
- The user interface input consists of the following:
- radio buttons to enable binary category selection (Present | Absent) to identify the status of the test case for each of the ten clinical symptoms and behavioral indicators
- YELLOW_FINGERS
- ANXIETY
- PEER_PRESSURE
- FATIGUE
- ALLERGY
- WHEEZING
- ALCOHOL_CONSUMING
- COUGHING
- SWALLOWING_DIFFICULTY
- CHEST_PAIN
- action button to:
- process study population data as baseline
- process user input as test case
- render all entries into visualization charts
- execute all computations, estimations and predictions
- render test case prediction into logistic probability plot
- radio buttons to enable binary category selection (Present | Absent) to identify the status of the test case for each of the ten clinical symptoms and behavioral indicators
- The user interface ouput consists of the following:
- count plots to:
- provide a visualization of the proportion of lung cancer categories (Yes | No) by status (Present | Absent) as baseline
- indicate the entries made from the user input to visually assess the test case characteristics against the study population
- logistic curve plot to:
- provide a visualization of the baseline logistic regression probability curve using the study population with lung cancer categories (Yes | No)
- indicate the estimated risk index and lung cancer probability of the test case into the baseline logistic regression probability curvee
- summary table to:
- present the computed risk index, estimated lung cancer probability and predicted risk category for the test case
- count plots to:
1.1.2 Survival Prediction ¶
This project implements the Cox Proportional Hazards Regression, Cox Net Survival, Survival Tree, Random Survival Forest, and Gradient Boosted Survival models as independent base learners using various helpful packages in Python to estimate the survival probabilities of right-censored survival time and status responses. The resulting predictions derived from the candidate models were evaluated in terms of their discrimination power using the Harrel's Concordance Index metric. Penalties including Ridge Regularization and Elastic Net Regularization were evaluated to impose constraints on the model coefficient updates, as applicable. Additionally, survival probability functions were estimated for model risk-groups and sampled individual cases.
- The complete model development process was consolidated in this Jupyter Notebook.
- All associated datasets and code files were stored in this GitHub Project Repository.
- The final model was deployed as a prototype application with a web interface via Streamlit.
1.1.2.1 Data Background ¶
- The original dataset comprised rows representing observations and columns representing variables.
- The target variables contain both numeric and dichotomous categorical data types:
- DEATH_EVENT (Categorical: 0, Censored | 1, Death)
- TIME (Numeric: Days)
- The complete set of 11 predictor variables contain both numeric and categorical data types:
- AGE (Numeric: Years)
- ANAEMIA (Categorical: 0, Absent | 1 Present)
- CREATININE_PHOSPHOKINASE (Numeric: Percent)
- DIABETES (Categorical: 0, Absent | 1 Present)
- EJECTION_FRACTION (Numeric: Percent)
- HIGH_BLOOD_PRESSURE (Categorical: 0, Absent | 1 Present)
- PLATELETS (Numeric: kiloplatelets/mL)
- SERUM_CREATININE (Numeric: mg/dL)
- SERUM_SODIUM (Numeric: mEq/L)
- SEX (Categorical: 0, Female | 1, Male)
- SMOKING (Categorical: 0, Absent | 1 Present)
- Exploratory data analysis identified a subset of 6 predictor variables that was significantly associated with the target variables and subsequently used as the final model predictors:
- AGE (Numeric: Years)
- ANAEMIA (Categorical: 0, Absent | 1 Present)
- EJECTION_FRACTION (Numeric: Percent)
- HIGH_BLOOD_PRESSURE (Categorical: 0, Absent | 1 Present)
- SERUM_CREATININE (Numeric: mg/dL)
- SERUM_SODIUM (Numeric: mEq/L)
1.1.2.2 Model Background ¶
- The model development process involved evaluating different Model Structures. Hyperparameter tuning was conducted using the 5-fold cross-validation method with optimal model performance determined using the Harrel's concordance index.
- Cox proportional hazards regression model developed from the original data.
- Cox net survival model developed from the original data.
- Survival tree model developed from the original data.
- Random survival forest model, developed from the original data.
- Gradient boosted survival model developed from the original data.
- The cox proportional hazards regression model developed from the original data was selected as the final model by demonstrating the most stable Harrel's concordance index across the different internal and external validation sets.
- The final model configuration for the cox proportional hazards regression model is described as follows:
- alpha = 10
1.1.2.3 Deployment Background ¶
- The prediction model was deployed using a web application hosted at Streamlit.
- The user interface input consists of the following:
- range sliders to enable numerical input to measure the characteristics of the test case for certain cardiovascular, hematologic and metabolic markers:
- AGE
- EJECTION_FRACTION
- SERUM_CREATININE
- SERUM_SODIUM
- radio buttons to enable binary category selection (Present | Absent) to identify the status of the test case for certain hematologic and cardiovascular markers:
- ANAEMIA
- HIGH_BLOOD_PRESSURE
- action button to:
- process study population data as baseline
- process user input as test case
- render all entries into visualization charts
- execute all computations, estimations and predictions
- render test case prediction into the survival probability plot
- range sliders to enable numerical input to measure the characteristics of the test case for certain cardiovascular, hematologic and metabolic markers:
- The user interface ouput consists of the following:
- Kaplan-Meier plots to:
- provide a baseline visualization of the survival profiles of the various feature categories (Yes | No or High | Low) estimated from the study population given the survival time and event status
- Indicate the entries made from the user input to visually assess the survival probabilities of the test case characteristics against the study population across all time points
- survival probability plot to:
- provide a visualization of the baseline survival probability profile using each observation of the study population given the survival time and event status
- indicate the heart failure survival probabilities of the test case at different time points
- summary table to:
- present the estimated heart failure survival probabilities and predicted risk category for the test case
- Kaplan-Meier plots to:
1.1.3 Image Classification ¶
This project focuses on leveraging the Convolutional Neural Network Model using various helpful packages in Python for multiclass image classification by directly learning hierarchical features from raw pixel data. The CNN models were designed to extract low- and high-level features for differentiating between image categories. Various hyperparameters, including the number of layers, filter size, and number of dense layer weights, were systematically evaluated to optimize the model architecture. Image Augmentation techniques were employed to increase the diversity of training images and improve the model's ability to generalize. To enhance model performance and robustness, various regularization techniques were explored, including Dropout, Batch Normalization, and their Combinations. These methods helped mitigate overfitting and ensured stable learning. Callback functions such as Early Stopping, Learning Rate Reduction on Performance Plateaus, and Model Checkpointing were implemented to fine-tune the training process, optimize convergence, and prevent overtraining. Model evaluation was conducted using Precision, Recall, and F1 Score metrics to ensure both false positives and false negatives are considered, providing a more balanced view of model classification performance. Post-training, interpretability was emphasized through an advanced visualization technique using Gradient Class Activation Mapping (Grad-CAM), providing insights into the spatial and hierarchical features that influenced the model's predictions, offering a deeper understanding of the decision-making process.
- The complete model development process was consolidated in this Jupyter Notebook.
- All associated datasets and code files were stored in this GitHub Project Repository.
- The final model was deployed as a prototype application with a web interface via Streamlit.
1.1.3.1 Data Background ¶
- The original dataset comprised labeled images.
- The target variable is of multi-class categorical data type:
- CLASS (Categorical: No Tumor, Clear MRI | Pituitary, MRI Putuitary Tumors | Meningioma, MRI Meningioma Tumors | Glioma, MRI Glioma Tumors)
- The hierarchical representation of image features enables the model to transform raw pixel data into a meaningful and compact representation, allowing it to make accurate predictions during image classification. The different features automatically learned during the training process are given (but not limited to) as follows:
- OBJECT TEXTURE (Repetitive gradients among pixel intensities)
- OBJECT EDGE (Abrupt changes or transitions on pixel intensities)
- OBJECT PATTERN (Distinctive structural features in pixel intensities)
- OBJECT SHAPE (Spatial relationships and contours among pixel intensities)
- SPATIAL HIERARCHY (Layered abstract representations of spatial structures in image objects)
- SPATIAL LOCALIZATION (Boundaries and position of the object within the image)
1.1.3.2 Model Background ¶
- Candidate convolutional neural network models were formulated using Tensorflow Keras layers as follows:
- Convolutional Layer (Conv_2D) - extracts features from input images using convolutional filters
- Maximum Pooling Layer (MaxPooling2D) - Reduces spatial dimensions and downsamples feature maps
- Activation Layer (Activation)- Applies an activation function element-wise to the output
- Flatten Layer (Flatten) - Flattens the input to a 1D array, preparing for fully connected layers
- Dense Layer (Dense) - Fully connected layer for classification
- Different iterations of the model were formulated using variations in the inclusion or exclusion of the following regularization layers:
- Dropout Layer (Dropout) - randomly drops (sets to zero) a fraction of the neurons during training reducing co-dependencies between them
- Batch Normalization Layer (BatchNormalization) - adjusts and scales the inputs to a layer reducing the sensitivity to weight initialization choices
- A subset of hyperparameters for the different layers were fixed during model training including:
- kernel_size - setting used to define the local region the convolutional layer considers when processing the input
- activation - setting used to introduce non-linearity into the model, enabling it to learn complex relationships in the data
- pool_size - setting used to reduce the spatial dimensions of the feature maps to focus on the most important features
- padding - setting used to control the spatial size and shape for every convolutional operation at each stage
- optimizer - setting used to determine how the model's weights are updated during training
- batch_size - setting used to determine how many samples are used in each iteration of training
- loss - setting used to define the objective that the model seeks to minimize during training
- A subset of hyperparameters for the different layers were optimized during model training including:
- filters - setting used to capture spatial hierarchies and features in the input images
- units - setting used to process the flattened feature maps and determine the dimensionality of the output space
- learning_rate - setting used to determine the step size at each iteration during optimization
- Two CNN model structures were additionally evaluated as follows:
- Simple
- Lesser number of Conv_2D
- Lower values set for filters
- Lower values set for units
- Complex
- Higher number of Conv_2D
- Higher values set for filters
- Higher values set for units
- Simple
- The model development process involved combining different CNN Model Structures and Regularization Layers with optimal model performance determined using precision, recall and F1 score.
- Simple CNN model developed from augmented data with no regularization.
- Complex CNN model developed from augmented data with no regularization.
- Simple CNN model developed from augmented data with Dropout regularization.
- Complex CNN model developed from augmented data with Dropout regularization.
- Simple CNN model developed from augmented data with Batch Normalization regularization.
- Complex CNN model developed from augmented data with Batch Normalization regularization.
- Simple CNN model developed from augmented data with Dropout and Batch Normalization regularization.
- Complex CNN model developed from augmented data with Dropout and Batch Normalization regularization.
- The complex CNN model developed from augmented data with Dropout and Batch Normalization regularization was selected as the final model by demonstrating the best validation precision, recall and F1 score with minimal overfitting.
- The final CNN model contained 6,446,596 trainable parameters broken down per layer as follows:
- Conv2D: cdrbnr_complex_conv2d_0
- output size = 227x227x16
- number of parameters = 160
- MaxPooling2D: cdrbnr_complex_max_pooling2d_0
- output size = 113x113x16
- number of parameters = 0
- Conv2D: cdrbnr_complex_conv2d_1
- output size = 113x113x32
- number of parameters = 4,640
- MaxPooling2D: cdrbnr_complex_max_pooling2d_1
- output size = 56x56x32
- number of parameters = 0
- Conv2D: cdrbnr_complex_conv2d_2
- output size = 56x56x64
- number of parameters = 18,496
- BatchNormalization: cdrbnr_complex_batch_normalization
- output size = 56x56x64
- number of parameters = 256
- Activation: cdrbnr_complex_activation
- output size = 56x56x64
- number of parameters = 0
- MaxPooling2D: cdrbnr_complex_max_pooling2d_2
- output size = 28x28x64
- number of parameters = 0
- Flatten: cdrbnr_complex_flatten
- output size = 50,176
- number of parameters = 0
- Dense: cdrbnr_complex_dense_0
- output size = 128
- number of parameters = 6,422,656
- Dropout: cdrbnr_complex_dropout
- output size = 128
- number of parameters = 0
- Dense: cdrbnr_complex_dense_1
- output size = 4
- number of parameters = 516
- Conv2D: cdrbnr_complex_conv2d_0
1.1.3.3 Deployment Background ¶
- The prediction model was deployed using a web application hosted at Streamlit.
- The user interface input consists of the following:
- First action button to:
- randomly sample an MR image as a test case
- conduct image preprocessing
- display the RGB channels
- activates the second action button
- Second action button to:
- load fitted CNN model
- estimate image class probabilities
- predict class categories
- perform the Grad-CAM computation
- display the Grad-CAM visualization for all convolutional layers
- render test case prediction summary
- First action button to:
- The user interface ouput consists of the following:
- RGB plots to:
- provide a baseline visualization of the test case by channel
- Grad-CAM plots to:
- present insights into how the model progressively learns features (from low-level to high-level) for each convolutional layer, aiding in understanding spatial and hierarchical representation
- highlight image regions that influenced the model's decision the most allowing to verify whether the model focuses on relevant areas
- summary table to:
- indicate if the model prediction matches the ground truth
- present the estimated class probabilities and predicted class for the test case
- RGB plots to:
1.2. Application Programming Interface (API) Development Using the FastAPI Framework ¶
1.2.1 Categorical Classification ¶
1.2.1.1 API Building ¶
- An API code using the FastAPI framework was developed for deploying a categorical classification model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as FastAPI, HTTPException, and BaseModel for API development.
- Included libraries for data manipulation (numpy, pandas) and model loading (joblib).
- Defining File Paths
- Specified the MODELS_PATH to locate the pre-trained scikit-learn model.
- Loading the Pre-Trained Classification Model
- Loaded the pre-trained scikit-learn model (stacked_balanced_class_best_model_upsampled.pkl) using joblib.load.
- Handled potential errors during model loading with a try-except block.
- Defining Input Schemas
- Created a Pydantic BaseModel class to define the input schema for TestSample: For individual test cases, expecting a list of floats as input features.
- Created a Pydantic BaseModel class to define the input schema for TestBatch: For batch processing, expecting a list of lists of floats as input features.
- Initializing the FastAPI App
- Created a FastAPI instance (app) to define and serve API endpoints.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- Individual Prediction Endpoint (/predict-individual-logit-probability-class): A POST endpoint to compute the logit, probability, and risk category for an individual test case.
- Batch Prediction Endpoint (/predict-list-logit-probability-class): A POST endpoint to compute and sort logits and probabilities for a batch of test cases.
- Individual Prediction Logic
- Converted the input data into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s decision_function to compute the logit value.
- Used the predict_proba method to estimate the probability of the positive class.
- Classified the test case as "Low-Risk" or "High-Risk" based on a probability threshold of 0.50.
- Batch Prediction Logic
- Converted the batch input data into a pandas DataFrame with appropriate feature names.
- Computed logits and probabilities for all cases in the batch using decision_function and predict_proba.
- Sorted the logits and probabilities in ascending order for easier interpretation.
- Error Handling
- Implemented robust error handling for invalid inputs or prediction errors using HTTPException.
- Returned meaningful error messages and appropriate HTTP status codes (e.g., 400 for bad requests).
- Running the FastAPI App
- Used uvicorn to run the FastAPI app on localhost at port 8000.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported both individual and batch predictions, making the API versatile for different use cases.
- Provided logits, probabilities, and risk categories as outputs for interpretable results.
- Ensured input validation and error handling for robust API performance.
1.2.1.2 API Testing ¶
- The API code developed using the FastAPI framework deploying a categorical classification model was successfully tested with results presented as follows:
- Server Initialization: FastAPI application was started successfully, with Uvicorn running on http://127.0.0.1:8000, indicating that the server and its documentation are active and ready to process requests.
- Hot Reloading Activated: Uvicorn's reloader process (WatchFiles) was initialized, allowing real-time code changes without restarting the server.
- Server Process Started: The primary server process was assigned a process ID (17924), confirming successful application launch.
- Application Ready State: The server was shown to wait for incoming requests, ensuring all necessary components, including model loading, are successfully initialized.
- Root Endpoint Accessed (GET /):The API received a GET request at the root endpoint and responded with 200 OK, confirming that the service is running and accessible.
- Individual Prediction Request (POST /predict-individual-logit-probability-class): A POST request was processed successfully, returning 200 OK, indicating that the API correctly handled and responded to an individual classification request.
- Batch Prediction Request (POST /predict-list-logit-probability-class): The API successfully processed a POST request for batch classification, returning 200 OK, confirming that multiple test cases were handled correctly.
- Invalid Input Handling (POST /predict-individual-logit-probability-class): A malformed or incorrectly structured request resulted in a 422 Unprocessable Entity response, demonstrating the API's robust error-handling mechanism for invalid input formats.
##################################
# Loading Python Libraries
##################################
import requests
##################################
# Defining the base URL of the API
# for the categorical classification model
##################################
CC_FASTAPI_BASE_URL = "http://127.0.0.1:8000"
##################################
# Defining the input values for an individual test case
##################################
individual_test_case = {
"features_individual": [1, 0, 0, 0, 0, 1, 0, 0, 1, 1]
}
##################################
# Defining the input values for a batch of cases
##################################
batch_test_case = {
"features_list": [
[1, 0, 0, 0, 0, 1, 0, 0, 1, 1],
[1, 0, 1, 0, 1, 1, 0, 1, 1, 1]
]
}
##################################
# Generating a GET endpoint request for
# for validating API service connection
##################################
response = requests.get(f"{CC_FASTAPI_BASE_URL}/")
if response.status_code == 200:
print("Root Endpoint Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Root Endpoint Response: {'message': 'Welcome to the Categorical Classification API!'}
##################################
# Generating a POST endpoint request for
# computing the risk index,
# estimating the lung cancer probability,
# and predicting the risk category
# of an individual test case
##################################
response = requests.post(f"{CC_FASTAPI_BASE_URL}/predict-individual-logit-probability-class", json=individual_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'logit': -1.2117837409390746, 'probability': 0.22938559072691203, 'risk_class': 'Low-Risk'}
##################################
# Sending a POST endpoint request for
# computing the risk index,
# estimating the lung cancer probability,
# and predicting the risk category
# of a list of train cases
##################################
response = requests.post(f"{CC_FASTAPI_BASE_URL}/predict-list-logit-probability-class", json=batch_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'logit': [-1.2117837409390746, 3.4784950973590973], 'probability': [0.22938559072691203, 0.9700696569701589], 'logit_sorted': [-1.2117837409390746, 3.4784950973590973], 'probability_sorted': [0.22938559072691203, 0.9700696569701589]}
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
malformed_test_case = {"features": [1, 0, 1]}
response = requests.post(f"{CC_FASTAPI_BASE_URL}/predict-individual-logit-probability-class", json=malformed_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Error: 422 {"detail":[{"type":"missing","loc":["body","features_individual"],"msg":"Field required","input":{"features":[1,0,1]}}]}
1.2.2 Survival Prediction ¶
1.2.2.1 API Building ¶
- An API code using the FastAPI framework was developed for deploying a survival prediction model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as FastAPI, HTTPException, and BaseModel for API development.
- Included libraries for survival analysis (sksurv, lifelines), data manipulation (numpy, pandas), and visualization (matplotlib).
- Used io and base64 for encoding and handling image outputs.
- Defining File Paths
- Specified the MODELS_PATH and PARAMETERS_PATH to locate the pre-trained survival model and related parameters.
- Loading the Pre-Trained Survival Model
- Loaded the pre-trained Cox Proportional Hazards (CoxPH) model (coxph_best_model.pkl) using joblib.load.
- Handled potential errors during model loading with a try-except block.
- Loading Model Parameters
- Loaded the median values for numeric features (numeric_feature_median_list.pkl) to support feature binning.
- Loaded the risk group threshold (coxph_best_model_risk_group_threshold.pkl) for categorizing patients into "High-Risk" and "Low-Risk" groups.
- Defining Input Schemas
- Created a Pydantic BaseModel class to define input schema for TestSample: For individual test cases, expecting a list of floats as input features.
- Created a Pydantic BaseModel class to define input schema for TrainList: For batch processing, expecting a list of lists of floats as input features.
- Created a Pydantic BaseModel class to define input schema for BinningRequest: For dichotomizing numeric features based on the median.
- Created a Pydantic BaseModel class to define input schema for KaplanMeierRequest: For generating Kaplan-Meier survival plots.
- Initializing the FastAPI App
- Created a FastAPI instance (app) to define and serve API endpoints.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- Individual Survival Prediction Endpoint (/compute-individual-coxph-survival-probability-class/): A POST endpoint to generate survival profiles, estimate survival probabilities, and predict risk categories for individual test cases.
- Batch Survival Profile Endpoint (/compute-list-coxph-survival-profile/): A POST endpoint to generate survival profiles for a batch of cases.
- Feature Binning Endpoint (/bin-numeric-model-feature/): A POST endpoint to dichotomize numeric features based on the median.
- Kaplan-Meier Plot Endpoint (/plot-kaplan-meier/): A POST endpoint to generate and return Kaplan-Meier survival plots.
- Individual Survival Prediction Logic
- Converted the input data into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s predict_survival_function to generate the survival function for the test case.
- Predicted the risk category ("High-Risk" or "Low-Risk") based on the model’s risk score and threshold.
- Interpolated survival probabilities at predefined time points (e.g., 50, 100, 150, 200, 250 days).
- Batch Survival Profile Logic
- Converted the batch input data into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s predict_survival_function to generate survival functions for all cases in the batch.
- Extracted and returned survival profiles for each case.
- Feature Binning Logic
- Converted the input data into a pandas DataFrame.
- Dichotomized the specified numeric feature into "Low" and "High" categories based on the median value.
- Returned the binned data as a list of dictionaries.
- Kaplan-Meier Plot Logic
- Converted the input data into a pandas DataFrame.
- Initialized a KaplanMeierFitter object to estimate survival probabilities.
- Plotted survival curves for different categories of the specified variable (e.g., "Low" vs. "High").
- Included an optional new case value for comparison in the plot.
- Saved the plot as a base64-encoded image and returned it in the API response.
- Error Handling
- Implemented robust error handling for invalid inputs or prediction errors using HTTPException.
- Returned meaningful error messages and appropriate HTTP status codes (e.g., 500 for server errors).
- Running the FastAPI App
- Used uvicorn to run the FastAPI app on localhost at port 8001.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported both individual and batch predictions, making the API versatile for different use cases.
- Provided survival probabilities, risk categories, and visualizations (Kaplan-Meier plots) for interpretable results.
- Enabled feature binning for categorical analysis of numeric features.
1.2.2.2 API Testing ¶
- The API code developed using the FastAPI framework deploying a survival prediction model was successfully tested with results presented as follows:
- Server Initialization: FastAPI application was started successfully, with Uvicorn running on http://127.0.0.1:8001, indicating that the server and its documentation are active and ready to process requests.
- Hot Reloading Activated: Uvicorn's reloader process (WatchFiles) was initialized, allowing real-time code changes without restarting the server.
- Server Process Started: The primary server process was assigned a process ID (18676), confirming successful application launch.
- Application Ready State: The server was shown to wait for incoming requests, ensuring all necessary components, including model loading, are successfully initialized.
- Root Endpoint Accessed (GET /): The API received a GET request at the root endpoint and responded with 200 OK, confirming that the service is running and accessible.
- Individual Survival Probability Request (POST /compute-individual-coxph-survival-probability-class/): A POST request was processed successfully, returning 200 OK, indicating that the API correctly computed survival probabilities and risk categorization for an individual test case.
- Batch Survival Profile Request (POST /compute-list-coxph-survival-profile/): The API successfully processed a POST request for batch survival profile computation, returning 200 OK, confirming that multiple test cases were handled correctly.
- Feature Binning Request (POST /bin-numeric-model-feature/): A POST request was successfully executed, returning 200 OK, confirming that the API correctly categorized numeric model features into dichotomous bins.
- Kaplan-Meier Plot Request (POST /plot-kaplan-meier/): The API successfully processed a POST request, returning 200 OK, indicating that a Kaplan-Meier survival plot was generated and returned as a base64-encoded image.
- Invalid Input Handling (POST /compute-individual-coxph-survival-probability-class/): A malformed or incorrectly structured request resulted in a 422 Unprocessable Entity response, demonstrating the API's robust error-handling mechanism for invalid input formats.
##################################
# Loading Python Libraries
##################################
import requests
import json
import pandas as pd
import base64
from IPython.display import display
from PIL import Image
##################################
# Defining the base URL of the API
# for the survival prediction model
##################################
SP_FASTAPI_BASE_URL = "http://127.0.0.1:8001"
##################################
# Defining the input values for an individual test case
##################################
single_test_case = {
"features_individual": [43, 0, 75, 1, 0.75, 100]
}
##################################
# Defining the input values for a batch of cases
##################################
train_list = {
"features_list": [
[43, 0, 75, 1, 0.75, 100],
[70, 1, 20, 1, 0.75, 100]
]
}
##################################
# Defining the input values for a batch of cases for binning request
##################################
bin_request = {
"X_original_list": [
{"AGE": -0.10, "EJECTION_FRACTION": -0.10, "SERUM_CREATININE ": -0.10, "SERUM_SODIUM": -0.10},
{"AGE": 0.20, "EJECTION_FRACTION": 0.20, "SERUM_CREATININE ": 0.20, "SERUM_SODIUM": 0.20},
{"AGE": 0.90, "EJECTION_FRACTION": 0.90, "SERUM_CREATININE ": 0.90, "SERUM_SODIUM": 0.90}
],
"numeric_feature": "AGE"
}
##################################
# Defining the input values for a batch of cases for Kaplan-Meier plotting
##################################
km_request = {
"df": [
{"TIME": 0, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 25, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 50, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 100, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 125, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 150, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 175, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 200, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 225, "DEATH_EVENT": 1, "AGE": "Low"},
{"TIME": 250, "DEATH_EVENT": 1, "AGE": "Low"},
{"TIME": 0, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 25, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 50, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 100, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 125, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 150, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 175, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 200, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 225, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 250, "DEATH_EVENT": 1, "AGE": "High"},
],
"cat_var": "AGE",
"new_case_value": "Low"
}
##################################
# Generating a GET endpoint request for
# for validating API service connection
##################################
response = requests.get(f"{SP_FASTAPI_BASE_URL}/")
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'message': 'Welcome to the Survival Prediction API!'}
##################################
# Sending a POST endpoint request for
# generating the heart failure survival profile,
# estimating the heart failure survival probabilities,
# and predicting the risk category
# of an individual test case
##################################
response = requests.post(f"{SP_FASTAPI_BASE_URL}/compute-individual-coxph-survival-probability-class/", json=single_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'survival_function': [0.9973812917524568, 0.9920416812438736, 0.9893236791425079, 0.972381113071464, 0.9693179903073035, 0.9631930672135339, 0.9631930672135339, 0.9600469571766689, 0.9600469571766689, 0.9568596864927983, 0.9536305709158891, 0.9471625843882805, 0.93729581350105, 0.9338986486591409, 0.93048646553474, 0.9270645831787163, 0.9202445006124622, 0.9167715111530355, 0.9132845175345189, 0.9097550958520674, 0.9097550958520674, 0.9097550958520674, 0.9060810720432387, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.8985598696587259, 0.8985598696587259, 0.8985598696587259, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8901959645503091, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8764677174183527, 0.8764677174183527, 0.8764677174183527, 0.8764677174183527, 0.8709113650481243, 0.8709113650481243, 0.8652494086650531, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8459534502216807, 0.8389821875092403, 0.8319419786276306, 0.8246669811915435, 0.8099879066057215, 0.8099879066057215, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7337716342207724, 0.7337716342207724, 0.7337716342207724, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696], 'survival_time': [50, 100, 150, 200, 250], 'survival_probabilities': [90.97550958520674, 87.64677174183527, 84.59534502216806, 78.48178617845467, 70.70184115456696], 'risk_category': 'Low-Risk'}
##################################
# Sending a POST endpoint request for
# generating the heart failure survival profile and
# estimating the heart failure survival probabilities
# of a list of train cases
##################################
response = requests.post(f"{SP_FASTAPI_BASE_URL}/compute-list-coxph-survival-profile/", json=train_list)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'survival_profiles': [[0.9973812917524568, 0.9920416812438736, 0.9893236791425079, 0.972381113071464, 0.9693179903073035, 0.9631930672135339, 0.9631930672135339, 0.9600469571766689, 0.9600469571766689, 0.9568596864927983, 0.9536305709158891, 0.9471625843882805, 0.93729581350105, 0.9338986486591409, 0.93048646553474, 0.9270645831787163, 0.9202445006124622, 0.9167715111530355, 0.9132845175345189, 0.9097550958520674, 0.9097550958520674, 0.9097550958520674, 0.9060810720432387, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.8985598696587259, 0.8985598696587259, 0.8985598696587259, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8901959645503091, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8764677174183526, 0.8764677174183526, 0.8764677174183526, 0.8764677174183526, 0.8709113650481243, 0.8709113650481243, 0.8652494086650531, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8459534502216807, 0.8389821875092403, 0.8319419786276306, 0.8246669811915435, 0.8099879066057215, 0.8099879066057215, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7337716342207724, 0.7337716342207724, 0.7337716342207724, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695], [0.9761144218801228, 0.928980888267716, 0.905777064852962, 0.7724242339590301, 0.7502787164583535, 0.7076872741961866, 0.7076872741961866, 0.6866593185026403, 0.6866593185026403, 0.6659260634219393, 0.6454915885099762, 0.6062342264686207, 0.5504405490863784, 0.5323184765768243, 0.5146536440658629, 0.49746533888245986, 0.464726413395405, 0.4488047347163327, 0.433309930422515, 0.4181140392975694, 0.4181140392975694, 0.4181140392975694, 0.4028020116306455, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.373006008573048, 0.373006008573048, 0.373006008573048, 0.35785929480690143, 0.35785929480690143, 0.35785929480690143, 0.35785929480690143, 0.3421927616040032, 0.3117176431598899, 0.3117176431598899, 0.3117176431598899, 0.3117176431598899, 0.29651072362871467, 0.29651072362871467, 0.29651072362871467, 0.29651072362871467, 0.2796248763802668, 0.2796248763802668, 0.2633052706162029, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.213871875082776, 0.19816204676152543, 0.18334919195124752, 0.16908728217620728, 0.1432835564355214, 0.1432835564355214, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.10710184631976834, 0.10710184631976834, 0.10710184631976834, 0.10710184631976834, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.05761125190131533, 0.05761125190131533, 0.05761125190131533, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404]]}
##################################
# Sending a POST endpoint request for
# creating dichotomous bins for the numeric features
# of a list of train cases
##################################
response = requests.post(f"{SP_FASTAPI_BASE_URL}/bin-numeric-model-feature/", json=bin_request)
if response.status_code == 200:
display("Response:", pd.DataFrame(response.json()))
else:
print("Error:", response.status_code, response.text)
'Response:'
AGE | EJECTION_FRACTION | SERUM_CREATININE | SERUM_SODIUM | |
---|---|---|---|---|
0 | Low | -0.1 | -0.1 | -0.1 |
1 | High | 0.2 | 0.2 | 0.2 |
2 | High | 0.9 | 0.9 | 0.9 |
##################################
# Sending a POST endpoint request for
# plotting the estimated survival profiles
# using Kaplan-Meier Plots
##################################
response = requests.post(f"{SP_FASTAPI_BASE_URL}/plot-kaplan-meier/", json=km_request)
if response.status_code == 200:
plot_data = response.json()["plot"]
# Decoding and displaying the plot
img = base64.b64decode(plot_data)
with open("kaplan_meier_plot.png", "wb") as f:
f.write(img)
display(Image.open("kaplan_meier_plot.png"))
else:
print("Error:", response.status_code, response.text)
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
malformed_test_case = {"features": [43, 0, 75, 1, 0.75]}
response = requests.post(f"{SP_FASTAPI_BASE_URL}/compute-individual-coxph-survival-probability-class", json=malformed_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Error: 422 {"detail":[{"type":"missing","loc":["body","features_individual"],"msg":"Field required","input":{"features":[43,0,75,1,0.75]}}]}
1.2.3 Image Classification ¶
1.2.3.1 API Building ¶
- An API code using the FastAPI framework was developed for deploying a CNN image classification model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as FastAPI, File, UploadFile, HTTPException, and JSONResponse for API development.
- Included libraries for deep learning (tensorflow, keras), image processing (PIL, cv2), and visualization (matplotlib).
- Used io, os, random, math, and base64 for file handling, random seed setting, and encoding.
- Setting Random Seed
- Defined a set_seed function to ensure reproducibility by setting random seeds for numpy, tensorflow, keras, and Python’s random module.
- Enabled deterministic operations in TensorFlow to ensure consistent results across runs.
- Defining File Paths
- Specified the MODELS_PATH to locate the pre-trained CNN model and other related files.
- Loading the Pre-Trained CNN Model
- Loaded the pre-trained CNN model (cdrbnr_complex_best_model.keras) using keras.models.load_model.
- Handled potential errors during model loading with a try-except block.
- Recreating the CNN Model Using Functional API
- Recreated the CNN model using Keras’ Functional API to access intermediate layers for Grad-CAM visualization.
- Defined the input layer and sequentially applied layers from the original model to reconstruct it.
- Compiling the Functional API Model
- Compiled the recreated model with the same loss function (categorical_crossentropy), optimizer (adam), and metrics (Recall) as the original model.
- Creating Gradient Models for Grad-CAM
- Created three gradient models (grad_model_first_conv2d, grad_model_second_conv2d, grad_model_third_conv2d) to extract outputs from the first, second, and third convolutional layers.
- Compiled these models with the same parameters as the main model.
- Initializing the FastAPI App
- Created a FastAPI instance (app) to define and serve API endpoints.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- File Upload Test Endpoint (/test-file-upload/): A POST endpoint to test file upload functionality, returning file metadata (filename, content type, size).
- Image Classification Endpoint (/predict-image-category-class-probability/): A POST endpoint to predict the image category and estimate class probabilities for an uploaded JPEG image.
- Grad-CAM Visualization Endpoint (/visualize-image-gradcam/): A POST endpoint to generate and return Grad-CAM heatmaps for the uploaded image.
- Image Preprocessing
- Read and resized uploaded images to the required input size (227x227 pixels).
- Converted images to grayscale and normalized pixel values to the range (0, 1).
- Expanded image dimensions to match the model’s input shape.
- Model Inference
- Used the pre-trained CNN model to predict the image category and class probabilities.
- Mapped predicted class indices to human-readable labels (e.g., "No Tumor", "Pituitary", "Meningioma", "Glioma").
- Grad-CAM Heatmap Generation
- Defined a make_gradcam_heatmap function to compute Grad-CAM heatmaps using TensorFlow’s GradientTape.
- Generated heatmaps for the first, second, and third convolutional layers.
- Superimposed heatmaps onto the original image using a color map (turbo) and transparency.
- Visualization and Encoding
- Created a multi-panel plot to display Grad-CAM heatmaps for all three convolutional layers.
- Saved the plot to a buffer, encoded it as a base64 string, and returned it in the API response.
- Error Handling
- Implemented robust error handling for invalid file uploads (e.g., missing files, non-JPEG images).
- Used HTTPException to return appropriate error messages and status codes (e.g., 400 for bad requests, 500 for server errors).
- Running the FastAPI App
- Used uvicorn to run the FastAPI app on localhost at port 8002.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported JPEG image uploads and validated file types.
- Provided both classification results and visual explanations (Grad-CAM) for model predictions.
- Ensured reproducibility and deterministic behavior through random seed setting.
1.2.3.2 API Testing ¶
- The API code developed using the FastAPI framework deploying a CNN image classification model was successfully tested with results presented as follows:
- Server Initialization: FastAPI application was started successfully, confirming that the server and its documentation are active and ready to process requests.
- Hot Reloading Activated: Uvicorn's reloader process (WatchFiles) was initialized, allowing real-time code changes without restarting the server.
- Server Process Started: The primary server process was assigned a process ID (16508), indicating successful application launch.
- Application Ready State: The server was shown to wait for incoming requests, ensuring all necessary components, including model loading, are successfully initialized.
- Root Endpoint Accessed (GET /): The API received a GET request at the root endpoint and responded with 200 OK, confirming that the service is running and accessible.
- File Upload Test (POST /test-file-upload/): A POST request for file uploads was successfully processed, returning metadata and 200 OK, verifying the API's ability to handle file uploads.
- Image Classification Prediction (POST /predict-image-category-class-probability/): The API processed an image classification request, returning predicted class probabilities with a 200 OK response.
- Grad-CAM Visualization (POST /visualize-image-gradcam/): A POST request for Grad-CAM visualization was handled successfully, generating a heatmap overlay and returning 200 OK.
- Invalid Input Handling (POST /test-file-upload/): A malformed or unsupported request resulted in a 400 Bad Request response, demonstrating the API's robust validation and error-handling mechanisms.
##################################
# Loading Python Libraries
##################################
import requests
import json
import base64
from IPython.display import display
from PIL import Image
import matplotlib.pyplot as plt
import io
from tensorflow.keras.utils import load_img
import os
import mimetypes
##################################
# Defining the base URL of the API
# for the image classification model
##################################
IC_FASTAPI_BASE_URL = "http://127.0.0.1:8002"
##################################
# Defining the file path for an individual test image
##################################
IMAGES_PATH = r"image_classification_study\images"
image_path = (os.path.join("..",IMAGES_PATH, "test_image.jpg"))
##################################
# Automatically determining the filename and content type
##################################
image_path_filename = os.path.basename(image_path)
image_path_content_type, _ = mimetypes.guess_type(image_path)
##################################
# Visualizing the individual test image
##################################
try:
image = Image.open(image_path)
print(f"Image File Path: {image_path}")
print(f"Image Format: {image.format}")
print(f"Image Size: {image.size}")
print(f"Image Mode: {image.mode}")
except Exception as e:
print(f"Error loading image: {e}")
plt.imshow(image)
plt.axis('off')
plt.title("Test Image")
plt.show()
Image File Path: ..\image_classification_study\images\test_image.jpg Image Format: JPEG Image Size: (215, 234) Image Mode: RGB
##################################
# Generating a GET endpoint request for
# for validating API service connection
##################################
response = requests.get(f"{IC_FASTAPI_BASE_URL}/")
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'message': 'Welcome to the Image Classification API!'}
##################################
# Sending a POST endpoint request for
# ensuring that the file upload mechanism is working
# by returning the the file metadata
##################################
with open(image_path, "rb") as file:
files = {"file": (image_path_filename, file, image_path_content_type)}
response = requests.post(f"{IC_FASTAPI_BASE_URL}/test-file-upload/", files=files)
if response.status_code == 200:
result = response.json()
print("File Upload Test Result:")
print(f"Filename: {result['filename']}")
print(f"Content Type: {result['content_type']}")
print(f"Size: {result['size']} bytes")
else:
print(f"Error: {response.status_code} - {response.text}")
File Upload Test Result: Filename: test_image.jpg Content Type: image/jpeg Size: 12103 bytes
##################################
# Sending a POST endpoint request for
# predicting the image category and
# estimating class probabilities
# of an individual test image
##################################
with open(image_path, "rb") as file:
files = {"file": ("image.jpg", file, "image/jpeg")}
response = requests.post(f"{IC_FASTAPI_BASE_URL}/predict-image-category-class-probability/", files=files)
if response.status_code == 200:
result = response.json()
print("Prediction Result:")
print(f"Predicted Class: {result['predicted_class']}")
print("Probabilities:")
for cls, prob in result["probabilities"].items():
print(f"{cls}: {prob:.5f}%")
else:
print(f"Error: {response.status_code} - {response.text}")
Prediction Result: Predicted Class: Meningioma Probabilities: No Tumor: 12.94989% Glioma: 0.02788% Meningioma: 87.02222% Pituitary: 0.00002%
##################################
# Sending a POST endpoint request for
# formulating the gradient class activation map
# from the output of the first to third convolutional layers and
# and superimposing on the actual image
##################################
with open(image_path, "rb") as file:
files = {"file": ("image.jpg", file, "image/jpeg")}
response = requests.post(f"{IC_FASTAPI_BASE_URL}/visualize-image-gradcam/", files=files)
if response.status_code == 200:
plot_data = response.json()["plot"]
# Decoding and displaying the plot
img = base64.b64decode(plot_data)
with open("image_gradcam_plot.png", "wb") as f:
f.write(img)
display(Image.open("image_gradcam_plot.png"))
else:
print(f"Error: {response.status_code} - {response.text}")
##################################
# Defining the file path for an individual test image
##################################
IMAGES_PATH = r"image_classification_study\images"
malformed_image_path = (os.path.join("..",IMAGES_PATH, "test_image.png"))
##################################
# Automatically determining the filename and content type
##################################
malformed_image_path_filename = os.path.basename(image_path)
malformed_image_path_content_type, _ = mimetypes.guess_type(malformed_image_path)
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
with open(malformed_image_path, "rb") as file:
files = {"file": (malformed_image_path_filename, file, malformed_image_path_content_type)}
response = requests.post(f"{IC_FASTAPI_BASE_URL}/test-file-upload/", files=files)
if response.status_code == 200:
result = response.json()
print("File Upload Test Result:")
print(f"Filename: {result['filename']}")
print(f"Content Type: {result['content_type']}")
print(f"Size: {result['size']} bytes")
else:
print(f"Error: {response.status_code} - {response.text}")
Error: 400 - {"detail":"File must be a JPEG image"}
1.3. Application Programming Interface (API) Development Using the Flask Framework ¶
1.3.1 Categorical Classification ¶
1.3.1.1 API Building ¶
- An API code using the Flask framework was developed for deploying a categorical classification model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as Flask, request, and jsonify for API development.
- Included libraries for data manipulation (numpy, pandas) and model loading (joblib).
- Integrated Flasgger for Swagger API documentation.
- Defining File Paths
- Specified the MODELS_PATH to locate the pre-trained scikit-learn model.
- Loading the Pre-Trained Classification Model
- Loaded the pre-trained scikit-learn model (stacked_balanced_class_best_model_upsampled.pkl) using joblib.load.
- Handled potential errors during model loading with a try-except block.
- Defining Input Schemas
- Created a Pydantic BaseModel class to define the input schema for TestSample: For individual test cases, expecting a list of floats as input features.
- Created a Pydantic BaseModel class to define the input schema for TestBatch: For batch processing, expecting a list of lists of floats as input features.
- Initializing the Flask App
- Created a Flask instance (app) to define and serve API endpoints.
- Integrated Swagger for automatic API documentation.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- Individual Prediction Endpoint (/predict-individual-logit-probability-class): A POST endpoint to compute the logit, probability, and risk category for an individual test case.
- Batch Prediction Endpoint (/predict-list-logit-probability-class): A POST endpoint to compute and sort logits and probabilities for a batch of test cases.
- Individual Prediction Logic
- Extracted input data from the request body and converted it into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s decision_function to compute the logit value.
- Used the predict_proba method to estimate the probability of the positive class.
- Classified the test case as "Low-Risk" or "High-Risk" based on a probability threshold of 0.50.
- Batch Prediction Logic
- Extracted batch input data from the request body and converted it into a pandas DataFrame with appropriate feature names.
- Computed logits and probabilities for all cases in the batch using decision_function and predict_proba.
- Sorted the logits and probabilities in ascending order for easier interpretation.
- Error Handling
- Implemented robust error handling for invalid inputs or prediction errors using jsonify.
- Returned meaningful error messages and appropriate HTTP status codes (e.g., 400 for bad requests).
- Running the Flask App
- Used app.run to run the Flask app on localhost at port 5000.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported both individual and batch predictions, making the API versatile for different use cases.
- Provided logits, probabilities, and risk categories as outputs for interpretable results.
- Ensured input validation and error handling for robust API performance.
1.3.1.2 API Testing ¶
- The API code developed using the Flask framework deploying a categorical classification model was successfully tested with results presented as follows:
- Server Initialization: Flask application started successfully, serving the 'categorical_classification_flaskapi' application and its documentation.
- Debug Mode Disabled: The application ran with Debug mode: off, indicating a production-like environment but still using Flask’s built-in server.
- Development Server Warning: A warning advised against using Flask’s built-in development server in production and recommended a WSGI server instead.
- Server Running on Local and Network Addresses: The API was accessible at http://127.0.0.1:5000 (localhost) and http://192.168.6.49:5000 (local network), confirming successful binding to all available interfaces (0.0.0.0).
- Root Endpoint Accessed (GET /): A GET request at / returned 200 OK, verifying that the API service is running and accessible.
- Individual Prediction Request (POST /predict-individual-logit-probability-class): A POST request was successfully processed with 200 OK, indicating correct handling and response for a single test case.
- Batch Prediction Request (POST /predict-list-logit-probability-class): A POST request for batch classification was handled successfully, returning 200 OK, confirming multiple test cases were processed correctly.
- Invalid Input Handling (POST /predict-individual-logit-probability-class): A malformed request resulted in 400 Bad Request, demonstrating the API’s ability to catch and handle incorrect input formats.
##################################
# Loading Python Libraries
##################################
import requests
##################################
# Defining the base URL of the API
# for the categorical classification model
##################################
CC_FLASKAPI_BASE_URL = "http://127.0.0.1:5000/"
##################################
# Defining the input values for an individual test case
##################################
individual_test_case = {
"features_individual": [1, 0, 0, 0, 0, 1, 0, 0, 1, 1]
}
##################################
# Defining the input values for a batch of cases
##################################
batch_test_case = {
"features_list": [
[1, 0, 0, 0, 0, 1, 0, 0, 1, 1],
[1, 0, 1, 0, 1, 1, 0, 1, 1, 1]
]
}
##################################
# Generating a GET endpoint request for
# validating API service connection
##################################
response = requests.get(f"{CC_FLASKAPI_BASE_URL}/")
if response.status_code == 200:
print("Root Endpoint Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Root Endpoint Response: {'message': 'Welcome to the Categorical Classification API!'}
##################################
# Generating a POST endpoint request for
# computing the risk index,
# estimating the lung cancer probability,
# and predicting the risk category
# of an individual test case
##################################
response = requests.post(f"{CC_FLASKAPI_BASE_URL}/predict-individual-logit-probability-class", json=individual_test_case)
if response.status_code == 200:
print("Individual Test Case Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Individual Test Case Response: {'logit': -1.2117837409390746, 'probability': 0.22938559072691203, 'risk_class': 'Low-Risk'}
##################################
# Sending a POST endpoint request for
# computing the risk index,
# estimating the lung cancer probability,
# and predicting the risk category
# of a list of train cases
##################################
response = requests.post(f"{CC_FLASKAPI_BASE_URL}/predict-list-logit-probability-class", json=batch_test_case)
if response.status_code == 200:
print("Batch Test Case Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Batch Test Case Response: {'logit': [-1.2117837409390746, 3.4784950973590973], 'logit_sorted': [-1.2117837409390746, 3.4784950973590973], 'probability': [0.22938559072691203, 0.9700696569701589], 'probability_sorted': [0.22938559072691203, 0.9700696569701589]}
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
malformed_test_case = {"features": [1, 0, 1]}
response = requests.post(f"{CC_FLASKAPI_BASE_URL}/predict-individual-logit-probability-class", json=malformed_test_case)
if response.status_code == 200:
print("Malformed Test Case Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Error: 400 {"error":"'features_individual'"}
1.3.2 Survival Prediction ¶
1.3.2.1 API Building ¶
- An API code using the Flask framework was developed for deploying a survival prediction model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as Flask, request, and jsonify for API development.
- Included libraries for survival analysis (sksurv, lifelines), data manipulation (numpy, pandas), and visualization (matplotlib).
- Used io and base64 for encoding and handling image outputs.
- Integrated Flasgger for Swagger API documentation.
- Defining File Paths
- Specified the MODELS_PATH and PARAMETERS_PATH to locate the pre-trained survival model and related parameters.
- Loading the Pre-Trained Survival Model
- Loaded the pre-trained Cox Proportional Hazards (CoxPH) model (coxph_best_model.pkl) using joblib.load.
- Handled potential errors during model loading with a try-except block.
- Loading Model Parameters
- Loaded the median values for numeric features (numeric_feature_median_list.pkl) to support feature binning.
- Loaded the risk group threshold (coxph_best_model_risk_group_threshold.pkl) for categorizing patients into "High-Risk" and "Low-Risk" groups.
- Initializing the Flask App
- Created a Flask instance (app) to define and serve API endpoints.
- Integrated Swagger for automatic API documentation.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- Individual Survival Prediction Endpoint (/compute-individual-coxph-survival-probability-class/): A POST endpoint to generate survival profiles, estimate survival probabilities, and predict risk categories for individual test cases.
- Batch Survival Profile Endpoint (/compute-list-coxph-survival-profile/): A POST endpoint to generate survival profiles for a batch of cases.
- Feature Binning Endpoint (/bin-numeric-model-feature/): A POST endpoint to dichotomize numeric features based on the median.
- Kaplan-Meier Plot Endpoint (/plot-kaplan-meier/): A POST endpoint to generate and return Kaplan-Meier survival plots.
- Individual Survival Prediction Logic
- Extracted input data from the request body and converted it into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s predict_survival_function to generate the survival function for the test case.
- Predicted the risk category ("High-Risk" or "Low-Risk") based on the model’s risk score and threshold.
- Interpolated survival probabilities at predefined time points (e.g., 50, 100, 150, 200, 250 days).
- Batch Survival Profile Logic
- Extracted batch input data from the request body and converted it into a pandas DataFrame with appropriate feature names.
- Used the pre-trained model’s predict_survival_function to generate survival functions for all cases in the batch.
- Extracted and returned survival profiles for each case.
- Feature Binning Logic
- Extracted input data from the request body and converted it into a pandas DataFrame.
- Dichotomized the specified numeric feature into "Low" and "High" categories based on the median value.
- Returned the binned data as a list of dictionaries.
- Kaplan-Meier Plot Logic
- Extracted input data from the request body and converted it into a pandas DataFrame.
- Initialized a KaplanMeierFitter object to estimate survival probabilities.
- Plotted survival curves for different categories of the specified variable (e.g., "Low" vs. "High").
- Included an optional new case value for comparison in the plot.
- Saved the plot as a base64-encoded image and returned it in the API response.
- Error Handling
- Implemented robust error handling for invalid inputs or prediction errors using jsonify.
- Returned meaningful error messages and appropriate HTTP status codes (e.g., 400 for bad requests, 500 for server errors).
- Running the Flask App
- Used app.run to run the Flask app on localhost at port 5001.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported both individual and batch predictions, making the API versatile for different use cases.
- Provided survival probabilities, risk categories, and visualizations (Kaplan-Meier plots) for interpretable results.
- Enabled feature binning for categorical analysis of numeric features.
1.3.2.2 API Testing ¶
- The API code developed using the Flask framework deploying a survival prediction model was successfully tested with results presented as follows:
- Server Initialization: Flask application started successfully, serving the 'survival_prediction_flaskapi' application and its documentation.
- Debug Mode Disabled: The application ran with Debug mode: off, indicating a stable deployment configuration.
- Development Server Warning: A warning advised against using Flask’s built-in development server in production and recommended a WSGI server instead.
- Server Running on Local and Network Addresses: The API was accessible at http://127.0.0.1:5001 (localhost) and http://192.168.6.49:5001 (local network), confirming successful binding to all available interfaces (0.0.0.0).
- Root Endpoint Accessed (GET /): A GET request at / returned 200 OK, verifying that the API service was running and accessible.
- Individual Survival Prediction Request (POST /compute-individual-coxph-survival-probability-class/): A POST request was successfully processed with 200 OK, indicating correct handling and response for an individual survival probability estimation.
- Batch Survival Profile Request (POST /compute-list-coxph-survival-profile/): A POST request for batch survival profile estimation was handled successfully, returning 200 OK, confirming multiple test cases were processed correctly.
- Numeric Feature Binning Request (POST /bin-numeric-model-feature/): A POST request was successfully processed with 200 OK, confirming correct handling of numeric feature dichotomization.
- Kaplan-Meier Plot Request (POST /plot-kaplan-meier/): A POST request was successfully processed with 200 OK, returning a Kaplan-Meier survival curve plot encoded in Base64 format.
- Invalid Input Handling (POST /compute-individual-coxph-survival-probability-class/): A malformed request resulted in 400 Bad Request, demonstrating the API’s ability to catch and handle incorrect input formats.
##################################
# Loading Python Libraries
##################################
import requests
import json
import pandas as pd
import base64
from IPython.display import Image, display
##################################
# Defining the base URL of the API
# for the survival prediction model
##################################
SP_FLASKAPI_BASE_URL = "http://127.0.0.1:5001"
##################################
# Defining the input values for an individual test case
##################################
single_test_case = {
"features_individual": [43, 0, 75, 1, 0.75, 100]
}
##################################
# Defining the input values for a batch of cases
##################################
train_list = {
"features_list": [
[43, 0, 75, 1, 0.75, 100],
[70, 1, 20, 1, 0.75, 100]
]
}
##################################
# Defining the input values for a batch of cases for binning request
##################################
bin_request = {
"X_original_list": [
{"AGE": -0.10, "EJECTION_FRACTION": -0.10, "SERUM_CREATININE": -0.10, "SERUM_SODIUM": -0.10},
{"AGE": 0.20, "EJECTION_FRACTION": 0.20, "SERUM_CREATININE": 0.20, "SERUM_SODIUM": 0.20},
{"AGE": 0.90, "EJECTION_FRACTION": 0.90, "SERUM_CREATININE": 0.90, "SERUM_SODIUM": 0.90}
],
"numeric_feature": "AGE"
}
##################################
# Defining the input values for a batch of cases for Kaplan-Meier plotting
##################################
km_request = {
"df": [
{"TIME": 0, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 25, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 50, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 100, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 125, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 150, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 175, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 200, "DEATH_EVENT": 0, "AGE": "Low"},
{"TIME": 225, "DEATH_EVENT": 1, "AGE": "Low"},
{"TIME": 250, "DEATH_EVENT": 1, "AGE": "Low"},
{"TIME": 0, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 25, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 50, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 100, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 125, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 150, "DEATH_EVENT": 0, "AGE": "High"},
{"TIME": 175, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 200, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 225, "DEATH_EVENT": 1, "AGE": "High"},
{"TIME": 250, "DEATH_EVENT": 1, "AGE": "High"},
],
"cat_var": "AGE",
"new_case_value": "Low"
}
##################################
# Generating a GET endpoint request for
# for validating API service connection
##################################
response = requests.get(f"{SP_FLASKAPI_BASE_URL}/")
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'message': 'Welcome to the Survival Prediction API!'}
##################################
# Sending a POST endpoint request for
# generating the heart failure survival profile,
# estimating the heart failure survival probabilities,
# and predicting the risk category
# of an individual test case
##################################
response = requests.post(f"{SP_FLASKAPI_BASE_URL}/compute-individual-coxph-survival-probability-class/", json=single_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'risk_category': 'Low-Risk', 'survival_function': [0.9973812917524568, 0.9920416812438736, 0.9893236791425079, 0.972381113071464, 0.9693179903073035, 0.9631930672135339, 0.9631930672135339, 0.9600469571766689, 0.9600469571766689, 0.9568596864927983, 0.9536305709158891, 0.9471625843882805, 0.93729581350105, 0.9338986486591409, 0.93048646553474, 0.9270645831787163, 0.9202445006124622, 0.9167715111530355, 0.9132845175345189, 0.9097550958520674, 0.9097550958520674, 0.9097550958520674, 0.9060810720432387, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.8985598696587259, 0.8985598696587259, 0.8985598696587259, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8901959645503091, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8764677174183527, 0.8764677174183527, 0.8764677174183527, 0.8764677174183527, 0.8709113650481243, 0.8709113650481243, 0.8652494086650531, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8593884303802698, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8459534502216807, 0.8389821875092403, 0.8319419786276306, 0.8246669811915435, 0.8099879066057215, 0.8099879066057215, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7337716342207724, 0.7337716342207724, 0.7337716342207724, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696, 0.7070184115456696], 'survival_probabilities': [90.97550958520674, 87.64677174183527, 84.59534502216806, 78.48178617845467, 70.70184115456696], 'survival_time': [50, 100, 150, 200, 250]}
##################################
# Sending a POST endpoint request for
# generating the heart failure survival profile and
# estimating the heart failure survival probabilities
# of a list of train cases
##################################
response = requests.post(f"{SP_FLASKAPI_BASE_URL}/compute-list-coxph-survival-profile/", json=train_list)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'survival_profiles': [[0.9973812917524568, 0.9920416812438736, 0.9893236791425079, 0.972381113071464, 0.9693179903073035, 0.9631930672135339, 0.9631930672135339, 0.9600469571766689, 0.9600469571766689, 0.9568596864927983, 0.9536305709158891, 0.9471625843882805, 0.93729581350105, 0.9338986486591409, 0.93048646553474, 0.9270645831787163, 0.9202445006124622, 0.9167715111530355, 0.9132845175345189, 0.9097550958520674, 0.9097550958520674, 0.9097550958520674, 0.9060810720432387, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.9024157452999795, 0.8985598696587259, 0.8985598696587259, 0.8985598696587259, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8945287485160898, 0.8901959645503091, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8812352215018253, 0.8764677174183526, 0.8764677174183526, 0.8764677174183526, 0.8764677174183526, 0.8709113650481243, 0.8709113650481243, 0.8652494086650531, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8593884303802697, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8528574859874233, 0.8459534502216807, 0.8389821875092403, 0.8319419786276306, 0.8246669811915435, 0.8099879066057215, 0.8099879066057215, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7943979200335176, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7848178617845467, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7741993572193384, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7555469848652164, 0.7337716342207724, 0.7337716342207724, 0.7337716342207724, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695, 0.7070184115456695], [0.9761144218801228, 0.928980888267716, 0.905777064852962, 0.7724242339590301, 0.7502787164583535, 0.7076872741961866, 0.7076872741961866, 0.6866593185026403, 0.6866593185026403, 0.6659260634219393, 0.6454915885099762, 0.6062342264686207, 0.5504405490863784, 0.5323184765768243, 0.5146536440658629, 0.49746533888245986, 0.464726413395405, 0.4488047347163327, 0.433309930422515, 0.4181140392975694, 0.4181140392975694, 0.4181140392975694, 0.4028020116306455, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.38802639234746467, 0.373006008573048, 0.373006008573048, 0.373006008573048, 0.35785929480690143, 0.35785929480690143, 0.35785929480690143, 0.35785929480690143, 0.3421927616040032, 0.3117176431598899, 0.3117176431598899, 0.3117176431598899, 0.3117176431598899, 0.29651072362871467, 0.29651072362871467, 0.29651072362871467, 0.29651072362871467, 0.2796248763802668, 0.2796248763802668, 0.2633052706162029, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.24731169874453887, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.23051507888001282, 0.213871875082776, 0.19816204676152543, 0.18334919195124752, 0.16908728217620728, 0.1432835564355214, 0.1432835564355214, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.119778195679268, 0.10710184631976834, 0.10710184631976834, 0.10710184631976834, 0.10710184631976834, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.09446095671842468, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.07544024472233593, 0.05761125190131533, 0.05761125190131533, 0.05761125190131533, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404, 0.040906392932898404]]}
##################################
# Sending a POST endpoint request for
# creating dichotomous bins for the numeric features
# of a list of train cases
##################################
response = requests.post(f"{SP_FLASKAPI_BASE_URL}/bin-numeric-model-feature/", json=bin_request)
if response.status_code == 200:
display("Response:", pd.DataFrame(response.json()))
else:
print("Error:", response.status_code, response.text)
'Response:'
AGE | EJECTION_FRACTION | SERUM_CREATININE | SERUM_SODIUM | |
---|---|---|---|---|
0 | Low | -0.1 | -0.1 | -0.1 |
1 | High | 0.2 | 0.2 | 0.2 |
2 | High | 0.9 | 0.9 | 0.9 |
##################################
# Sending a POST endpoint request for
# plotting the estimated survival profiles
# using Kaplan-Meier Plots
##################################
response = requests.post(f"{SP_FLASKAPI_BASE_URL}/plot-kaplan-meier/", json=km_request)
if response.status_code == 200:
plot_data = response.json()["plot"]
# Decoding and displaying the plot
img = base64.b64decode(plot_data)
with open("kaplan_meier_plot.png", "wb") as f:
f.write(img)
display(Image("kaplan_meier_plot.png"))
else:
print("Error:", response.status_code, response.text)
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
malformed_test_case = {"features": [43, 0, 75, 1, 0.75]}
response = requests.post(f"{SP_FLASKAPI_BASE_URL}/compute-individual-coxph-survival-probability-class/", json=malformed_test_case)
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
Error: 400 {"error":"Missing 'features_individual' in request"}
1.3.3 Image Classification ¶
1.3.3.1 API Building ¶
- An API code using the Flask framework was developed for deploying a CNN image classification model with the steps described as follows:
- Loading Python Libraries
- Imported necessary libraries such as Flask, request, jsonify, and send_file for API development.
- Included libraries for deep learning (tensorflow, keras), image processing (PIL, cv2), and visualization (matplotlib).
- Used io, os, random, math, and base64 for file handling, random seed setting, and encoding.
- Integrated Flasgger for Swagger API documentation.
- Setting Random Seed
- Defined a set_seed function to ensure reproducibility by setting random seeds for numpy, tensorflow, keras, and Python’s random module.
- Enabled deterministic operations in TensorFlow to ensure consistent results across runs.
- Defining File Paths
- Specified the MODELS_PATH to locate the pre-trained CNN model and other related files.
- Loading the Pre-Trained CNN Model
- Loaded the pre-trained CNN model (cdrbnr_complex_best_model.keras) using keras.models.load_model.
- Handled potential errors during model loading with a try-except block.
- Recreating the CNN Model Using Functional API
- Recreated the CNN model using Keras’ Functional API to access intermediate layers for Grad-CAM visualization.
- Defined the input layer and sequentially applied layers from the original model to reconstruct it.
- Compiling the Functional API Model
- Compiled the recreated model with the same loss function (categorical_crossentropy), optimizer (adam), and metrics (Recall) as the original model.
- Creating Gradient Models for Grad-CAM
- Created three gradient models (grad_model_first_conv2d, grad_model_second_conv2d, grad_model_third_conv2d) to extract outputs from the first, second, and third convolutional layers.
- Compiled these models with the same parameters as the main model.
- Initializing the Flask App
- Created a Flask instance (app) to define and serve API endpoints.
- Integrated Swagger for automatic API documentation.
- Defining API Endpoints
- Root Endpoint (/): A simple GET endpoint to validate API service connectivity.
- File Upload Test Endpoint (/test-file-upload/): A POST endpoint to test file upload functionality, returning file metadata (filename, content type, size).
- Image Classification Endpoint (/predict-image-category-class-probability/): A POST endpoint to predict the image category and estimate class probabilities for an uploaded JPEG image.
- Grad-CAM Visualization Endpoint (/visualize-image-gradcam/): A POST endpoint to generate and return Grad-CAM heatmaps for the uploaded image.
- Image Preprocessing
- Read and resized uploaded images to the required input size (227x227 pixels).
- Converted images to grayscale and normalized pixel values to the range (0, 1).
- Expanded image dimensions to match the model’s input shape.
- Model Inference
- Used the pre-trained CNN model to predict the image category and class probabilities.
- Mapped predicted class indices to human-readable labels (e.g., "No Tumor", "Pituitary", "Meningioma", "Glioma").
- Grad-CAM Heatmap Generation
- Defined a make_gradcam_heatmap function to compute Grad-CAM heatmaps using TensorFlow’s GradientTape.
- Generated heatmaps for the first, second, and third convolutional layers.
- Superimposed heatmaps onto the original image using a color map (turbo) and transparency.
- Visualization and Encoding
- Created a multi-panel plot to display Grad-CAM heatmaps for all three convolutional layers.
- Saved the plot to a buffer, encoded it as a base64 string, and returned it in the API response.
- Error Handling
- Implemented robust error handling for invalid file uploads (e.g., missing files, non-JPEG images).
- Used jsonify to return appropriate error messages and status codes (e.g., 400 for bad requests, 500 for server errors).
- Running the Flask App
- Used app.run to run the Flask app on localhost at port 5002.
- Loading Python Libraries
- Key features of the API code included the following:
- Supported JPEG image uploads and validated file types.
- Provided both classification results and visual explanations (Grad-CAM) for model predictions.
- Ensured reproducibility and deterministic behavior through random seed setting.
1.3.3.2 API Testing ¶
- The API code developed using the Flask framework deploying a CNN image classification model was successfully tested with results presented as follows:
- Server Initialization: Flask application started successfully, serving the 'image_classification_flaskapi' application and its documentation.
- Debug Mode Disabled: The application ran with Debug mode: off, indicating a production-like environment but still using Flask’s built-in server.
- Development Server Warning: A warning advised against using Flask’s built-in development server in production and recommends a WSGI server instead.
- Server Running on Local Address: The API was accessible at http://127.0.0.1:5002 (localhost), confirming successful binding to the local interface.
- Root Endpoint Accessed (GET /): A GET request at / returned 200 OK, verifying that the API service was running and accessible.
- File Upload Endpoint Accessed (POST /test-file-upload/): A POST request at /test-file-upload/ returns 200 OK, confirming correct handling of file upload metadata.
- Image Classification Endpoint Accessed (POST /predict-image-category-class-probability/): A POST request for image category class and probability prediction returned 200 OK, demonstrating accurate classification and probability estimation for the uploaded image.
- Grad-CAM Visualization Endpoint Accessed (POST /visualize-image-gradcam/): A POST request for image gradcam visualization returned 200 OK, showing successful generation of Grad-CAM visualizations for the uploaded image.
- Invalid File Upload Handling (POST /test-file-upload/): An invalid file upload resulted in 400 Bad Request, indicating proper error handling when incorrect file formats are uploaded.
##################################
# Loading Python Libraries
##################################
import requests
import json
import base64
from IPython.display import display
from PIL import Image
import matplotlib.pyplot as plt
import io
from tensorflow.keras.utils import load_img
import os
import mimetypes
##################################
# Defining the base URL of the API
# for the image classification model
##################################
IC_FLASKAPI_BASE_URL = "http://127.0.0.1:5002"
##################################
# Defining the file path for an individual test image
##################################
IMAGES_PATH = r"image_classification_study\images"
image_path = (os.path.join("..",IMAGES_PATH, "test_image.jpg"))
##################################
# Automatically determining the filename and content type
##################################
image_path_filename = os.path.basename(image_path)
image_path_content_type, _ = mimetypes.guess_type(image_path)
##################################
# Visualizing the individual test image
##################################
try:
image = Image.open(image_path)
print(f"Image File Path: {image_path}")
print(f"Image Format: {image.format}")
print(f"Image Size: {image.size}")
print(f"Image Mode: {image.mode}")
except Exception as e:
print(f"Error loading image: {e}")
plt.imshow(image)
plt.axis('off')
plt.title("Test Image")
plt.show()
Image File Path: ..\image_classification_study\images\test_image.jpg Image Format: JPEG Image Size: (215, 234) Image Mode: RGB
##################################
# Generating a GET endpoint request for
# for validating API service connection
##################################
response = requests.get(f"{IC_FLASKAPI_BASE_URL}/")
if response.status_code == 200:
display("Response:", response.json())
else:
print("Error:", response.status_code, response.text)
'Response:'
{'message': 'Welcome to the Image Classification API!'}
##################################
# Sending a POST endpoint request for
# ensuring that the file upload mechanism is working
# by returning the the file metadata
##################################
with open(image_path, "rb") as file:
files = {"file": (image_path_filename, file, image_path_content_type)}
response = requests.post(f"{IC_FLASKAPI_BASE_URL}/test-file-upload/", files=files)
if response.status_code == 200:
result = response.json()
print("File Upload Test Result:")
print(f"Filename: {result['filename']}")
print(f"Content Type: {result['content_type']}")
print(f"Size: {result['size']} bytes")
else:
print(f"Error: {response.status_code} - {response.text}")
File Upload Test Result: Filename: test_image.jpg Content Type: image/jpeg Size: 12103 bytes
##################################
# Sending a POST endpoint request for
# predicting the image category and
# estimating class probabilities
# of an individual test image
##################################
with open(image_path, "rb") as file:
files = {"file": ("image.jpg", file, "image/jpeg")}
response = requests.post(f"{IC_FLASKAPI_BASE_URL}/predict-image-category-class-probability/", files=files)
if response.status_code == 200:
result = response.json()
print("Prediction Result:")
print(f"Predicted Class: {result['predicted_class']}")
print("Probabilities:")
for cls, prob in result["probabilities"].items():
print(f"{cls}: {prob:.5f}%")
else:
print(f"Error: {response.status_code} - {response.text}")
Prediction Result: Predicted Class: Meningioma Probabilities: Glioma: 0.02788% Meningioma: 87.02222% No Tumor: 12.94989% Pituitary: 0.00002%
##################################
# Sending a POST endpoint request for
# formulating the gradient class activation map
# from the output of the first to third convolutional layers and
# and superimposing on the actual image
##################################
with open(image_path, "rb") as file:
files = {"file": ("image.jpg", file, "image/jpeg")}
response = requests.post(f"{IC_FLASKAPI_BASE_URL}/visualize-image-gradcam/", files=files)
if response.status_code == 200:
plot_data = response.json()["plot"]
# Decoding and displaying the plot
img = base64.b64decode(plot_data)
with open("image_gradcam_plot.png", "wb") as f:
f.write(img)
display(Image.open("image_gradcam_plot.png"))
else:
print(f"Error: {response.status_code} - {response.text}")
##################################
# Defining the file path for an individual test image
##################################
IMAGES_PATH = r"image_classification_study\images"
malformed_image_path = (os.path.join("..",IMAGES_PATH, "test_image.png"))
##################################
# Automatically determining the filename and content type
##################################
malformed_image_path_filename = os.path.basename(image_path)
malformed_image_path_content_type, _ = mimetypes.guess_type(malformed_image_path)
##################################
# Sending a POST endpoint request
# using malformed data to evaluate
# the API's error handling function
##################################
malformed_image_path = (os.path.join("..",IMAGES_PATH, "test_image.png"))
with open(malformed_image_path, "rb") as file:
files = {"file": (malformed_image_path_filename, file, malformed_image_path_content_type)}
response = requests.post(f"{IC_FLASKAPI_BASE_URL}/test-file-upload/", files=files)
if response.status_code == 200:
result = response.json()
print("File Upload Test Result:")
print(f"Filename: {result['filename']}")
print(f"Content Type: {result['content_type']}")
print(f"Size: {result['size']} bytes")
else:
print(f"Error: {response.status_code} - {response.text}")
Error: 400 - {"error":"File must be a JPEG image"}
1.4. Consolidated Findings ¶
- Pre-trained models covering categorical classification, survival prediction and image classification were separately loaded and integrated into both the FastAPI and Flask frameworks.
- RESTful API endpoints using these frameworks were developed for each project to process prediction logic, handle input requests, and return diverse outputs.
- The APIs were designed to handle a variety of input types, such as file uploads for image classification and structured data for survival and risk prediction.
- Output types included strings (e.g., risk categories), floats (e.g., probabilities), and visualizations (e.g., Grad-CAM heatmaps and Kaplan-Meier plots).
- Robust error handling was implemented across all projects to ensure graceful degradation and meaningful error messages. This included validating input types, checking for missing fields, and handling exceptions during model inference.
- Both FastAPI and Flask frameworks were observed to:
- be efficient for building APIs to serve machine learning models via HTTP endpoints.
- integrate well with Python-based machine learning and deep learning libraries like TensorFlow, Scikit-learn and Scikit-survival.
- load, preprocess, and run inference on machine and deep learning models.
- allow serving models as RESTful APIs, handling requests, and returning model predictions.
- handle GET and POST requests, JSON payloads, and return responses.
- Comparing both frameworks, observations were as follows:
- FastAPI can be leveraged for its asynchronous capabilities, built-in data validation with Pydantic, and automatic OpenAPI documentation that are ideal for high concurrency, real-time ML inference, or scalable APIs.
- Flask is simpler and more flexible but synchronous by default and required integration with Flasgger for Swagger documentation, making it better for small-scale applications with a lightweight setup or when performance isn’t a top priority.
2. Summary ¶
3. References ¶
- [Book] Building Machine Learning Powered Applications: Going From Idea to Product by Emmanuel Ameisen
- [Book] Designing Machine Learning Systems: An Iterative Process for Production-Ready Applications by Chip Huyen
- [Book] Machine Learning Bookcamp: Build a Portfolio of Real-Life Projects by Alexey Grigorev and Adam Newmark
- [Book] Building Machine Learning Pipelines: Automating Model Life Cycles with TensorFlow by Hannes Hapke and Catherine Nelson
- [Book] Hands-On APIs for AI and Data Science: Python Development with FastAPI by Ryan Day
- [Book] Managing Machine Learning Projects: From Design to Deployment by Simon Thompson
- [Book] Building Data Science Applications with FastAPI: Develop, Manage, and Deploy Efficient Machine Learning Applications with Python by François Voron
- [Book] Microservice APIs: Using Python, Flask, FastAPI, OpenAPI and More by Jose Haro Peralta
- [Book] Machine Learning Engineering with Python: Manage the Lifecycle of Machine Learning odels using MLOps with Practical Examples by Andrew McMahon
- [Book] Introducing MLOps: How to Scale Machine Learning in the Enterprise by Mark Treveil, Nicolas Omont, Clément Stenac, Kenji Lefevre, Du Phan, Joachim Zentici, Adrien Lavoillotte, Makoto Miyazaki and Lynn Heidmann
- [Book] Practical Python Backend Programming: Build Flask and FastAPI Applications, Asynchronous Programming, Containerization and Deploy Apps on Cloud by Tim Peters
- [Python Library API] NumPy by NumPy Team
- [Python Library API] pandas by Pandas Team
- [Python Library API] seaborn by Seaborn Team
- [Python Library API] matplotlib.pyplot by MatPlotLib Team
- [Python Library API] matplotlib.image by MatPlotLib Team
- [Python Library API] matplotlib.offsetbox by MatPlotLib Team
- [Python Library API] itertools by Python Team
- [Python Library API] operator by Python Team
- [Python Library API] sklearn.experimental by Scikit-Learn Team
- [Python Library API] sklearn.impute by Scikit-Learn Team
- [Python Library API] sklearn.linear_model by Scikit-Learn Team
- [Python Library API] sklearn.preprocessing by Scikit-Learn Team
- [Python Library API] scipy by SciPy Team
- [Python Library API] sklearn.tree by Scikit-Learn Team
- [Python Library API] sklearn.ensemble by Scikit-Learn Team
- [Python Library API] sklearn.svm by Scikit-Learn Team
- [Python Library API] sklearn.metrics by Scikit-Learn Team
- [Python Library API] sklearn.model_selection by Scikit-Learn Team
- [Python Library API] imblearn.over_sampling by Imbalanced-Learn Team
- [Python Library API] imblearn.under_sampling by Imbalanced-Learn Team
- [Python Library API] SciKit-Survival by SciKit-Survival Team
- [Python Library API] SciKit-Learn by SciKit-Learn Team
- [Python Library API] StatsModels by StatsModels Team
- [Python Library API] SciPy by SciPy Team
- [Python Library API] Lifelines by Lifelines Team
- [Python Library API] tensorflow by TensorFlow Team
- [Python Library API] keras by Keras Team
- [Python Library API] pil by Pillow Team
- [Python Library API] glob by glob Team
- [Python Library API] cv2 by OpenCV Team
- [Python Library API] os by os Team
- [Python Library API] random by random Team
- [Python Library API] keras.models by TensorFlow Team
- [Python Library API] keras.layers by TensorFlow Team
- [Python Library API] keras.wrappers by TensorFlow Team
- [Python Library API] keras.utils by TensorFlow Team
- [Python Library API] keras.optimizers by TensorFlow Team
- [Python Library API] keras.preprocessing.image by TensorFlow Team
- [Python Library API] keras.callbacks by TensorFlow Team
- [Python Library API] keras.metrics by TensorFlow Team
- [Python Library API] sklearn.metrics by Scikit-Learn Team
- [Python Library API] Streamlit by Streamlit Team
- [Python Library API] Streamlit Community Cloud by Streamlit Team
- [Article] ML - Deploy Machine Learning Models Using FastAPI by Dorian Machado (Medium)
- [Article] Deploying Machine Learning Models Using FastAPI by Kevin Njagi (Medium)
- [Article] Deploy Machine Learning API with FastAPI for Free by Aniket Maurya (Lightning.AI)
- [Article] How to Use FastAPI for Machine Learning by Cheuk Ting Ho (JetBrains.Com)
- [Article] Deploying and Hosting a Machine Learning Model with FastAPI and Heroku by Michael Herman (TestDriven.IO)
- [Article] A Practical Guide to Deploying Machine Learning Models by Bala Priya (MachineLearningMastery.Com)
- [Article] Using FastAPI to Deploy Machine Learning Models by Carl Handlin (Medium)
- [Article] How to Deploy a Machine Learning Model by Maarten Grootendorst (MaartenGrootendorst.Com)
- [Article] Accelerating Machine Learning Deployment: Unleashing the Power of FastAPI and Docker by Pratyush Khare (Medium)
- [Article] Containerize and Deploy ML Models with FastAPI & Docker by Hemachandran Dhinakaran (Medium)
- [Article] Quick Tutorial to Deploy Your ML models using FastAPI and Docker by Shreyansh Singh (GitHub)
- [Article] How to Deploying Machine Learning Models in Production by Umair Akram (Medium)
- [Article] Deploying a Machine Learning Model with FastAPI: A Comprehensive Guide by Muhammad Naveed Arshad (Medium)
- [Article] Deploy Machine Learning Model with REST API using FastAPI by Yusuf Berki Yazıcıoğlu (Medium)
- [Article] Deploying An ML Model With FastAPI — A Succinct Guide by Yash Prakash (Medium)
- [Article] How to Build a Machine Learning App with FastAPI: Dockerize and Deploy the FastAPI Application to Kubernetes by Bravin Wasike (Dev.TO)
- [Article] Building a Machine Learning Model API with Flask: A Step-by-Step Guide by Nilesh Shinde (Medium)
- [Article] Deploying Your Machine Learning Model as a REST API Using Flask by Emmanuel Oludare (Medium)
- [Article] Machine Learning Model Deployment on Heroku Using Flask by Charu Makhijani (Medium)
- [Article] Model Deployment using Flask by Ravindra Sharma (Medium)
- [Article] Deploy a Machine Learning Model using Flask: Step-By-Step by Claudio Sabato (CodeFather.Tech)
- [Article] How to Deploy a Machine Learning Model using Flask? by DataDance.AI Team (DataDance.AI)
- [Article] A Comprehensive Guide on Deploying Machine Learning Models with Flask by MachineLearningModels.Org Team (MachineLearningModels.Org)
- [Article] How to Deploy Machine Learning Models with Flask and Docker by Usama Malik (Medium)
- [Article] Deploying Machine Learning Models with Flask: A Step-by-Step Guide by Sukma Hanifa (Medium)
- [Article] Machine Learning Model Deployment on Heroku Using Flask by Charu Makhijani (Medium)
- [Article] Complete Guide on Model Deployment with Flask and Heroku by Tarek Ghanoum (Medium)
- [Article] Turning Machine Learning Models into APIs in Python by Sayak Paul (DataCamp)
- [Article] Machine Learning, Pipelines, Deployment and MLOps Tutorial by Moez Ali (DataCamp)
- [Video Tutorial] Machine Learning Models Deployment with Flask and Docker by Data Science Dojo (YouTube)
- [Video Tutorial] Deploy Machine Learning Model Flask by Stats Wire (YouTube)
- [Video Tutorial] Deploy Machine Learning Models with Flask | Using Render to host API and Get URL :Step-By-Step Guide by Prachet Shah (YouTube)
- [Video Tutorial] Deploy Machine Learning Model using Flask by Krish Naik (YouTube)
- [Video Tutorial] Deploy Your ML Model Using Flask Framework by MSFTImagine (YouTube)
- [Video Tutorial] Build a Machine Learning App From Scratch with Flask & Docker by Patrick Loeber (YouTube)
- [Video Tutorial] Deploying a Machine Learning Model to a Web with Flask and Python Anywhere by Prof. Phd. Manoel Gadi (YouTube)
- [Video Tutorial] End To End Machine Learning Project With Deployment Using Flask by Data Science Diaries (YouTube)
- [Video Tutorial] Publish ML Model as API or Web with Python Flask by Python ML Daily (YouTube)
- [Video Tutorial] Deploy a Machine Learning Model using Flask API to Heroku by Jackson Yuan (YouTube)
- [Video Tutorial] Deploying Machine Learning Model with FlaskAPI - CI/CD for ML Series by Anthony Soronnadi (YouTube)
- [Video Tutorial] Deploy ML model as Webservice | ML model deployment | Machine Learning | Data Magic by Data Magic (YouTube)
- [Video Tutorial] Deploying Machine Learning Model Using Flask by DataMites (YouTube)
- [Video Tutorial] ML Model Deployment With Flask On Heroku | How To Deploy Machine Learning Model With Flask | Edureka by Edureka (YouTube)
- [Video Tutorial] ML Model Deployment with Flask | Machine Learning & Data Science by Dan Bochman (YouTube)
- [Video Tutorial] How to Deploy ML Solutions with FastAPI, Docker, & AWS by Shaw Talebi (YouTube)
- [Video Tutorial] Deploy ML models with FastAPI, Docker, and Heroku | Tutorial by AssemblyAI (YouTube)
- [Video Tutorial] Machine Learning Model Deployment Using FastAPI by TheOyinbooke (YouTube)
- [Video Tutorial] Creating APIs For Machine Learning Models with FastAPI by NeuralNine (YouTube)
- [Video Tutorial] How To Deploy Machine Learning Models Using FastAPI-Deployment Of ML Models As API’s by Krish Naik (YouTube)
- [Video Tutorial] Machine Learning Model with FastAPI, Streamlit and Docker by CodeTricks (YouTube)
- [Video Tutorial] FastAPI Machine Learning Model Deployment | Python | FastAPI by Stats Wire (YouTube)
- [Video Tutorial] Deploying Machine Learning Models - Full Guide by NeuralNine (YouTube)
- [Video Tutorial] Model Deployment FAST API - Docker | Machine Learning Model Deployment pipeline | FastAPI VS Flask by 360DigiTMG (YouTube)
- [Video Tutorial] Build an AI app with FastAPI and Docker - Coding Tutorial with Tips by Patrick Loeber (YouTube)
- [Video Tutorial] Create a Deep Learning API with Python and FastAPI by DataQuest (YouTube)
- [Video Tutorial] Fast API Machine Learning Web App Tutorial + Deployment on Heroku by Greg Hogg (YouTube)
- [Course] Deeplearning.AI Machine Learning in Production by DeepLearning.AI Team (Coursera)
- [Course] IBM AI Workflow: Enterprise Model Deployment by IBM Team (Coursera)
- [Course] DataCamp Machine Learning Engineer Track by DataCamp Team (DataCamp)
- [Course] DataCamp Designing Machine Learning Workflows in Python by DataCamp Team (DataCamp)
- [Course] DataCamp Building APIs in Python by DataCamp Team (DataCamp)
from IPython.display import display, HTML
display(HTML("<style>.rendered_html { font-size: 15px; font-family: 'Trebuchet MS'; }</style>"))