Skip to content

toni-ramchandani/AINotesForTesting

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Note: For this tutorial, you will need an API key and tokens to evaluate closed-source models. Ensure you have:

  • An API key from the respective service provider (e.g., OpenAI).
  • Sufficient tokens or credits to make API calls and evaluate the models.

You can obtain an API key by signing up on the provider's platform and following their instructions for API access. Make sure to review the pricing and usage limits to manage your costs effectively.

Although we have covered a lot of open source examples as well, you need to practise with both Open source and closed source to learn well

How to Purchase OpenAI API Key Tokens and Create an API Key

Step 1: Create an OpenAI Account

  1. Visit the OpenAI website.
  2. Sign up for an account using your email address.

Step 2: Create an API Key

Step 3: Set Up Billing

Step 4: Review Usage and Limits

Further Readings

Numpy


Pandas


CUDA


Machine Learning


Deep Learning Neural Networks


Generative AI


LLM


Evaluating LLMs


Prompt engineering

-Prompt Engineering for Generative AI | Machine Learning - Google Developers. https://developers.google.com/machine-learning/resources/prompt-eng.


Vector databases:


Retrieval-Augmented Generation (RAG)


List of tools for testing ai ml apps

Tool Application
Giskard Automated testing for bias, fairness, robustness; integrates with CI/CD pipelines.
DeepXplore White-box testing of deep learning models; finds incorrect behaviors under different conditions.
SHAP Explaining model predictions by attributing feature importance.
CleverHans Adversarial example testing for model robustness.
ART (Adversarial Robustness Toolbox) Securing AI models against adversarial attacks and enhancing model robustness.
FOOLBOX Benchmarking adversarial robustness and model evaluation under attack scenarios.
LangChain Integrating LLMs and creating pipelines for generative model testing.
MLflow Tracking, comparing, and managing machine learning experiments and models.
Weights & Biases Monitoring and visualizing model performance, logging experiments, and tracking metrics.
Facets Visualizing and understanding dataset distributions and model performance.
TensorFlow Model Analysis Provides tools for evaluating models, including fairness and performance analysis.
DataRobot MLOps Operationalizing machine learning models, monitoring, and automated testing in production.
Deepchecks Open-source tool for data validation and model testing, including drift detection and fairness checks.
Great Expectations Data validation and pipeline quality checks for ensuring consistent input data during model training.
TruEra Automated testing, model debugging, and explainability; ensures fairness and improves model quality.
LakeFS Version control for data lakes, allowing for reproducible data pipelines in machine learning projects.
Feast Open-source feature store that helps manage, store, and serve features for ML models in production.
Alibi Detect Open-source library for drift detection, outlier detection, and model monitoring.
Seldon Core Machine learning deployment platform that supports monitoring and testing of production models.
Kubeflow Manages ML workflows, model training, and serves models; useful for testing AI/ML pipelines.
DeepMind’s Polycoder Testing AI code generation and ensuring the quality of ML-based code generation models.
H2O.ai Driverless AI Automated machine learning platform focusing on model explainability, bias detection, and fairness.
Neptune.ai Experiment tracking tool for collaborative development and monitoring AI models in production.
ExplainX.ai Focuses on model explainability and fairness, helping ensure transparent AI/ML decisions.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published