Slides from my talk at MLOps World' 21
Deploying AI models in production and scaling the ML services is still a big challenge. In this talk we will cover details of how to deploy your AI models, best practices for the deployment scenarios, and techniques for performance optimization and scaling the ML services. Come join us to learn how you can jumpstart the journey of taking your PyTorch models from Research to production.
AI and ML Series - Introduction to Generative AI and LLMs - Session 1DianaGray10
Session 1
👉This first session will cover an introduction to Generative AI & harnessing the power of large language models. The following topics will be discussed:
Introduction to Generative AI & harnessing the power of large language models.
What’s generative AI & what’s LLM.
How are we using it in our document understanding & communication mining models?
How to develop a trustworthy and unbiased AI model using LLM & GenAI.
Personal Intelligent Assistant
Speakers:
📌George Roth - AI Evangelist at UiPath
📌Sharon Palawandram - Senior Machine Learning Consultant @ Ashling Partners & UiPath MVP
📌Russel Alfeche - Technology Leader RPA @qBotica & UiPath MVP
Managing the Machine Learning Lifecycle with MLflowDatabricks
ML development brings many new complexities beyond the traditional software development lifecycle. MLflow is an open-source project from Databricks aiming to solve some of these challenges such as experiment tracking, reproducibility, model packaging, deployment, and governance, in order to manage and accelerate the lifecycle of your ML projects.
Developing a Knowledge Graph of your Competency, Skills, and Knowledge at NASANeo4j
The document discusses NASA's efforts to create a knowledge graph to map workforce skills. It explains that NASA is asking questions like what are the current workforce skill sets, how skills can be grouped, and how individuals can identify new skills or careers. The document outlines NASA's process for developing the knowledge graph which includes understanding the domain, defining a model, gathering data from sources like O*NET, and analyzing the data to create graphs that can compare occupations and recommend career changes or training. The knowledge graph is intended to help with workforce planning, identifying expertise, and outlining career paths.
MLOps Bridging the gap between Data Scientists and Ops.Knoldus Inc.
Through this session we're going to introduce the MLOps lifecycle and discuss the hidden loopholes that can affect the MLProject. Then we are going to discuss the ML Model lifecycle and discuss the problem with training. We're going to introduce the MLFlow Tracking module in order to track the experiments.
MLflow is an MLOps tool that enables data scientist to quickly productionize their Machine Learning projects. To achieve this, MLFlow has four major components which are Tracking, Projects, Models, and Registry. MLflow lets you train, reuse, and deploy models with any library and package them into reproducible steps. MLflow is designed to work with any machine learning library and require minimal changes to integrate into an existing codebase. In this session, we will cover the common pain points of machine learning developers such as tracking experiments, reproducibility, deployment tool and model versioning. Ready to get your hands dirty by doing quick ML project using mlflow and release to production to understand the ML-Ops lifecycle.
Use MLflow to manage and deploy Machine Learning model on Spark Herman Wu
MLflow is a platform for managing the machine learning lifecycle, including tracking experiments, packaging models into reproducible projects, and deploying models into production. It provides APIs and tools to help with experiment tracking, model deployment, and model monitoring. MLflow supports many frameworks like PyTorch, TensorFlow, Keras, and Spark ML. It integrates with cloud platforms like Azure ML, AWS SageMaker, and Databricks to enable deployment and management of models on those platforms.
This workshop presentation from Enterprise Knowledge team members Joe Hilger, Founder and COO, and Sara Nash, Technical Analyst, was delivered on June 8, 2020 as part of the Data Summit 2020 virtual conference. The 3-hour workshop provided an interdisciplinary group of participants with a definition of what a knowledge graph is, how it is implemented, and how it can be used to increase the value of your organization’s datas. This slide deck gives an overview of the KM concepts that are necessary for the implementation of knowledge graphs as a foundation for Enterprise Artificial Intelligence (AI). Hilger and Nash also outlined four use cases for knowledge graphs, including recommendation engines and natural language query on structured data.
Data Scientists and Machine Learning practitioners, nowadays, seem to be churning out models by the dozen and they continuously experiment to find ways to improve their accuracies. They also use a variety of ML and DL frameworks & languages , and a typical organization may find that this results in a heterogenous, complicated bunch of assets that require different types of runtimes, resources and sometimes even specialized compute to operate efficiently.
But what does it mean for an enterprise to actually take these models to "production" ? How does an organization scale inference engines out & make them available for real-time applications without significant latencies ? There needs to be different techniques for batch (offline) inferences and instant, online scoring. Data needs to be accessed from various sources and cleansing, transformations of data needs to be enabled prior to any predictions. In many cases, there maybe no substitute for customized data handling with scripting either.
Enterprises also require additional auditing and authorizations built in, approval processes and still support a "continuous delivery" paradigm whereby a data scientist can enable insights faster. Not all models are created equal, nor are consumers of a model - so enterprises require both metering and allocation of compute resources for SLAs.
In this session, we will take a look at how machine learning is operationalized in IBM Data Science Experience (DSX), a Kubernetes based offering for the Private Cloud and optimized for the HortonWorks Hadoop Data Platform. DSX essentially brings in typical software engineering development practices to Data Science, organizing the dev->test->production for machine learning assets in much the same way as typical software deployments. We will also see what it means to deploy, monitor accuracies and even rollback models & custom scorers as well as how API based techniques enable consuming business processes and applications to remain relatively stable amidst all the chaos.
Speaker
Piotr Mierzejewski, Program Director Development IBM DSX Local, IBM
Episode 2: The LLM / GPT / AI Prompt / Data Engineer RoadmapAnant Corporation
In this episode we'll discuss the different flavors of prompt engineering in the LLM/GPT space. According to your skill level you should be able to pick up at any of the following:
Leveling up with GPT
1: Use ChatGPT / GPT Powered Apps
2: Become a Prompt Engineer on ChatGPT/GPT
3: Use GPT API with NoCode Automation, App Builders
4: Create Workflows to Automate Tasks with NoCode
5: Use GPT API with Code, make your own APIs
6: Create Workflows to Automate Tasks with Code
7: Use GPT API with your Data / a Framework
8: Use GPT API with your Data / a Framework to Make your own APIs
9: Create Workflows to Automate Tasks with your Data /a Framework
10: Use Another LLM API other than GPT (Cohere, HuggingFace)
11: Use open source LLM models on your computer
12: Finetune / Build your own models
Series: Using AI / ChatGPT at Work - GPT Automation
Are you a small business owner or web developer interested in leveraging the power of GPT (Generative Pretrained Transformer) technology to enhance your business processes?
If so, Join us for a series of events focused on using GPT in business. Whether you're a small business owner or a web developer, you'll learn how to leverage GPT to improve your workflow and provide better services to your customers.
The catalyst for the success of automobiles came not through the invention of the car but rather through the establishment of an innovative assembly line. History shows us that the ability to mass produce and distribute a product is the key to driving adoption of any innovation, and machine learning is no different. MLOps is the assembly line of Machine Learning and in this presentation we will discuss the core capabilities your organization should be focused on to implement a successful MLOps system.
Emeli Dral (Evidently AI) – Analyze it: production monitoring for machine lea...Codiax
Machine learning models in production are susceptible to failures from data and concept drift issues that can degrade model performance over time. Effective monitoring of model health and data quality is needed to detect such issues. Key aspects to monitor include model accuracy, data schema changes, data distribution shifts, broken data pipelines, and the impact of concept drift. Starting simply by adding basic machine learning metrics to existing service monitoring for memory, latency, and uptime is pragmatic. The appropriate monitoring approach depends on factors like available resources, use case importance, and system complexity.
Simplifying Model Management with MLflowDatabricks
<p>Last summer, Databricks launched MLflow, an open source platform to manage the machine learning lifecycle, including experiment tracking, reproducible runs and model packaging. MLflow has grown quickly since then, with over 120 contributors from dozens of companies, including major contributions from R Studio and Microsoft. It has also gained new capabilities such as automatic logging from TensorFlow and Keras, Kubernetes integrations, and a high-level Java API. In this talk, we’ll cover some of the new features that have come to MLflow, and then focus on a major upcoming feature: model management with the MLflow Model Registry. Many organizations face challenges tracking which models are available in the organization and which ones are in production. The MLflow Model Registry provides a centralized database to keep track of these models, share and describe new model versions, and deploy the latest version of a model through APIs. We’ll demonstrate how these features can simplify common ML lifecycle tasks.</p>
Knowledge Graphs and Generative AI_GraphSummit Minneapolis Sept 20.pptxNeo4j
This document discusses using knowledge graphs to ground large language models (LLMs) and improve their abilities. It begins with an overview of generative AI and LLMs, noting their opportunities but also challenges like lack of knowledge and inability to verify sources. The document then proposes using a knowledge graph like Neo4j to provide context and ground LLMs, describing how graphs can be enriched with algorithms, embeddings and other data. Finally, it demonstrates how contextual searches and responses can be improved by retrieving relevant information from the knowledge graph to augment LLM responses.
This document discusses metadata and the importance of metadata management. It introduces Apache Atlas as an open source platform for metadata management and governance. Key points include:
- Metadata is important for data reuse, analytics, and governance. It provides context and meaning about data.
- Current reality is that metadata is often not well supported or integrated across tools. Apache Atlas aims to provide an open, unified approach.
- Apache Atlas has graduated to a top-level Apache project. It provides a type-agnostic metadata store and interfaces that can be accessed by various tools.
- The vision is for an open ecosystem where metadata is shared and federated across repositories from different vendors and tools.
Why do the majority of Data Science projects never make it to production?Itai Yaffe
María de la Fuente (Solutions Architect Manager for IMEA) @ Databricks
While most companies understand the value creation of leveraging data and are taking on board an AI strategy, only 13% of the data science projects make it to production successfully.
Besides the well-known skills gap in the market, we need to level up our end-to-end approach and cover all aspects involved when working with AI.
In this session, we will discuss the main obstacles to overcome and how we can avoid the major pitfalls to ensure our data science journey becomes successful.
Apache Spark is a fast and general engine for large-scale data processing. It was created by UC Berkeley and is now the dominant framework in big data. Spark can run programs over 100x faster than Hadoop in memory, or more than 10x faster on disk. It supports Scala, Java, Python, and R. Databricks provides a Spark platform on Azure that is optimized for performance and integrates tightly with other Azure services. Key benefits of Databricks on Azure include security, ease of use, data access, high performance, and the ability to solve complex analytics problems.
AI as a Service, Build Shared AI Service Platforms Based on Deep Learning Tec...Databricks
I will share the vision and the production journey of how we build enterprise shared AI As A Service platforms with distributed deep learning technologies. Including those topics:
1) The vision of Enterprise Shared AI As A Service and typical AI services use cases at FinTech industry
2) The high level architecture design principles for AI As A Service
3) The technical evaluation journey to choose an enterprise deep learning framework with comparisons, such as why we choose Deep learning framework based on Spark ecosystem
4) Share some production AI use cases, such as how we implemented new Users-Items Propensity Models with deep learning algorithms with Spark,improve the quality , performance and accuracy of offer and campaigns design, targeting offer matching and linking etc.
5) Share some experiences and tips of using deep learning technologies on top of Spark , such as how we conduct Intel BigDL into a real production.
TensorFlow meetup: Keras - Pytorch - TensorFlow.jsStijn Decubber
Slides from the TensorFlow meetup hosted on October 9th at the ML6 offices in Ghent. Join our Meetup group for updates and future sessions: https://www.meetup.com/TensorFlow-Belgium/
Scaling Up AI Research to Production with PyTorch and MLFlowDatabricks
PyTorch, the popular open-source ML framework, has continued to evolve rapidly since the introduction of PyTorch 1.0, which brought an accelerated workflow from research to production.
GDG Cloud Southlake #16: Priyanka Vergadia: Scalable Data Analytics in Google...James Anderson
Do you know The Cloud Girl? She makes the cloud come alive with pictures and storytelling.
The Cloud Girl, Priyanka Vergadia, Chief Content Officer @Google, joins us to tell us about Scaleable Data Analytics in Google Cloud.
Maybe, with her explanation, we'll finally understand it!
Priyanka is a technical storyteller and content creator who has created over 300 videos, articles, podcasts, courses and tutorials which help developers learn Google Cloud fundamentals, solve their business challenges and pass certifications! Checkout her content on Google Cloud Tech Youtube channel.
Priyanka enjoys drawing and painting which she tries to bring to her advocacy.
Check out her website The Cloud Girl: https://thecloudgirl.dev/ and her new book: https://www.amazon.com/Visualizing-Google-Cloud-Illustrated-References/dp/1119816327
Since its beginning, the Performance Advisory Council aims to promote engagement between various experts from around the world, to create relevant, value-added content sharing between members. For Neotys, to strengthen our position as a thought leader in load & performance testing. During this event, 12 participants convened in Chamonix (France) exploring several topics on the minds of today’s performance tester such as DevOps, Shift Left/Right, Test Automation, Blockchain and Artificial Intelligence.
Reproducible AI using MLflow and PyTorchDatabricks
Model reproducibility is becoming the next frontier for successful AI models building and deployments for both Research and Production scenarios. In this talk, we will show you how to build reproducible AI models and workflows using PyTorch and MLflow that can be shared across your teams, with traceability and speed up collaboration for AI projects.
Artem Koval presented on cloud-native MLOps frameworks. MLOps is a process for deploying and monitoring machine learning models through continuous integration and delivery. It addresses fairness, explainability, model monitoring, and human intervention. Modern MLOps frameworks focus on these areas as well as data labeling, testing, and observability. Different levels of MLOps are needed depending on an organization's size, from lightweight for small teams to enterprise-level for large companies with many models. Human-centered AI should be incorporated at all levels by involving humans throughout the entire machine learning process.
Revolutionary container based hybrid cloud solution for MLPlatform
Ness' data science platform, NextGenML, puts the entire machine learning process: modelling, execution and deployment in the hands of data science teams.
The entire paradigm approaches collaboration around AI/ML, being implemented with full respect for best practices and commitment to innovation.
Kubernetes (onPrem) + Docker, Azure Kubernetes Cluster (AKS), Nexus, Azure Container Registry(ACR), GlusterFS
Workflow
Argo->Kubeflow
DevOps
Helm, kSonnet, Kustomize,Azure DevOps
Code Management & CI/CD
Git, TeamCity, SonarQube, Jenkins
Security
MS Active Directory, Azure VPN, Dex (K8s) integrated with GitLab
Machine Learning
TensorFlow (model training, boarding, serving), Keras, Seldon
Storage (Azure)
Storage Gen1 & Gen2, Data Lake, File Storage
ETL (Azure)
Databricks, Spark on K8, Data Factory (ADF), HDInsight (Kafka and Spark), Service Bus (ASB)
Lambda functions & VMs, Cache for Redis
Monitoring and Logging
Graphana, Prometeus, GrayLog
The document provides an overview of parallel development and Microsoft's investments in parallel computing technologies. It discusses the difficulty of writing parallel code and introduces some of Microsoft's tools and APIs to help developers write parallel and concurrent applications more easily, including the Task Parallel Library (TPL) and Parallel LINQ (PLINQ). It encourages developers to experiment with and provide feedback on these new parallel programming models and tools.
Automated ML Workflow for Distributed Big Data Using Analytics Zoo (CVPR2020 ...Jason Dai
This document summarizes a CVPR 2020 tutorial on the Analytics Zoo platform for automated machine learning workflows for distributed big data using Apache Spark. The tutorial covers an overview of Analytics Zoo and the BigDL distributed deep learning framework. It demonstrates distributed training of deep learning models using TensorFlow and PyTorch on Spark, and features of Analytics Zoo like end-to-end pipelines, ML workflow for automation, and model deployment with cluster serving. Real-world use cases applying Analytics Zoo at companies like SK Telecom, Midea, and MasterCard are also presented.
This document discusses moving machine learning models from prototype to production. It outlines some common problems with the current workflow where moving to production often requires redevelopment from scratch. Some proposed solutions include using notebooks as APIs and developing analytics that are accessed via an API. It also discusses different data science platforms and architectures for building end-to-end machine learning systems, focusing on flexibility, security, testing and scalability for production environments. The document recommends a custom backend integrated with Spark via APIs as the best approach for the current project.
DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergenceinside-BigData.com
In this deck, Johann Lombardi from Intel presents: DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergence.
"Intel has been building an entirely open source software ecosystem for data-centric computing, fully optimized for Intel® architecture and non-volatile memory (NVM) technologies, including Intel Optane DC persistent memory and Intel Optane DC SSDs. Distributed Asynchronous Object Storage (DAOS) is the foundation of the Intel exascale storage stack. DAOS is an open source software-defined scale-out object store that provides high bandwidth, low latency, and high I/O operations per second (IOPS) storage containers to HPC applications. It enables next-generation data-centric workflows that combine simulation, data analytics, and AI."
Unlike traditional storage stacks that were primarily designed for rotating media, DAOS is architected from the ground up to make use of new NVM technologies, and it is extremely lightweight because it operates end-to-end in user space with full operating system bypass. DAOS offers a shift away from an I/O model designed for block-based, high-latency storage to one that inherently supports fine- grained data access and unlocks the performance of next- generation storage technologies.
Watch the video: https://youtu.be/wnGBW31yhLM
Learn more: https://www.intel.com/content/www/us/en/high-performance-computing/daos-high-performance-storage-brief.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Helixa uses serverless machine learning architectures to power an audience intelligence platform. It ingests large datasets and uses machine learning models to provide insights. Helixa's machine learning system is built on AWS serverless services like Lambda, Glue, Athena and S3. It features a data lake for storage, a feature store for preprocessed data, and uses techniques like map-reduce to parallelize tasks. Helixa aims to build scalable and cost-effective machine learning pipelines without having to manage servers.
The document describes a Bucharest Big Data Meetup occurring on June 5th. The meetup will include two tech talks: one on productionizing machine learning from 7:00-7:40 PM, and another on a technology comparison of databases vs blockchains from 7:40-8:15 PM. The meetup will conclude from 8:15-8:45 PM with pizza and drinks sponsored by Netopia.
Introducing MlFlow: An Open Source Platform for the Machine Learning Lifecycl...DataWorks Summit
Specialized tools for machine learning development and model governance are becoming essential. MlFlow is an open source platform for managing the machine learning lifecycle. Just by adding a few lines of code in the function or script that trains their model, data scientists can log parameters, metrics, artifacts (plots, miscellaneous files, etc.) and a deployable packaging of the ML model. Every time that function or script is run, the results will be logged automatically as a byproduct of those lines of code being added, even if the party doing the training run makes no special effort to record the results. MLflow application programming interfaces (APIs) are available for the Python, R and Java programming languages, and MLflow sports a language-agnostic REST API as well. Over a relatively short time period, MLflow has garnered more than 3,300 stars on GitHub , almost 500,000 monthly downloads and 80 contributors from more than 40 companies. Most significantly, more than 200 companies are now using MLflow. We will demo MlFlow Tracking , Project and Model components with Azure Machine Learning (AML) Services and show you how easy it is to get started with MlFlow on-prem or in the cloud.
Presented at IDEAS SoCal on Oct 20, 2018. I discuss main approaches of deploying data science engines to production and provide sample code for the comprehensive approach of real time scoring with MLeap and Spark ML.
Big Data for Testing - Heading for Post Process and AnalyticsOPNFV
Yujun Zhang, ZTE Corporation, Donald Hunter, Cisco, Trevor Cooper, Intel
The testing community created tens of testing projects, hundreds of testing cases, thousands of testing jobs. Huge amount of testing data has been produced. What comes next, then?
The testing community puts in place tools and procedures to declare testcases/projects, normalize and upload results. These tools and procedures have been adopted so we now have lots of data covering lots of scenarios, hardware, installers.
In this presentation, we shall discuss the stakes and challenges of result post processing.
* How analytics can provide valuable inputs to the community, end users or upstream projects.
* How can we produce accurate indicators, reports and graphs, focus on interpreting / consuming test results.
* How can we get the best of breeds of our result mine?
The document outlines Neo4j's product strategy and roadmap. It discusses trends like increasing cloud adoption and the blending of transactional and analytical use cases. The roadmap focuses on cloud-first capabilities, ease of use for developers, trusted fundamentals of the database, and enabling AI through graph algorithms and knowledge graphs. Key announcements include new graph algorithms, change data capture for integration, autonomous clustering for scalability, and innovations in graph embeddings and generative AI integration.
Data Agility—A Journey to Advanced Analytics and Machine Learning at ScaleDatabricks
Hari Subramanian presented on Uber's journey to enable data agility and advanced analytics at scale. He discussed Uber's large and growing data platform that processes millions of daily trips and terabytes of data. He then described Uber's Data Science Workbench, which aims to democratize data science by providing self-service access to infrastructure, tools, and data to support various users from data scientists to business analysts. Finally, he presented a case study on COTA, a deep learning model for customer support ticketing that was developed and deployed using Uber's data platform and workflow.
Similar to Scaling AI in production using PyTorch (20)
Profiling PyTorch for Efficiency & Sustainabilitygeetachauhan
From my talk at the Data & AI summit - latest update on the PyTorch Profiler and how you can use it for optimizations for efficiency. Talk also dives into the future and what we need to do together as an industry to move towards Sustainable AI
Building AI with Security Privacy in Mindgeetachauhan
The document discusses building AI with security and privacy in mind. It covers privacy challenges in AI like tensions between data privacy and model training. It then discusses various privacy preserving machine learning techniques like homomorphic encryption, differential privacy, secure multi-party computation, on-device computation, and federated learning. The document provides examples of how each technique works. It concludes by discussing tools and techniques for starting a privacy journey in AI and provides resources to learn more.
Building AI with Security and Privacy in mindgeetachauhan
The document discusses building AI with security and privacy in mind. It covers privacy challenges in AI like tensions between data privacy and model training. It then discusses various privacy preserving machine learning techniques like homomorphic encryption, differential privacy, secure multi-party computation, on-device computation, and federated learning. The document provides examples of how each technique works. It concludes by discussing tools and techniques for starting a privacy journey in AI and provides resources to learn more.
Building Interpretable & Secure AI Systems using PyTorchgeetachauhan
Slides from my talk at Deep Learning World 2020. The talk covered use cases, special challenges and solutions for building Interpretable and Secure AI systems using Pytorch.
- Tools for building Interpretable models
- How to build secure, privacy preserving AI models with Pytorch
- Use cases and insights from the field
Slides from Talk @ Intel IoT DevFest IV
With both Facebook and Google's recent shift in direction towards a "Future is Private" world, learn how you too can train and deploy your AI models in a privacy-preserving way, with Decentralized AI and a combination of AI and Blockchain. These techniques will become even more rampant as we move into a world where users will own their own data and companies will start using “ethically sourced data” and move towards a path for Ethical AI for the IoT space.
In this session, you will learn:
- Use cases for Decentralized AI, with combined benefits of AI + Blockchain for IoT applications
- Federated learning & related privacy-preserving AI model training techniques for IoT applications
- How to build Ethical AI solutions for IoT using these techniques
Draper Accelerator Talk Slides - convering convergence of of AI and Blockchain and how it solves challenges for IoT, Ai@Edge and Data Ethics and User Data Monetization.
Decentralized AI: Convergence of AI + Blockchain geetachauhan
Santa Clara IoT Expo talk slides - convering convergence of of AI and Blockchain and how it solves challenges for IoT, Ai@Edge and Data Ethics and User Data Monetization
Decentralized AI: Convergence of Blockchain + AIgeetachauhan
This document discusses the convergence of blockchain and AI through decentralized AI approaches. It outlines challenges with centralized AI models regarding privacy, influence, economics and transparency. Decentralized solutions proposed include federated learning, blockchain, homomorphic encryption, and data marketplaces. Blockchain provides an open, trustless network to replace centralized authorities and enable applications like data exchanges, AI marketplaces and distributed machine learning across devices. Overall the goal is to democratize AI and data through user ownership and control.
Decentralized AI: Convergence of Blockchain + AIgeetachauhan
As we move into the world where User's will own their own data, and companies will use "Ethically Sourced Data", there will be a rampant need for Decentralized AI. And, combining with Blockchain one gets viable Business Models. This talk covers use cases for convergence of Blockchain and AI.
Talk @ ACM SF Bayarea Chapter on Deep Learning for medical imaging space.
The talk covers use cases, special challenges and solutions for Deep Learning for Medical Image Analysis using Tensorflow+Keras. You will learn about:
- Use cases for Deep Learning in Medical Image Analysis
- Different DNN architectures used for Medical Image Analysis
- Special purpose compute / accelerators for Deep Learning (in the Cloud / On-prem)
- How to parallelize your models for faster training of models and serving for inferenceing.
- Optimization techniques to get the best performance from your cluster (like Kubernetes/ Apache Mesos / Spark)
- How to build an efficient Data Pipeline for Medical Image Analysis using Deep Learning
- Resources to jump start your journey - like public data sets, common models used in Medical Image Analysis
The document discusses deep learning techniques for financial technology (FinTech) applications. It begins with examples of current deep learning uses in FinTech like trading algorithms, fraud detection, and personal finance assistants. It then covers topics like specialized compute hardware for deep learning training and inference, optimization techniques for CPUs and GPUs, and distributed training approaches. Finally, it discusses emerging areas like FPGA and quantum computing and provides resources for practitioners to start with deep learning for FinTech.
NIPS - Deep learning @ Edge using Intel's NCSgeetachauhan
The document discusses using Intel's Neural Compute Stick for deep learning at the edge. It introduces the Neural Compute Stick, which enables computer vision and AI capabilities in small, low power devices. It then provides an overview of deep learning and discusses how to build IoT applications using the Neural Compute Stick SDK. Examples of use cases for edge intelligence in IoT are also presented.
Best Practices for On-Demand HPC in Enterprisesgeetachauhan
Traditionally HPC has been popular in Scientific domains, but not in most other Enterprises. With the advent of on-demand-HPC in cloud and growing adoption of Deep Learning, HPC should now be a standard platform for any Enterprise leading with AI and Machine Learning. This session will cover the best practices for building your own on-demand HPC cluster for Enterprise workloads along with key use cases where Enterprises will benefit from HPC solution.
Deep learning @ Edge using Intel's Neural Compute Stickgeetachauhan
Talk @ Intel Global IoT DevFest, Nov 2017
The new generation of hardware accelerators are enabling rich AI driven, Intelligent IoT solutions @ the edge.
The talk showcased how to use Intel's latest Nervana Compute Stick for accelerating deep learning IoT solutions. It also covered use cases and code details for running Deep Learning models on Intel's Nervana Compute Stick.
Distributed deep learning optimizations - AI WithTheBestgeetachauhan
Learn how to optimize Tensorflow for your Intel CPU and techniques for distributed deep learning for both model training and inferencing. Talk @ AI WithTheBest
Distributed deep learning optimizationsgeetachauhan
The document discusses optimizations for distributed deep learning. It covers challenges like latency, cost and power consumption when scaling deep learning models. It then discusses specialized compute like Google TPUs and optimizations for CPU, GPU and inference workloads. Techniques like data parallelism, model parallelism, quantization and clustering are presented. Emerging areas like FPGA, neuromorphic and quantum computing are also mentioned.
Intel optimized tensorflow, distributed deep learninggeetachauhan
This document discusses optimizations for running TensorFlow on Intel CPUs for deep learning. It outlines techniques for compiling TensorFlow from source with CPU optimizations, using proper data formats and batch sizes, and reading data with queues to leverage multi-core CPUs. It also covers distributed deep learning using TensorFlow Estimators, parameter servers, and model parallelism to distribute graphs across multiple machines. Resources for further information on Intel optimizations, installing libraries, and distributed TensorFlow are provided.
Navigating Post-Quantum Blockchain: Resilient Cryptography in Quantum Threatsanupriti
In the rapidly evolving landscape of blockchain technology, the advent of quantum computing poses unprecedented challenges to traditional cryptographic methods. As quantum computing capabilities advance, the vulnerabilities of current cryptographic standards become increasingly apparent.
This presentation, "Navigating Post-Quantum Blockchain: Resilient Cryptography in Quantum Threats," explores the intersection of blockchain technology and quantum computing. It delves into the urgent need for resilient cryptographic solutions that can withstand the computational power of quantum adversaries.
Key topics covered include:
An overview of quantum computing and its implications for blockchain security.
Current cryptographic standards and their vulnerabilities in the face of quantum threats.
Emerging post-quantum cryptographic algorithms and their applicability to blockchain systems.
Case studies and real-world implications of quantum-resistant blockchain implementations.
Strategies for integrating post-quantum cryptography into existing blockchain frameworks.
Join us as we navigate the complexities of securing blockchain networks in a quantum-enabled future. Gain insights into the latest advancements and best practices for safeguarding data integrity and privacy in the era of quantum threats.
Interaction Latency: Square's User-Centric Mobile Performance MetricScyllaDB
Mobile performance metrics often take inspiration from the backend world and measure resource usage (CPU usage, memory usage, etc) and workload durations (how long a piece of code takes to run).
However, mobile apps are used by humans and the app performance directly impacts their experience, so we should primarily track user-centric mobile performance metrics. Following the lead of tech giants, the mobile industry at large is now adopting the tracking of app launch time and smoothness (jank during motion).
At Square, our customers spend most of their time in the app long after it's launched, and they don't scroll much, so app launch time and smoothness aren't critical metrics. What should we track instead?
This talk will introduce you to Interaction Latency, a user-centric mobile performance metric inspired from the Web Vital metric Interaction to Next Paint"" (web.dev/inp). We'll go over why apps need to track this, how to properly implement its tracking (it's tricky!), how to aggregate this metric and what thresholds you should target.
Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Em...Erasmo Purificato
Slide of the tutorial entitled "Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Emerging Trends" held at UMAP'24: 32nd ACM Conference on User Modeling, Adaptation and Personalization (July 1, 2024 | Cagliari, Italy)
Sustainability requires ingenuity and stewardship. Did you know Pigging Solutions pigging systems help you achieve your sustainable manufacturing goals AND provide rapid return on investment.
How? Our systems recover over 99% of product in transfer piping. Recovering trapped product from transfer lines that would otherwise become flush-waste, means you can increase batch yields and eliminate flush waste. From raw materials to finished product, if you can pump it, we can pig it.
Quality Patents: Patents That Stand the Test of TimeAurora Consulting
Is your patent a vanity piece of paper for your office wall? Or is it a reliable, defendable, assertable, property right? The difference is often quality.
Is your patent simply a transactional cost and a large pile of legal bills for your startup? Or is it a leverageable asset worthy of attracting precious investment dollars, worth its cost in multiples of valuation? The difference is often quality.
Is your patent application only good enough to get through the examination process? Or has it been crafted to stand the tests of time and varied audiences if you later need to assert that document against an infringer, find yourself litigating with it in an Article 3 Court at the hands of a judge and jury, God forbid, end up having to defend its validity at the PTAB, or even needing to use it to block pirated imports at the International Trade Commission? The difference is often quality.
Quality will be our focus for a good chunk of the remainder of this season. What goes into a quality patent, and where possible, how do you get it without breaking the bank?
** Episode Overview **
In this first episode of our quality series, Kristen Hansen and the panel discuss:
⦿ What do we mean when we say patent quality?
⦿ Why is patent quality important?
⦿ How to balance quality and budget
⦿ The importance of searching, continuations, and draftsperson domain expertise
⦿ Very practical tips, tricks, examples, and Kristen’s Musts for drafting quality applications
https://www.aurorapatents.com/patently-strategic-podcast.html
Performance Budgets for the Real World by Tammy EvertsScyllaDB
Performance budgets have been around for more than ten years. Over those years, we’ve learned a lot about what works, what doesn’t, and what we need to improve. In this session, Tammy revisits old assumptions about performance budgets and offers some new best practices. Topics include:
• Understanding performance budgets vs. performance goals
• Aligning budgets with user experience
• Pros and cons of Core Web Vitals
• How to stay on top of your budgets to fight regressions
Data Protection in a Connected World: Sovereignty and Cyber Securityanupriti
Delve into the critical intersection of data sovereignty and cyber security in this presentation. Explore unconventional cyber threat vectors and strategies to safeguard data integrity and sovereignty in an increasingly interconnected world. Gain insights into emerging threats and proactive defense measures essential for modern digital ecosystems.
Video traffic on the Internet is constantly growing; networked multimedia applications consume a predominant share of the available Internet bandwidth. A major technical breakthrough and enabler in multimedia systems research and of industrial networked multimedia services certainly was the HTTP Adaptive Streaming (HAS) technique. This resulted in the standardization of MPEG Dynamic Adaptive Streaming over HTTP (MPEG-DASH) which, together with HTTP Live Streaming (HLS), is widely used for multimedia delivery in today’s networks. Existing challenges in multimedia systems research deal with the trade-off between (i) the ever-increasing content complexity, (ii) various requirements with respect to time (most importantly, latency), and (iii) quality of experience (QoE). Optimizing towards one aspect usually negatively impacts at least one of the other two aspects if not both. This situation sets the stage for our research work in the ATHENA Christian Doppler (CD) Laboratory (Adaptive Streaming over HTTP and Emerging Networked Multimedia Services; https://athena.itec.aau.at/), jointly funded by public sources and industry. In this talk, we will present selected novel approaches and research results of the first year of the ATHENA CD Lab’s operation. We will highlight HAS-related research on (i) multimedia content provisioning (machine learning for video encoding); (ii) multimedia content delivery (support of edge processing and virtualized network functions for video networking); (iii) multimedia content consumption and end-to-end aspects (player-triggered segment retransmissions to improve video playout quality); and (iv) novel QoE investigations (adaptive point cloud streaming). We will also put the work into the context of international multimedia systems research.
Implementations of Fused Deposition Modeling in real worldEmerging Tech
The presentation showcases the diverse real-world applications of Fused Deposition Modeling (FDM) across multiple industries:
1. **Manufacturing**: FDM is utilized in manufacturing for rapid prototyping, creating custom tools and fixtures, and producing functional end-use parts. Companies leverage its cost-effectiveness and flexibility to streamline production processes.
2. **Medical**: In the medical field, FDM is used to create patient-specific anatomical models, surgical guides, and prosthetics. Its ability to produce precise and biocompatible parts supports advancements in personalized healthcare solutions.
3. **Education**: FDM plays a crucial role in education by enabling students to learn about design and engineering through hands-on 3D printing projects. It promotes innovation and practical skill development in STEM disciplines.
4. **Science**: Researchers use FDM to prototype equipment for scientific experiments, build custom laboratory tools, and create models for visualization and testing purposes. It facilitates rapid iteration and customization in scientific endeavors.
5. **Automotive**: Automotive manufacturers employ FDM for prototyping vehicle components, tooling for assembly lines, and customized parts. It speeds up the design validation process and enhances efficiency in automotive engineering.
6. **Consumer Electronics**: FDM is utilized in consumer electronics for designing and prototyping product enclosures, casings, and internal components. It enables rapid iteration and customization to meet evolving consumer demands.
7. **Robotics**: Robotics engineers leverage FDM to prototype robot parts, create lightweight and durable components, and customize robot designs for specific applications. It supports innovation and optimization in robotic systems.
8. **Aerospace**: In aerospace, FDM is used to manufacture lightweight parts, complex geometries, and prototypes of aircraft components. It contributes to cost reduction, faster production cycles, and weight savings in aerospace engineering.
9. **Architecture**: Architects utilize FDM for creating detailed architectural models, prototypes of building components, and intricate designs. It aids in visualizing concepts, testing structural integrity, and communicating design ideas effectively.
Each industry example demonstrates how FDM enhances innovation, accelerates product development, and addresses specific challenges through advanced manufacturing capabilities.
Are you interested in dipping your toes in the cloud native observability waters, but as an engineer you are not sure where to get started with tracing problems through your microservices and application landscapes on Kubernetes? Then this is the session for you, where we take you on your first steps in an active open-source project that offers a buffet of languages, challenges, and opportunities for getting started with telemetry data.
The project is called openTelemetry, but before diving into the specifics, we’ll start with de-mystifying key concepts and terms such as observability, telemetry, instrumentation, cardinality, percentile to lay a foundation. After understanding the nuts and bolts of observability and distributed traces, we’ll explore the openTelemetry community; its Special Interest Groups (SIGs), repositories, and how to become not only an end-user, but possibly a contributor.We will wrap up with an overview of the components in this project, such as the Collector, the OpenTelemetry protocol (OTLP), its APIs, and its SDKs.
Attendees will leave with an understanding of key observability concepts, become grounded in distributed tracing terminology, be aware of the components of openTelemetry, and know how to take their first steps to an open-source contribution!
Key Takeaways: Open source, vendor neutral instrumentation is an exciting new reality as the industry standardizes on openTelemetry for observability. OpenTelemetry is on a mission to enable effective observability by making high-quality, portable telemetry ubiquitous. The world of observability and monitoring today has a steep learning curve and in order to achieve ubiquity, the project would benefit from growing our contributor community.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/07/intels-approach-to-operationalizing-ai-in-the-manufacturing-sector-a-presentation-from-intel/
Tara Thimmanaik, AI Systems and Solutions Architect at Intel, presents the “Intel’s Approach to Operationalizing AI in the Manufacturing Sector,” tutorial at the May 2024 Embedded Vision Summit.
AI at the edge is powering a revolution in industrial IoT, from real-time processing and analytics that drive greater efficiency and learning to predictive maintenance. Intel is focused on developing tools and assets to help domain experts operationalize AI-based solutions in their fields of expertise.
In this talk, Thimmanaik explains how Intel’s software platforms simplify labor-intensive data upload, labeling, training, model optimization and retraining tasks. She shows how domain experts can quickly build vision models for a wide range of processes—detecting defective parts on a production line, reducing downtime on the factory floor, automating inventory management and other digitization and automation projects. And she introduces Intel-provided edge computing assets that empower faster localized insights and decisions, improving labor productivity through easy-to-use AI tools that democratize AI.
Hire a private investigator to get cell phone recordsHackersList
Learn what private investigators can legally do to obtain cell phone records and track phones, plus ethical considerations and alternatives for addressing privacy concerns.
AC Atlassian Coimbatore Session Slides( 22/06/2024)apoorva2579
This is the combined Sessions of ACE Atlassian Coimbatore event happened on 22nd June 2024
The session order is as follows:
1.AI and future of help desk by Rajesh Shanmugam
2. Harnessing the power of GenAI for your business by Siddharth
3. Fallacies of GenAI by Raju Kandaswamy
In this follow-up session on knowledge and prompt engineering, we will explore structured prompting, chain of thought prompting, iterative prompting, prompt optimization, emotional language prompts, and the inclusion of user signals and industry-specific data to enhance LLM performance.
Join EIS Founder & CEO Seth Earley and special guest Nick Usborne, Copywriter, Trainer, and Speaker, as they delve into these methodologies to improve AI-driven knowledge processes for employees and customers alike.
Knowledge and Prompt Engineering Part 2 Focus on Prompt Design Approaches
Scaling AI in production using PyTorch
1. 1 7 J U N E 2 0 2 1
S C A L I N G A I I N P R O D U C T I O N
U S I N G P Y T O R C H
G E E T A C H A U H A N
PyTorch Partner Engineering, Facebook AI
@ C H A U H A N G
2. MLOPS World 2021
A G E N D A 0 1
C H A L L E N G E S W I T H M L I N
P R O D U C T I O N
0 2
T O R C H S E R V E O V E R V I E W
0 3
B E S T P R A C T I C E S F O R P R O D U C T I O N
D E P L O Y M E N T
3. MLOps World 2021
P Y T O R C H C O M M U N I T Y G R O W T H
Source: https://paperswithcode.com/trends
4. MLOps World 2021
●
●
●
Cloud / On-Prem
Preprocessing
Application
Application logic
Application logic
Postprocessing
. . .
. . .
. . .
Performance Ease of use
Cost efficiency Deployment at scale
C H A L L E N G E S W I T H M L I N D E P L O Y M E N T
5. MLOps World 2021
INFERENCE AT SCALE
Deploying and managing models in production is
di
ffi
cult.
Some of the pain points include:
Loading and managing multiple models, on multiple
servers or end devices
Running pre-processing and post-processing code on
prediction requests.
How to log, monitor and secure predictions
What happens when you hit scale?
6. MLOps World 2021
TORCHSERVE
Easily deploy PyTorch models in production at scale
D E F A U LT H A N D L E R S
F O R C O M M O N T A S K S
L O W L AT E N C Y M O D E L
S E R V I N G
W O R K S W I T H A N Y M L
E N V I R O N M E N T
7. MLOps World 2021
• Default handlers for common use
cases (e.g., image segmentation,
text classification) along with
custom handlers support for other
use cases and a Model Zoo
• Multi-model serving, Model
versioning and ability to roll back
to an earlier version
• Automatic batching of individual
inferences across HTTP requests
• Logging including common
metrics, and the ability to
incorporate custom metrics
• Robust HTTP APIS -
Management and Inference
model1.pth
model1.pth
model1.pth
torch-model-archiver
HTTP
HTTP
http://localhost:8080/ …
http://localhost:8081/ …
Logging Metrics
model1.mar model2.mar model3.mar
model4.mar model5.mar
<path>/model_store
Inference API
Management API
TorchServe
Metrics API
Inference
API
Serving Model 3
Serving Model 2
Serving Model 1
torchserve --start
TORCHSERVE
8. T O R C H S E R V E D E T A I L :
M O D E L H A N D L E R S
TorchServe has default model handlers that
perform boilerplate data transforms for
common cases:
• Image Classification
• Image Segmentation
• Object Detection
• Text Classification
You can also create custom model handlers
for any model and inference task.
import torch
class MyModelHandler(object):
def initialize(self, context):
# get GPU status & device handle
# load model & supporting files (vocabularies etc.)
def preprocess(self, data):
# put incoming data into tensor
# transform as needed for your model
def inference(self, context):
# do predictions
def postprocess(self, output):
# process inference output, e.g. extracting top K
# package output for web delivery
def handle(self, context):
if not _service.initialized:
_service.initialize(context)
if data is None:
return None
data = _service.preprocess(data)
data = _service.inference(data)
data = _service.postprocess(data)
return data
9. M O D E L A R C H I V E
torch-model-archiver cli tool for packaging all
model artifacts into a single deployment unit
• model checkpoints or model definition file
with state_dict
• torchscript and eager mode support
• Extra files like vocab, config, index_to_name
mapping
torch-model-archiver
—model-name BERTSeqClassification_Torchscript
--version 1.0
--serialized-file Transformer_model/traced_model.pt
--handler ./Transformer_handler_generalized.py
--extra-files "./setup_config.json,./
Seq_classification_artifacts/index_to_name.json"
setup.config
{
“model_name": "bert-base-uncased",
“mode": "sequence_classification",
“do_lower_case": "True",
“num_labels": "2",
“save_mode": "torchscript",
“max_length": "150"
}
torchserve --start
--model-store model_store
—-models <path-to model-file/s3-url/azure-blob-url>
https://github.com/pytorch/serve/tree/master/model-archiver#creating-a-model-archive
10. D Y N A M I C B A T C H I N G
Via Custom Handlers
• Model Configuration based
• batch_size Max batch size
• max_batch_delay The max batch delay time
TorchServe waits to
receive batch_size number of requests
• (Coming soon) Batching support in default
handlers
curl localhost:8081/models/resnet-152
{
"modelName": "resnet-152",
"modelUrl": "https://s3.amazonaws.com/model-server/
model_archive_1.0/examples/resnet-152-batching/resnet-152.ma
"runtime": "python",
"minWorkers": 1,
"maxWorkers": 1,
"batchSize": 8,
"maxBatchDelay": 10,
"workers": [
{
"id": "9008",
"startTime": "2019-02-19T23:56:33.907Z",
"status": "READY",
"gpu": false,
"memoryUsage": 607715328
}
]
}
https://github.com/pytorch/serve/blob/master/docs/batch_inference_with_ts.md
11. M E T R I C S
Out of box metrics with ability to extend
• CPU, Disk, Memory utilization
• Requests type count
• ts.metrics class for extension
• Types supported - Size, percentage, counter,
general metric
• Prometheus metrics support available
# Access context metrics as follows
metrics = context.metrics
# Create Dimension Object
from ts.metrics.dimension import Dimension
# Dimensions are name value pairs
dim1 = Dimension(name, value)
.
dimN= Dimension(name_n, value_n)
# Add Distance as a metric
# dimensions = [dim1, dim2, dim3, ..., dimN]
metrics.add_metric('DistanceInKM', distance, 'km',
dimensions=dimensions)
# Add Image size as a size metric
metrics.add_size('SizeOfImage', img_size, None, 'MB', dimensions)
# Add MemoryUtilization as a percentage metric
metrics.add_percent('MemoryUtilization', utilization_percent, None,
dimensions)
# Create a counter with name 'LoopCount' and dimensions
metrics.add_counter('LoopCount', 1, None, dimensions)
# Log custom metrics
for metric in metrics.store:
logger.info("[METRICS]%s", str(metric))
https://github.com/pytorch/serve/blob/master/docs/metrics.md
12. MLOps World 2021
RECENT FEATURES
+ Ensemble Model support, Captum Model Interpretability
+ Kubeflow Pipelines /KFServing Integration with Auto-scaling and Canary rollout on any cloud/on-prem
+ GCP Vertex AI Serverless pipelines
+ MLflow Integration
+ Prometheus Integration with Grafana
+ Multiple nodes on EC2, Autoscaling on SageMaker/EKS, AWS Inferentia support
+ MMF, NMT, DeepLapV3 new examples
13. Deployment
models
Optimizations Resilience Measurement
Responsible AI
Standalon
e
Primary backu
p
Orchestratio
n
Cloud vs.
on-premises
Performance vs.
latency
TorchScript profilin
g
Offline vs. real-tim
e
Cost
Robust endpoin
t
Auto-scalin
g
Canary
deployment
s
A / B testing
Metric
s
Model
performanc
e
Interpretabilit
y
Feedback loop
Fairnes
s
Human-centered
design
B E S T P R A C T I C E S F O R P R O D U C T I O N D E P L O Y M E N T S
14. MLOps World 2021
Fairness by design
• Measure skewness of data, model bias, data bias; identify relevant metrics
• Transparency, Explainable AI, inclusive design
Human-centered design
• Consider AI-driven decisions and their impact on people at the time of model design
• Provide ability to have human recourse vs. full automation – for example, need to avoid a mortgage
applications AI rejecting people of certain category or race
• Computer vision models measure results based on demographics; for example, include support for different
skin tones, age groups
R E S P O N S I B L E A I
15. MLOps World 2021
• Build with performance vs. latency goals in mind
• Reduce size of the model: Quantization, pruning, mixed precision training
• Reduce latency: TorchScript model; use SnakeViz profiler
• Evaluate GPU vs. CPU for low latency
• Evaluate REST vs. gRPC for your prediction service
O P T I M I Z A T I O N S
16. MLOps World 2021
fp32 accuracy int8 accuracy change Technique CPU inference speed up
ResNet50 76.1
Top-1, Imagenet
-0.2
75.9
Post Training
2x
214ms ➙102ms,
Intel Skylake-DE
MobileNetV2 71.9
Top-1, Imagenet
-0.3
71.6
Quantization-Aware
Training
4x
75ms ➙18ms
OnePlus 5, Snapdragon 835
Translate / FairSeq 32.78
BLEU, IWSLT 2014 de-en
0.0
32.78
Dynamic
(weights only)
4x
for encoder
Intel Skylake-SE
These models and more available on TorchHub - https://pytorch.org/hub/
QUANTIZATION
18. MLOps World 2021
B E R T
M O D E L
P R O F I L I N G
Torchscript Mode
4x speedup
19. MLOps World 2021
Offline vs. real-time predictions
• Offline: Dynamic batching
• Online: Async processing – push/poll
• Pre-computed predictions for certain elements
Cost optimizations
• Spot Instances for offline
• Autoscaling based on metrics, on-demand cluster
• Evaluate AI Accelerators supported like AWS Inferentia for lower cost point
O P T I M I Z A T I O N S ( C O N T D . )
20. MLOps World 2021
Develop
,
Test
Production
Staging
,
Experiments
Hybrid Cloud
On-prem Cloud Managed
Install from Source
Standalone
Docker
Large Scale
Production
MLflow, Kubeflow
Kubernetes, Kubeflow/KFserving
Primary/Backup, ML Microservices
Autoscaling, Canary Rollouts
Minikub
e
Self managed Docker AWS CloudFormation
CLOUD VMs/ Containers
Microservices behind
API Gateway
CLOUD VMs/ Containers
AWS SageMaker
Endpoints, BYOC
AWS SageMaker
EKS/AKS/GKE
AWS SageMaker/ GCP
AI Platform
Serverless Functions
GCP Vertex AI,
AWS SageMaker
Canary Rollouts
Databricks
Managed MLflow
D E P L O Y I N G M O D E L S I N P R O D U C T I O N
21. MLOps World 2021
Create robust endpoint for serving, for example, SageMaker endpoint
Auto-scaling with orchestration deployments, multi-node for EC2, and other scenarios
Canary deployments, test new version of a model on small subset before making
default
Shadow inference, deploy new version of model in parallel
A / B testing of different versions of model
R E S I L L I E N C E
22. MLOps World 2021
Define model performance metrics, such as accuracy, while designing the AI service;
use-case specific
Add custom metrics as appropriate
Use CloudWatch or Prometheus dashboards for monitoring model performance
Model interpretability analysis via Captum
Deploy with a feedback loop, if model accuracy drops over time or new version,
analyze issues like concept drift, stale data, etc.
M E A S U R E M E N T
23. MLOps World 2021
Understand
Align
Mitigate
Monitor
Measure
Stakeholder conversations to find
consensus and outline measurement and
mitigation plans
Analyze model performance,
label bias, outcomes, and other
relevant signals
Address observed
issues in dataset,
models, policies, etc
How might the product’s goals, its policy,
and its implementation affect users from
different subgroups? Identify contextual
definitions of fairness
Monitor effect of mitigations on
subgroups, and ensure fairness
analysis holds as product adapts
FAIRNESS BY DESIGN
24. CAPTUM
Text Contributions: 7.54
Image Contributions: 11.19
Total Contributions: 18.73
0 200 400 600 800
400
300
200
100
0
S U P P O R T F O R AT T R I B U T I O N A LG O R I T H M S
T O I N T E R P R E T:
• Output predictions with respect to inputs
• Output predictions with respect to layers
• Neurons with respect to inputs
• Currently provides gradient & perturbation based
approaches (e.g. Integrated Gradients)
Model interpretability library for PyTorch
https://captum.ai/
26. MLOps World 2021
COMMUNIT Y PROJECTS https://github.com/cceyda/torchserve-dashboard
https://github.com/Unity-Technologies/SynthDet
https://medium.com/pytorch/how-wadhwani-ai-uses-pytorch-
to-empower-cotton-farmers-14397f4c9f2b
27. MLOps World 2021
FUTURE RELEASES
+ Improved memory and resource usage for better scalability
+ C++ Backend for lower latency
+ Enhanced profiling tools