m AI and ML Concepts 1



Key Points

  1. 2 contexts: in-process analytics and stand alone analytics


Learning Roadmap 


Free AI course list

https://www.linkedin.com/posts/dirk-zee_free-ai-courses-save-1000s-of-hours-1-activity-7125048945776451586-iMDg?utm_source=share&utm_medium=member_desktop


data-science-4-dummies-v3-2022.pdf.   link.  << good beginner concepts on data science engineering / solutions./ concepts


https://objectcomputing.com/files/5715/6095/8662/Slide_Deck_Groovy_for_Data_Science_Webinar.pdf  - Groovy for Data Science - migrate v3x

http://glaforge.appspot.com/article/machine-learning-apis-with-apache-groovy - article on Groovy ML

https://www.udemy.com/course/python-for-data-science-and-machine-learning-bootcamp/ - basic ML and Python programming course covers Spark etc as well

http://ciml.info/ - ML algorithms

More Advanced

https://www.udemy.com/course/pytorch-for-deep-learning-with-python-bootcamp/

ML use cases

_ml-use-cases-The Big Book of Machine Learning Use Case.pdf


AI and Digital Transformation Competencies for Civil Servants

https://www.linkedin.com/posts/transformpartner_the-ai-and-digital-transformation-competency-ugcPost-7153356976977059840--Ee7?utm_source=share&utm_medium=member_desktop

ai-skills-Artificial Intelligence and Digital Transformation Competencies-study-2022,pdf.   link

ai-skills-Artificial Intelligence and Digital Transformation Competencies-study-2022.pdf. file


there is an unmet need to develop comprehensive digital competency frameworks that can:

 

1. Clearly identify the internal challenges a government faces in its digital transformation journey;

2. Propose specific competencies that can address those challenges,

The Artificial Intelligence and Digital Transformation Competency Framework includes three major Competency Domains:

 

1.        #DigitalPlanning and Design

2.        Data Use and Governance

3.        #DigitalManagement and Execution

 

The competency framework also includes five complementary Attitudes that enable civil servants to pursue digital transformation effectively:

1. Trust


2. Creativity


3. Adaptability


4. Curiosity


5. Experimentation

 

Each Competency Domain is structured around three Proficiency Levels: Basic, Medium and Advanced, and includes an ‘AI-specific level’ that aims to identify and unpack the major AI elements.

I'll argue Trust is actually a capability not an attitude


Learning Strategies for AI



https://www.youtube.com/watch?v=h2FDq3agImI

Summary of "How I'd Learn AI in 2024 (if I could start over)"

  1. Background and Context:

    • The speaker began studying AI in 2013 and has since worked as a freelance data scientist.
    • They have a YouTube channel with over 25,000 subscribers, sharing their AI journey and knowledge.
    • Emphasis on the growing AI market (expected to reach nearly $2 trillion by 2030) and the ease of entry into the field, especially with pre-trained models from OpenAI.
  2. Understanding AI and Choosing a Path:

    • AI is a broad term, encompassing machine learning, deep learning, and data science.
    • The speaker encourages understanding AI beyond popular misconceptions and choosing between coding and no-code/low-code tools.
    • They stress the importance of technical understanding for those aspiring to build reliable AI applications.
  3. Technical Roadmap and Learning Approach:

    • The roadmap focuses on learning by doing and reverse engineering, rather than solely theoretical understanding.
    • Key areas include setting up a working environment, learning Python, and understanding essential libraries like NumPy, pandas, and Matplotlib.
    • Emphasis on practical learning through projects and portfolio building, with resources like Kaggle and Project Pro mentioned.
  4. Specialization and Knowledge Sharing:

    • After gaining foundational knowledge and experience, the speaker advises choosing a specialization in AI.
    • Sharing knowledge and teaching others is highlighted as a way to deepen one's understanding and contribute to the AI community.
  5. Final Steps and Community Engagement:

    • The importance of applying knowledge in real-world scenarios, embracing challenges, and continuous learning.
    • Encourages joining communities of like-minded individuals and staying updated with the rapidly evolving field of AI and data science.
    • Offers a free resource with a complete AI learning roadmap, including training videos and instructions.




References

Reference_description_with_linked_URLs_________________________Notes_________________________________________________________


http://ciml.info/Machine Learning overview focused on algorithms for different problem types

Math Foundations for AI ML

Statistics Foundations for AI ML

http://ciml.info/

https://github.com/hal3/ciml/

ML Course 1
https://www.datacamp.com/courses/preparing-for-statistics-interview-
questions-in-python?fbclid=IwAR29QSzZqoJEaarVtOoSRimPGOFbPbVYGLQ3
j2nyqt4PZ74AmkcJqILip94
Interview questions for statistics in python
https://courses.edx.org/dashboardedX courses for ML
https://thenextweb.com/podium/2019/11/11/machine-learning-algorithms-and-the-art-of-hyperparameter-selection/ML Algorithm concepts
https://www.linkedin.com/feed/update/urn:li:activity:6617271768493191168/ML Intro Concepts links - Linkedin
Guide_to_Open_Source_AI.pdfGuide_to_Open_Source_AI.pdf
machine-learning-for-dummies-w_wile255.pdfmachine-learning-for-dummies-w_wile255.pdf


ML use cases
The Big Book of Data Science Use Cases.pdfThe Big Book of Data Science Use Cases.pdf
https://www.gartner.com/smarterwithgartner/top-3-benefits-of-ai-projects/Gartner - 3 key AI benefits
https://alyce.ai/application/files/7615/6933/9987/ALYCE_Survey_.pdfAlyce.AI - object computing survey with use cases
AI_2019-news-from-the-batch-includes-AV-info-jmason.pdfAI status - 2019 - The Batch
https://www.forbes.com/sites/janakirammsv/2019/01/01/an-executives-guide-to-
understanding-cloud-
based-machine-learning-services/#6e7c6c043e3e
Forbes - overview of cloud machine learning services

https://hbr.org/2020/10/a-practical-guide-to-building-ethical-ai

Practical Guide to Building Ethical AI ** HBR pdf

Practical Guide to Building Ethical AI ** HBR






Machine Learning Basics

https://www.quantinsti.com/blog/machine-learning-basics

Machine Learning Basics

https://www.techrepublic.com/article/machine-learning-the-smart-persons-guide/

Executive summary 1 on ML

https://lfdl.io/projects/

Linux Foundation Open AI

Deep Learning Solutions

http://ciml.info/

Machine Learning Math Course free ***

https://www.udemy.com/hands-on-introduction-to-artificial-intelligenceai/learn/lecture/12130906#overviewUdemy free ML course **

file:///C:/Users/Jim%20Mason/Google%20Drive/_docs/howto/data/mlearn/Guide_to_Open_Source_AI.pdf

Linux F Guide to OS AI frameworks - good starting point

https://objectcomputing.com/files/5715/6095/8662/Slide_Deck_Groovy_for_Data_Science_Webinar.pdf

Groovy for Data Science

http://glaforge.appspot.com/article/machine-learning-apis-with-apache-groovyML apis from Groovy ( Java ) integrating Google AI - ML services **
machine-learning-with-python-v2-2024.pdf.   linkmachine-learning-with-python-v2-2024 **

https://www.datacamp.com/community/tutorials/finance-python-trading?fbclid=IwAR1n33gRfaNvZv02RR5wCrSNkS4RqxJQqipBG
QuqMyWc6_akSPQ2hAkFp6c

DataCamp course - Python for Algorithmic Trading

https://www.datasciencecentral.com/profiles/blogs/new-books-and-resources-for-dsc-members

datasciencecentral.com free ebooks

https://www.slideshare.net/carologic/ai-and-machine-learning-demystified-by-carol-smith-at-midwest-ux-2017

https://drive.google.com/open?id=1oa3bMB6KHDbo6ftgZMWSfnj8FpwQWCrm

AI and Machine Learning Demystified by Carol Smith at Midwest UX 2017

https://drive.google.com/open?id=1sPBXhjtvEt54BlyAcW_Mz1PMETzCTLdvmachine learning for dummies ebook

https://d2wvfoqc9gyqzf.cloudfront.net/content/uploads/2018/09/Ng-MLY01-13.pdf

https://drive.google.com/open?id=1wDgmF9V3A7zeBXslgmRube2_cQ0XzxrM

Deep Learning Strategies and Project mgt - Andrew Ng
https://drive.google.com/open?id=1H3ivxDSLUxsw2CQAXUhdyzMPGZdnAj6jPython Tensorflow Tutorial - datacamp 2018
python-tensorflow-tutorial1-datacamp-Convolutional Neural Networks with TensorFlowpython-tensorflow-tutorial1-datacamp-Convolutional Neural Networks with TensorFlow

https://www.datacamp.com/community/tutorials/cnn-tensorflow-python

https://drive.google.com/open?id=1H3ivxDSLUxsw2CQAXUhdyzMPGZdnAj6j

Tutorial: Tensorflow neural network example in Python: Datacamp

https://www.datacamp.com/community/tutorials/tensorflow-tutorial

https://drive.google.com/open?id=1pC1805LpYe4el3mCT7qw0wzqV-vP3sv4

Tutorial: Tensorflow basics : Datacamp


https://drive.google.com/open?id=1b2bh7XJ_MkVG0C1kx7IITOqK0XfFh8WKContainers for ML workloads


billable courses
https://www.udemy.com/course/python-for-data-science-and-machine-learning-bootcamp/Python ML covers Python basics - $12 - RECOMMENDED first course


https://university.cloudera.com/instructor-led-training/cloudera-data-scientist-trainingCloudera - $3200 - 4 days - pyspark - sparklr - spark2 env
https://www.udemy.com/course/pytorch-for-deep-learning-with-python-bootcamp/Pytorch bootcamp - Udemy - $12 - sharp focus on Pytorch -requires Python
https://www.udemy.com/course/python-for-data-science-and-machine-learning-bootcamp/Python ML covers Python basics - $12
https://www.udemy.com/course/the-complete-neural-networks-bootcamp-theory-applications/Complete Neural Networks bootcamp - $12 - more ML models, examples - requires Python, math




Google ML



GCP platform - free account

https://www.kaggle.com/

Google Kaggle - free ML test account

https://www.kaggle.com/learn/overview

Kaggle learning resources

https://techcrunch.com/2019/06/24/google-brings-together-bigquery-and-
kaggle-in-new-integration/?yptr=yahoo

Link Kaggle notebook to BigQuery results

https://www.datacamp.com/community/blog/keras-cheat-sheet?fbclid=IwAR2x0PgQEjsALKSIrlC_ZkADzXgqUyG9dJ_zAeh7h1c
VFrQQzcEjRxtWB98
Keras cheat sheet
https://keras.io/Keras Basics
https://drive.google.com/open?id=1Y5jf0CLBlhieO85Er0n8bsYnScnXEIvGTf.keras - tensor flow Keras gen 2

https://medium.com/sciforce/understanding-tensor-processing-units-10ff41f50e78

https://drive.google.com/open?id=1lFzNJcXvyuD41vkB1K_Gkftgz9o4DHaJ

Tensorflow concepts

https://cloud.google.com/blog/products/gcp/understanding-neural-networks-with-tensorflow-playground

https://drive.google.com/open?id=13_fj-IB2jwKIdBHquGm43gpYq4GhsCnl

Tensorflow exampe overview - playground - 2015

Tensorflow Tutorial
https://cloud.google.com/tpu/Google TPU AI processor overview
https://drive.google.com/open?id=1UWwyoHEJZi1f6mQo3SvpxgQz8WBDjuImAI hardware- GPU or TPU?
https://cloud.google.com/products/ai/building-blocks/Google AI building blocks
https://cloud.google.com/automl/Google auto translation bots


MLFlow - Linux Foundation - open source platform for the machine learning lifecycleopen source platform for the machine learning lifecycle
https://mlflow.orgMLFlow 
https://mlflow.org/#:~:text=MLflow%20is%20an%20open%20source,
and%20a%20central%20model%20registry.
MLFlow model registry
https://databricks.com/blog/2018/06/05/introducing-mlflow-an-open-source-machine-learning-platform.htmlOverview
https://mlflow.org/docs/latest/quickstart.htmlQuickstart
https://towardsdatascience.com/getting-started-with-mlflow-52eff8c09c61Getting Started article
https://github.com/mlflow/mlflowMLFlow on github




AWS ML
https://aws.amazon.com/machine-learning/e-guide/?trk=FB_AI_DeepLens_eGuide_v3&sc
_channel=PSM&sc_campaign=field&sc_publisher=
FB&sc_category=DeepLens&sc_country=US&sc_geo=NAMER&sc_outcome=aware&sc_
medium=FIELD-P%7CFB%7CSocial-P%7CAll%7CAW%7CMachine+Learning%7CDeepLens%7CUS%7CEN%7CImage&fbclid=
IwAR2h6I83BZsUuK0quB2bgT1uz6ux229cEDCQtlNpLwNQfvgpSybvuskc5p4
Business View of AI services on AWS

ML Training
https://aws.amazon.com/machine-learning/ML Overview

ML Apps, Services

ML Sagemaker - Deep Learning Utility

ML Tools - Keras, Tensorflow, Kubeflow




Azure ML










IBM ML

https://www.ibm.com/cloud/watson-studio/resources?cm_mmc=PSocial_Linkedin-_-Hybrid Cloud_
Data Science-_-WW_NA-_-1074735077_Tracking Pixel&cm_mmca1=000000RE&cm_mmca2=10009788&cm_mmca4=1074735077&cm_mmca5=
1078561140&cm_mmca6=d3617d3d-e5a9-4451-ba40-18c00fc18384

ibm watson

free access, ar









Facebook ML

https://ai.facebook.com/blog/dlrm-an-advanced-open-source-deep-learning-recommendation-model/

ML Recommendation Mgr open sourced



ML Tutorials

https://github.com/cloud-annotations/training 

https://www.linkedin.com/feed/update/urn:li:activity:6496499508409561088

real-time image object recognition customizable in your browser with TensorFlow.js and Python

demo real-time image recognition

https://rasa.com/

https://rasa.com/docs/

open source framework to build smart chatbots in python w RNN, NLU, NLP
http://ciml.info/Machine Learning overview focused on algorithms for different problem types






Data Science Tools

https://drive.google.com/drive/u/0/folders/1e5WecKq_87ahkcL3UM7EdOz73qviLyhW

https://objectcomputing.com/files/5715/6095/8662/Slide_Deck_Groovy_for_Data_Science_Webinar.pdf

Groovy Data Science - Paul King - OCI
https://campus.datacamp.com/courses/building-chatbots-in-python/chatbots-101?ex=1DataCamp - Python Chatbots course




https://www.cloudera.com/products/data-science-and-engineering/data-science-workbench.htmlCloudera DS workbench






Other ML providers and services


https://www.h2o.ai/products/h2o-driverless-ai/H2O AI services - platform with consulting services

https://rasa.com/

https://github.com/RasaHQ/rasa

Rasa - open-source smart assistant framework for chatbots

see Smart Assistant Integration for more on bulding assistants for NLP domains

https://www.nextplatform.com/2019/03/05/one-step-closer-to-deep-learning-on-neuromorphic-hardware/Deep Learning neural networks have a software training environment ( Whetstone ) that can work on multiple neural network platforms when ported. Efficient, mimics human neuron behavior.  Goal of this work to be able to leverage standard deep learning technologies to produce inferencing software capable of running in a much lower power envelope when deployed. 
https://www.sage.com/en-us/blog/wp-content/uploads/sites/2/2018/09/Artificial_Intelligence_In_2019_Sage_Handbook.pdfAI business concepts handbook 
https://www.forbes.com/insights-intelai/Very good series of articles on AI impacts across many industries including AI concepts as well
https://blockchain.ieee.org/technicalbriefs/march-2019/towards-advanced-artificial-intelligence-using-blockchain-technologiesAI and blockchain together article


AI Regulation Concepts


EUBOF on convergence of AI and Blockchain report - 2021 url

EUBOF on convergence of AI and Blockchain meeting recording


EU Regulatory Framework Proposal on AI - 2021 url

EU AI Regulatory Concepts today - 2022 url 

EU AI Regulatory Concepts today - 2022 url.pdf file


US AI Bill of Rights Proposal 2022 url

US-2022-Blueprint-for-an-AI-Bill-of-Rights.pdf file


UK AI Regulation Approach doc - 2022 url

UK-AI-regulation-approach-2022-faegredrinker.com-AI Regulation in the UK New Government Approach.pdf file

U.K. Government intends to “regulate the use of AI rather than the technology itself,”
UK establishing an approach to regulating AI - 2022 url
UK on EU AI regulations as risky 2022 url
ai-governance-for-citiies-2023-UN-MILA.pdf. link
AI-goverenance-2021-The_ghost_of_AI_governance_past_present_and_future.pdf file
EU AI Regulation flawed - 2022 article 

G7 2023 Statement on AI Governance.  url 

G7 2023 Statement on AI Governance.  linkj







Key Concepts


AI Terms

https://www.linkedin.com/posts/dirk-zee_%3F%3F%3F%3F%3F%3F%3F%3F%3F%3F%3F%3F-%3F%3F-dont-miss-activity-7137821904437948416-m_t_?utm_source=share&utm_medium=member_desktop

Don't miss the Essential Concepts:



- 𝗔𝗜: The overarching world of Artificial Intelligence, transforming industries.

- 𝗠𝗟: Machine Learning's role in shaping intelligent systems.

- 𝗗𝗲𝗲𝗽 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: Neural networks for human-like thinking.

- 𝗡𝗲𝘂𝗿𝗮𝗹 𝗡𝗲𝘁𝘄𝗼𝗿𝗸: Mimicking human brain functions for learning.

- 𝗦𝘂𝗽𝗲𝗿𝘃𝗶𝘀𝗲𝗱 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: Teaching computers with labelled examples.

- 𝗨𝗻𝘀𝘂𝗽𝗲𝗿𝘃𝗶𝘀𝗲𝗱 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: Machines finding patterns without labels.

- 𝗥𝗲𝗶𝗻𝗳𝗼𝗿𝗰𝗲𝗺𝗲𝗻𝘁 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: Trial-and-error learning for machines.

- 𝗡𝗟𝗣: Tech enabling computers to understand human language.

- 𝗖𝗼𝗺𝗽𝘂𝘁𝗲𝗿 𝗩𝗶𝘀𝗶𝗼𝗻: Machines interpreting visual information.

- 𝗖𝗵𝗮𝘁𝗯𝗼𝘁: Conversational AI for customer support and more.

- Hallucination - when Gen AI produces the wrong answer but is convinced it's correct

- 𝗜𝗢𝗧: Devices connected, sharing data for smart applications.

- 𝗖𝗹𝗼𝘂𝗱 𝗖𝗼𝗺𝗽𝘂𝘁𝗶𝗻𝗴: Remote storage, management, and data processing.

- 𝗕𝗶𝗮𝘀 𝗶𝗻 𝗔𝗜: Addressing unintentional biases in algorithms.

- 𝗔𝗹𝗴𝗼𝗿𝗶𝘁𝗵𝗺: Core of AI, the building block for intelligent systems.

- 𝗗𝗮𝘁𝗮 𝗠𝗶𝗻𝗶𝗻𝗴: Extracting patterns and insights from vast datasets.

- 𝗕𝗶𝗴 𝗗𝗮𝘁𝗮:Navigating challenges with massive and diverse data.

- 𝗥𝗼𝗯𝗼𝘁𝗶𝗰𝘀: Merging AI with physical machines for automation.

- 𝗔𝗹𝗴𝗼𝗿𝗶𝘁𝗵𝗺𝗶𝗰 𝗙𝗮𝗶𝗿𝗻𝗲𝘀𝘀: Ensuring fairness and avoiding bias in AI.

- 𝗧𝗿𝗮𝗻𝘀𝗳𝗲𝗿 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: Applying knowledge for enhanced AI efficiency.

- 𝗘𝗱𝗴𝗲 𝗖𝗼𝗺𝗽𝘂𝘁𝗶𝗻𝗴: Localised AI implementation for efficiency.

- 𝗘𝘅𝗽𝗹𝗮𝗶𝗻𝗮𝗯𝗹𝗲 𝗔𝗜: Making AI decisions transparent and understandable.

- 𝗚𝗔𝗡𝘀: AI creating realistic data through adversarial networks.

- 𝗘𝗱𝗴𝗲 𝗔𝗜: Localised AI for reduced reliance on centralised servers.

- 𝗔𝗜 𝗘𝘁𝗵𝗶𝗰𝘀: Guiding principles for responsible AI development.

Image preview




AI platforms

  1. Cloud ML platforms engineered for ML performance

    Recent hardware investments from Amazon, Google, Microsoft and Facebook, made ML infrastructure cheaper and efficient. Cloud providers are now offering custom hardware that’s highly optimized for running ML workloads in the cloud. Google’s TPU and Microsoft’s FPGA offerings are examples of custom hardware accelerators exclusively meant for ML jobs. When combined with the recent computing trends such as Kubernetes, ML infrastructure becomes an attractive choice for enterprises.

    Amazon EC2 Deep Learning AMI backed by NVIDIA GPU, Google Cloud TPU, Microsoft Azure Deep Learning VM based on NVIDIA GPU, and IBM GPU-based Bare Metal Servers are examples of niche IaaS for ML.



  2. Deep learning systems can learn new information, improve their own performance

    The complexity of the steps needed in a given decision making process all depend on the skills of the person or system making the decision. Just like we teach people basic skills before advanced, we teach systems the same way. Extending the metaphor, just like people, systems can also learn on their own:  for a given goal, collect relevant data and update their understanding ( the model ) of how best to solve a problem.
    In a driving example, once the system knows how to brake, how to steer, how to follow a digital map, how to sense objects within a range, asking the system to drive me to work is not a complex set of instructions and decisions. Systems, just like people, can build skills, knowledge and decision making capability leveraging what they already know.
     (edited)



Key Questions

  1. what is data science?
  2. how is it different from data analysis?
  3. what questions do data scientists answer?
  4. what questions do data analysts answer?
  5. how are both different from data architects and engineers?
  6. what are key AI use cases by industry?
  7. what are key benefits for AI and ML ?  VCR3S


ML / AI Operations Models


User Type - Human or Automated System for decision making ( eg a vehicle recognizing people crossing a road )

Operations Type - Standalone research ( which stock to buy long term ) or Embedded Operations ( decide when to buy power on a grid automatically )



4 Types of AI

https://www.govtech.com//computing/understanding-the-four-types-of-artificial-intelligence.html?utm_campaign=Re-engagement%20-%20GT&utm_medium=email&_hsmi=223352446&_hsenc=p2ANqtz--J8naIOn0aB0TgrqNWDE3kqq3yFEH5eQBWk7YF3DznZ8_0G9B8oKoXge_75XP67FNxYFcu89SblNADJRYFKjoXw0h7DA&utm_content=223352168&utm_source=hs_automation

govtech.com-Understanding the Four Types of Artificial Intelligence.pdf  link


Generative AI creates novel content

https://finance.yahoo.com/news/china-building-parallel-generative-ai-150129822.html

After text-to-image tools from Stability AI and OpenAI became the talk of the town, ChatGPT's ability to hold intelligent conversations is the new obsession in sectors across the board.


Generative AI has trust and quality issues: needs better trust, transparency, verifications with SLT

https://www.cnet.com/tech/google-lets-people-start-trying-bard-its-own-ai-chatbot/

SLT - Shared Ledger Technology - can make Generative AI better

It can add better: trust, transparency and verifications



Enterprise AI Solutions & Operations in the Cloud

ai-devops-ENTERPRISE AI in the Cloud-2024.pdf

PART I: INTRODUCTION
CHAPTER 1: ENTERPRISE TRANSFORMATION WITH AI IN THE CLOUD 3
CHAPTER 2: CASE STUDIES OF ENTERPRISE AI IN THE CLOUD 19

PART II: STRATEGIZING AND ASSESSING FOR AI
CHAPTER 3: ADDRESSING THE CHALLENGES WITH ENTERPRISE AI 31
CHAPTER 4: DESIGNING AI SYSTEMS RESPONSIBLY 41
CHAPTER 5: ENVISIONING AND ALIGNING YOUR AI STRATEGY 50
CHAPTER 6: DEVELOPING AN AI STRATEGY AND PORTFOLIO 57
CHAPTER 7: MANAGING STRATEGIC CHANGE 66

PART III: PLANNING AND LAUNCHING A PILOT PROJECT
CHAPTER 8: IDENTIFYING USE CASES FOR YOUR AI/ML PROJECT 79
CHAPTER 9: EVALUATING AI/ML PLATFORMS AND SERVICES 106
CHAPTER 10: LAUNCHING YOUR PILOT PROJECT 152

PART IV: BUILDING AND GOVERNING YOUR TEAM
CHAPTER 11: EMPOWERING YOUR PEOPLE THROUGH ORG CHANGE MANAGEMENT 163
CHAPTER 12: BUILDING YOUR TEAM 173

PART V: SETTING UP INFRASTRUCTURE AND MANAGING OPERATIONS
CHAPTER 13: SETTING UP AN ENTERPRISE AI CLOUD PLATFORM INFRASTRUCTURE 187
CHAPTER 14: OPERATING YOUR AI PLATFORM WITH MLOPS BEST PRACTICES 217

PART VI: PROCESSING DATA AND MODELING
CHAPTER 15: PROCESS DATA AND ENGINEER FEATURES IN THE CLOUD 243
CHAPTER 16: CHOOSING YOUR AI/ML ALGORITHMS 268
CHAPTER 17: TRAINING, TUNING, AND EVALUATING MODELS 315

PART VII: DEPLOYING AND MONITORING MODELS
CHAPTER 18: DEPLOYING YOUR MODELS INTO PRODUCTION 345
CHAPTER 19: MONITORING MODELS 361
CHAPTER 20: GOVERNING MODELS FOR BIAS AND ETHICS 377



Common ML / AI use cases

https://alyce.ai/application/files/7615/6933/9987/ALYCE_Survey_.pdf

Where is AI / ML used in 2019?

61% Predictive analytics
39% Image or video analysis
37% Customized product recommendations
33% Marketing personalization
29% Cyber security or fraud detection
18% Edge analytics
18% Manufacturing process optimization
14% Prescriptive advisors
14% Supply chain optimization
12% Database cleansing and maintenan



Selected Key AI use cases to deliver value


Recognition ( eg is the person walking into a cross walk now or not ? )

Correlation ( eg what is correlation between technical degrees and income levels ? )

Prediction ( eg given the top 7 cap stocks trends on revenue and EBITDA over the last 8 quarters, identify the minimum expected price growth with 95% confidence for each )

Optimization (eg given a delivery time liimit, find the lowest cost delivery route or find the used car value in my zip code )

Variances ( eg meeting or exceeding standards:  OKRs, KPIs, SLOs etc )

Gen AI ( eg for the top 3 cloud providers, create a detailed report that compares operationas costs given a use case with an online shopping app site using IAAS and an open-source database, API framework, 10,000 daily views, 1,000 order queries and transactions )



Report on AI use case examples

As an AI architect I want to create a report showing AI use cases  for EACH of these use case:  prediction, optimization, recognition, Generative AI, monitoring, correlation, variance analysis with detailed definitions, use case examples, the net value for the example, the governance capabilities for the example ( monitoring, bias, transparency, accountability ), steps to get started for the example and hyper links to each example. 

gai>> 

https://docs.google.com/document/d/1BPCSIz3t4yFqefFBmh4OZuT3i3BJkwbMwTC3o6KG798/edit?usp=drive_link










Common AI and ML benefits


  1. Risk
    1. risk reduction, mitigation, detection metrics
  2. Speed
    1. process, customer service calls, operations efficiency, customer satisfaction and response metrics
  3. Revenue
    1. churn, lead management, revenue, orders, web traffic, abandons, ux, user interactions metrics
  4. Resources
    1. resource efficiencies, critical resource availability metrics
  5. Value
    1. customer perspectives and value delivered metrics
  6. Quality
    1. quality improvements on process, data, services, user experiences, settlements metrics


Gartner - 3 key AI benefits

https://www.gartner.com/smarterwithgartner/top-3-benefits-of-ai-projects/



Common ML Engineering Foundations Course - free

https://end-to-end-machine-learning.teachable.com/p/000-foundational-skills

sign up - jm9y

https://end-to-end-machine-learning.teachable.com/


Resources



Data Science Concepts Course - Free 

https://end-to-end-machine-learning.teachable.com/p/data-science-concepts

Some useful things to know as you get started



Common ML Scientist Skill Sets

https://www.h2o.ai/careers/

A great candidate for Customer Data Scientist/Sales Engineer should:

  • Know Python, R, Java, Scala, Spark.
  • Have experience with Big Data including Hadoop, Spark, Kafka.
  • Have a working knowledge of ML algorithms for Regression and Classification problems.
    • Understanding of Supervised, Unsupervised, Deep learning techniques
    • Knowledge of XGBoost, Linear regression, GBM, GLM, LightGBM, Random Forest and other common ML algorithms 
    • Experience using TensorFlow, Keras, Scikit libraries for performing ML .



ML Algorithm Tuning Concepts

https://thenextweb.com/podium/2019/11/11/machine-learning-algorithms-and-the-art-of-hyperparameter-selection/

These algorithms automatically adjust (learn) their internal parameters based on data. However, there is a subset of parameters that is not learned and that have to be configured by an expert. Such parameters are often referred to as “hyperparameters” — and they have a big impact on our lives as the use of AI increases.

For example, the tree depth in a decision tree model and the number of layers in an artificial neural network are typical hyperparameters. The performance of a model can drastically depend on the choice of its hyperparameters. A decision tree can yield good results for moderate tree depth and have very bad performance for very deep trees.

The choice of the optimal hyperparameters is more art than science, if we want to run it manually. Indeed, the optimal selection of the hyperparameter values depends on the problem at hand.

Since the algorithms, the goals, the data types, and the data volumes change considerably from one project to another, there is no single best choice for hyperparameter values that fits all models and all problems. Instead, hyperparameters must be optimized within the context of each machine learning project.

In this article, we’ll start with a review of the power of an optimization strategy and then provide an overview of four commonly used optimization strategies:

  • Grid search
  • Random search
  • Hill climbing
  • Bayesian optimization

Optimizing ML model parameters

A typical optimization procedure defines the possible set of hyperparameters and the metric to be maximized or minimized for that particular problem. Hence, in practice, any optimization procedure follows these classical steps:

  • 1) Split the data at hand into training and test subsets
  • 2) Repeat optimization loop a fixed number of times or until a condition is met:
    • a) Select a new set of model hyperparameters
    • b) Train the model on the training subset using the selected set of hyperparameters
    • c) Apply the model to the test subset and generate the corresponding predictions
    • d) Evaluate the test predictions using the appropriate scoring metric for the problem at hand, such as accuracy or mean absolute error. Store the metric value that corresponds to the selected set of hyperparameters
  • 3) Compare all metric values and choose the hyperparameter set that yields the best metric value

The question is how to pass from step 2d back to step 2a for the next iteration; that is, how to select the next set of hyperparameters, making sure that it is actually better than the previous set. We would like our optimization loop to move toward a reasonably good solution, even though it may not be the optimal one. In other words, we want to be reasonably sure that the next set of hyperparameters is an improvement over the previous one.

A typical optimization procedure treats a machine learning model as a black box. That means at each iteration for each selected set of hyperparameters, all we are interested in is the model performance as measured by the selected metric. We do not need (want) to know what kind of magic happens inside the black box. We just need to move to the next iteration and iterate over the next performance evaluation, and so on.

The key factor in all different optimization strategies is how to select the next set of hyperparameter values in step 2a, depending on the previous metric outputs in step 2d. Therefore, for a simplified experiment, we omit the training and testing of the black box, and we focus on the metric calculation (a mathematical function) and the strategy to select the next set of hyperparameters. In addition, we have substituted the metric calculation with an arbitrary mathematical function and the set of model hyperparameters with the function parameters.

In this way, the optimization loop runs faster and remains as general as possible. One further simplification is to use a function with only one hyperparameter to allow for an easy visualization. Below is the function we used to demonstrate the four optimization strategies. We would like to emphasize that any other mathematical function would have worked as well.

Bayesian optimization

The Bayesian optimization strategy selects the next hyperparameter value based on the function outputs in the previous iterations, similar to the hill climbing strategy. Unlike hill climbing, Bayesian optimization looks at past iterations globally and not only at the last one.

There are typically two phases in this procedure:

  • During the first phase, called warm-up, hyperparameter values are generated randomly. After a user-defined number N of such random generations of hyperparameters, the second phase kicks in.
  • In the second phase, at each iteration, a “surrogate” model of type P(output | past hyperparameters) is estimated to describe the conditional probability of the output values on the hyperparameter values from past iterations. This surrogate model is much easier to optimize than the original function. Thus, the algorithm optimizes the surrogate and suggests the hyperparameter values at the maximum of the surrogate model as the optimal values for the original function as well. A fraction of the iterations in the second phase is also used to probe areas outside of the optimal region. This is to avoid the problem of local maxima.


Figure 4: Bayesian optimization of the hyperparameter values on a [0, 10] range. The color gradient reflects the position in the generated sequence of hyperparameter candidates. Whiter points correspond to hyperparameter values generated earlier on in the process; red points correspond to hyperparameter values generated later on. The gray points are generated in the first random phase of the strategy.Figure 4 demonstrates that the Bayesian optimization strategy uses the warm-up phase to define the most promising area and then selects the next values for the hyperparameters in that area.


You can also see that intense red points are clustered closer to the maximum, while pale red and white points are scattered. This demonstrates that the definition of the optimal region is improved with each iteration of the second phase.

Summary

We all know the importance of hyperparameter optimization while training a machine learning model. Since manual optimization is time-consuming and requires specific expert knowledge, we have explored four common automatic procedures for hyperparameter optimization.

In general, an automatic optimization procedure follows an iterative procedure in which at each iteration, the model is trained on a new set of hyperparameters and evaluated on the test set. At the end, the hyperparameter set corresponding to the best metric score is chosen as the optimal set. The question is how to select the next set of hyperparameters, ensuring that this is actually better than the previous set.

We have provided an overview of four commonly used optimization strategies: grid search, random search, hill climbing, and Bayesian optimization. They all have pros and cons, and we have briefly explained the differences by illustrating how they work on a simple toy use case. Now you are all set to go and try them out in a real-world machine learning problem.


Tensorflow -


https://www.tensorflow.org/

TensorFlow is an end-to-end open source platform for machine learning. It has a comprehensive, flexible ecosystem of tools, libraries and community resources that lets researchers push the state-of-the-art in ML and developers easily build and deploy ML powered applications

The core open source library to help you develop and train ML models. Get started quickly by running Colab notebooks directly in your browser.



For Java



For Javascript

https://www.tensorflow.org/js/

TensorFlow.js is a JavaScript library for training and deploying models in the browser and on Node.js.

Develop ML models in JavaScript, and use ML directly in the browser or in Node.js.


Run existing models

Use off-the-shelf JavaScript models or convert Python TensorFlow models to run in the browser or under Node.js.
Use official TensorFlow.js models
Convert Python models
Retrain existing models

Retrain pre-existing ML models using your own data.
Use Transfer Learning to customize models
Develop ML with JavaScript

Build and train models directly in JavaScript using flexible and intuitive APIs.
Get started with TensorFlow.js



Tensorflow Tutorials

https://www.tensorflow.org/tutorials

The TensorFlow tutorials are written as Jupyter notebooks and run directly in Google Colab—a hosted notebook environment that requires no setup. Click the Run in Google Colab button.

The best place to start is with the user-friendly Keras sequential API. Build models by plugging together building blocks. After these tutorials, read the Keras guide.

This "Hello, World!" notebook shows the Keras Sequential API and model.fit.

This notebook collection demonstrates basic machine learning tasks using Keras.

These tutorials use tf.data to load various data formats and build input pipelines.

More

The Keras functional and subclassing APIs provide a define-by-run interface for customization and advanced research. Build your model, then write the forward and backward pass. Create custom layers, activations, and training loops.

Many Libraries and Extensions available free

Explore libraries to build advanced models or methods using TensorFlow, and access domain-specific application packages that extend TensorFlow. This is a sample of the tutorials available for these projects.



Keras - Python Deep Learning Library

https://keras.io/

Keras is a high-level neural networks API, written in Python and capable of running on top of TensorFlow, CNTK, or Theano. It was developed with a focus on enabling fast experimentation. Being able to go from idea to result with the least possible delay is key to doing good research.

Use Keras if you need a deep learning library that:

  • Allows for easy and fast prototyping (through user friendliness, modularity, and extensibility).
  • Supports both convolutional networks and recurrent networks, as well as combinations of the two.
  • Runs seamlessly on CPU and GPU.

Read the documentation at Keras.io.

Keras is compatible with: Python 2.7-3.6.


Guiding principles

  • User friendliness. Keras is an API designed for human beings, not machines. It puts user experience front and center. Keras follows best practices for reducing cognitive load: it offers consistent & simple APIs, it minimizes the number of user actions required for common use cases, and it provides clear and actionable feedback upon user error.

  • Modularity. A model is understood as a sequence or a graph of standalone, fully configurable modules that can be plugged together with as few restrictions as possible. In particular, neural layers, cost functions, optimizers, initialization schemes, activation functions and regularization schemes are all standalone modules that you can combine to create new models.

  • Easy extensibility. New modules are simple to add (as new classes and functions), and existing modules provide ample examples. To be able to easily create new modules allows for total expressiveness, making Keras suitable for advanced research.

  • Work with Python. No separate models configuration files in a declarative format. Models are described in Python code, which is compact, easier to debug, and allows for ease of extensibility.


Getting started: 30 seconds to Keras

The core data structure of Keras is a model, a way to organize layers. The simplest type of model is the Sequential model, a linear stack of layers. For more complex architectures, you should use the Keras functional API, which allows to build arbitrary graphs of layers.

Here is the Sequential model:

from keras.models import Sequential

model = Sequential()

Stacking layers is as easy as .add():

from keras.layers import Dense

model.add(Dense(units=64, activation='relu', input_dim=100))
model.add(Dense(units=10, activation='softmax'))

Once your model looks good, configure its learning process with .compile():

model.compile(loss='categorical_crossentropy',
              optimizer='sgd',
              metrics=['accuracy'])

If you need to, you can further configure your optimizer. A core principle of Keras is to make things reasonably simple, while allowing the user to be fully in control when they need to (the ultimate control being the easy extensibility of the source code).

model.compile(loss=keras.losses.categorical_crossentropy,
              optimizer=keras.optimizers.SGD(lr=0.01, momentum=0.9, nesterov=True))

You can now iterate on your training data in batches:

# x_train and y_train are Numpy arrays --just like in the Scikit-Learn API.
model.fit(x_train, y_train, epochs=5, batch_size=32)

Alternatively, you can feed batches to your model manually:

model.train_on_batch(x_batch, y_batch)

Evaluate your performance in one line:

loss_and_metrics = model.evaluate(x_test, y_test, batch_size=128)

Or generate predictions on new data:

classes = model.predict(x_test, batch_size=128)

Building a question answering system, an image classification model, a Neural Turing Machine, or any other model is just as fast. The ideas behind deep learning are simple, so why should their implementation be painful?

For a more in-depth tutorial about Keras, you can check out:

In the examples folder of the repository, you will find more advanced models: question-answering with memory networks, text generation with stacked LSTMs, etc.


Installation

Before installing Keras, please install one of its backend engines: TensorFlow, Theano, or CNTK. We recommend the TensorFlow backend.

You may also consider installing the following optional dependencies:

Then, you can install Keras itself. There are two ways to install Keras:

  • Install Keras from PyPI (recommended):
sudo pip install keras

If you are using a virtualenv, you may want to avoid using sudo:

pip install keras
  • Alternatively: install Keras from the GitHub source:

First, clone Keras using git:

git clone https://github.com/keras-team/keras.git

Then, cd to the Keras folder and run the install command:

cd keras
sudo python setup.py install

Configuring your Keras back end

By default, Keras will use TensorFlow as its tensor manipulation library. Follow these instructions to configure the Keras backend.


Support

You can ask questions and join the development discussion:

You can also post bug reports and feature requests (only) in GitHub issues. Make sure to read our guidelines first.


AI solution devops concepts



Containers make operation and management of AI solutions safer, easier



article - Benefits-of-containers-for-AI-workloads

The benefits of containerization for AI development include dependency management, environment consistency, scaling, resource consumption efficiency, isolation, reproducibility, security, latency control, versioning and cost management.

1. Dependency management and app segregation

2. Consistent environment

3. Scalability with Kubernetes

4. Resource efficiency compared to pure VMs w lower platform lockin

5. Isolation - reasonably well

6. Security - over apps in shared server

7. Latency with separate services for independent scaling with limited overhead

8. Version control w separation of app instances for different configurations

9. Reproducibility with snapshots of the container environment as needed

10. Reduced staff costs - maybe

Containers integrate well with CI/CD pipelines, facilitating automated testing and deployment. With less manual intervention from development to deployment, human errors should be decreased. These aspects reduce potential expenses related to bug fixes and downtime.



Kubernetes provides common management for AI solutions across environments and platforms



Kubeflow

https://www.kubeflow.org

https://github.com/kubeflow/


The Machine Learning Toolkit for Kubernetes

https://www.kubeflow.org/docs/about/kubeflow/

The Kubeflow project is dedicated to making deployments of machine learning (ML) workflows on Kubernetes simple, portable and scalable. Our goal is not to recreate other services, but to provide a straightforward way to deploy best-of-breed open-source systems for ML to diverse infrastructures. Anywhere you are running Kubernetes, you should be able to run Kubeflow.

Getting started with Kubeflow

Read the Kubeflow overview for an introduction to the Kubeflow architecture and to see how you can use Kubeflow to manage your ML workflow.

Follow the getting-started guide to set up your environment and install Kubeflow.

What is Kubeflow?

Kubeflow is the machine learning toolkit for Kubernetes. Learn about Kubeflow use cases.

To use Kubeflow, the basic workflow is:

Download and run the Kubeflow deployment binary.
Customize the resulting configuration files.
Run the specified script to deploy your containers to your specific environment.

You can adapt the configuration to choose the platforms and services that you want to use for each stage of the ML workflow: data preparation, model training, prediction serving, and service management.

You can choose to deploy your Kubernetes workloads locally, on-premises, or to a cloud environment.

Read the Kubeflow overview for more details.

The Kubeflow mission

Our goal is to make scaling machine learning (ML) models and deploying them to production as simple as possible, by letting Kubernetes do what it’s great at:

Easy, repeatable, portable deployments on a diverse infrastructure (for example, experimenting on a laptop, then moving to an on-premises cluster or to the cloud)
Deploying and managing loosely-coupled microservices
Scaling based on demand

Because ML practitioners use a diverse set of tools, one of the key goals is to customize the stack based on user requirements (within reason) and let the system take care of the “boring stuff”. While we have started with a narrow set of technologies, we are working with many different projects to include additional tooling.

Ultimately, we want to have a set of simple manifests that give you an easy to use ML stack anywhere Kubernetes is already running, and that can self configure based on the cluster it deploys into.

History

Kubeflow started as an open sourcing of the way Google ran TensorFlow internally, based on a pipeline called TensorFlow Extended. It began as just a simpler way to run TensorFlow jobs on Kubernetes, but has since expanded to be a multi-architecture, multi-cloud framework for running entire machine learning pipelines.


Roadmaps

To see what’s coming up in future versions of Kubeflow, refer to the Kubeflow roadmap.

The following components also have roadmaps:

Arena
Fairing
Kubeflow Pipelines
KF Serving
Katib
MPI Operator


Components of Kubeflow

https://www.kubeflow.org/docs/components/

Logical components that make up Kubeflow
Central Dashboard

The central user interface (UI) in Kubeflow
Metadata

Tracking and managing metadata of machine learning workflows in Kubeflow
Jupyter Notebooks

Using Jupyter notebooks in Kubeflow
Frameworks for Training

Training of ML models in Kubeflow
Hyperparameter Tuning

Hyperparameter tuning of ML models in Kubeflow
Pipelines

ML Pipelines in Kubeflow
Tools for Serving

Serving of ML models in Kubeflow
Multi-Tenancy in Kubeflow

Multi-user isolation and identity access management (IAM)
Miscellaneous

Miscellaneous Kubeflow components


How to get started using Kubeflow.

https://www.kubeflow.org/docs/started/

How to get started using Kubeflow.
Kubeflow Overview

How Kubeflow helps you organize your ML workflow
Installing Kubeflow

Overview of installation choices for various environments
Cloud Installation

Instructions for installing Kubeflow on a public cloud
Kubernetes Installation

Instructions for installing Kubeflow on an existing Kubernetes cluster
Workstation Installation

Instructions for installing Kubeflow on a workstation or server



PyTorch


https://pytorch.org/

An open source machine learning framework that accelerates the path from research prototyping to production deployment.

PyTorch enables fast, flexible experimentation and efficient production through a user-friendly front-end, distributed training, and ecosystem of tools and libraries.

pre-reqs

Python
Juptyer notebooks ???
Data Sources


https://pytorch.org/features/

TorchScript

With TorchScript, PyTorch provides ease-of-use and flexibility in eager mode, while seamlessly transitioning to graph mode for speed, optimization, and functionality in C++ runtime environments.

my_script_module.save("my_script_module.pt")

TorchScript

With TorchScript, PyTorch provides ease-of-use and flexibility in eager mode, while seamlessly transitioning to graph mode for speed, optimization, and functionality in C++ runtime environments.

Distributed Training

Optimize performance in both research and production by taking advantage of native support for asynchronous execution of collective operations and peer-to-peer communication that is accessible from Python and C++.


Mobile (Experimental)

PyTorch supports an end-to-end workflow from Python to deployment on iOS and Android. It extends the PyTorch API to cover common preprocessing and integration tasks needed for incorporating ML in mobile applications.


Tools & Libraries

An active community of researchers and developers have built a rich ecosystem of tools and libraries for extending PyTorch and supporting development in areas from computer vision to reinforcement learning.


Native ONNX Support

Export models in the standard ONNX (Open Neural Network Exchange) format for direct access to ONNX-compatible platforms, runtimes, visualizers, and more.


Cloud Partners

PyTorch is well supported on major cloud platforms, providing frictionless development and easy scaling through prebuilt images, large scale training on GPUs, ability to run models in a production scale environment, and more.

PyTorch Tutorials

https://pytorch.org/tutorials/

To learn how to use PyTorch, begin with our Getting Started Tutorials. The 60-minute blitz is the most common starting point, and provides a broad view into how to use PyTorch from the basics all the way into constructing deep neural networks.

Some considerations:

We’ve added a new feature to tutorials that allows users to open the notebook associated with a tutorial in Google Colab. Visit this page for more information.
If you would like to do the tutorials interactively via IPython / Jupyter, each tutorial has a download link for a Jupyter Notebook and Python source code.
Additional high-quality examples are available, including image classification, unsupervised learning, reinforcement learning, machine translation, and many other applications, in PyTorch Examples.
You can find reference documentation for the PyTorch API and layers in PyTorch Docs or via inline help.
If you would like the tutorials section improved, please open a github issue here with your feedback.
Check out our PyTorch Cheat Sheet for additional useful information.
Finally, here’s a link to the PyTorch Release Notes


Cloudera ML Workbench Course Overview

https://university.cloudera.com/instructor-led-training/cloudera-data-scientist-training

$3200 for 4 days online


The workshop includes brief lectures, interactive demonstrations, hands-on exercises, and discussions covering topics including:

• Overview of data science and machine learning at scale

• Overview of the Hadoop ecosystem

• Working with HDFS data and Hive tables using Hue

• Introduction to Cloudera Data Science Workbench

• Overview of Apache Spark 2

• Reading and writing data

• Inspecting data quality

• Cleansing and transforming data

• Summarizing and grouping data

• Combining, splitting, and reshaping data

• Exploring data

• Configuring, monitoring, and troubleshooting Spark applications

• Overview of machine learning in Spark MLlib

• Extracting, transforming, and selecting features

• Building and evaluating regression models

• Building and evaluating classification models

• Building and evaluating clustering models

• Cross-validating models and tuning hyperparameters

• Building machine learning pipelines

• Deploying machine learning models


Technologies

Participants gain practical skills and hands-on experience with data science tools including:

• Spark, Spark SQL, and Spark MLlib

• PySpark and sparklyr

• Cloudera Data Science Workbench (CDSW)

• Hue


Prerequisites

Workshop participants should have a basic understanding of Python or R and some experience exploring and analyzing data and developing statistical or machine learning models. Knowledge of Hadoop or Spark is not required.


AIAAS - MLAAS - ML Services


Smart Assistant Services Agents ( SASA ) - provide real-time or batch decision analysis

DMX used car pricing assistant

Carbon Offset trading assistant - Optional AMM feature



Hyper.now example MLAAS

https://www2.microstrategy.com/producthelp/Current/HyperNowService/en-us/Content/tech_overview.htm#:~:text=Hyper.-,Now%20Technology%20Overview,throughout%20their%20standard%20business%20workflows.

HyperIntelligence is a proprietary analytics technology offered by MicroStrategy that enables users to consume contextual data and analytics within the websites, emails, enterprise applications, and web-based tools leveraged throughout their standard business workflows. This technology allows users to augment their workflows with relevant insights and one-click actions delivered via HyperIntelligence cards, which function as micro-dashboards that are automatically generated to convey relevant data and analytics associated with defined keyword attributes.

When such attribute(s) that you identify as keywords appear in context as supported by the HyperIntelligence product, the keyword triggers the associated card(s) to appear. The context in which HyperIntelligence cards may appear when triggered by defined keywords includes:

  • Within web-based content such as emails, websites, applications, or other browser-accessible instances;
  • Within emails, websites, or web applications;
  • Within Microsoft Office web-based applications such as Microsoft Excel; and
  • As notifications populated on Apple and Android mobile devices.

The contextual capabilities offered by Hyper.Now are comparable to the capabilities offered by the associated HyperIntelligence products offered in the enterprise MicroStrategy product catalog.



Generative Ai and Large Language Model Concepts



llm-compact-guide-to-large-language-models-databricks-2024.pdf file

some quotes from the guide >>

ChatGPT provides a nice user interface (or API) where users can feed prompts to one of many models (GPT-3.5, GPT-4, and more) and typically get a fast response. These are among the highest-performing models, trained on enormous data sets, and are capable of extremely complex tasks both from a technical standpoint, such as code generation, as well as from a creative perspective like writing poetry in a specific style.

open-source LLMs

Communities like Hugging Face gather hundreds of thousands of models from contributors that can help solve tons of specific use cases such as text generation, summarization and classification.

Databricks enhances MLFlow for specific use cases

On Databricks, for example, we’ve made improvements to open source frameworks like MLflow to make it very easy for someone with a bit of Python experience to pull any Hugging Face transformer model and use it as a Python object. Oftentimes, you can find an open source model that solves your specific problem that is orders of magnitude smaller than ChatGPT, allowing you to bring the model into your environment and host it yourself. This means that you can keep the data in your control for privacy and governance concerns as well as manage your costs

Smaller open-source LLMs target specific use cases with specific, private data

Another huge upside to using open source models is the ability to fine-tune them to your own data. Since you’re not dealing with a black box of a proprietary service, there are techniques that let you take open source models and train them to your specific data, greatly improving their performance on your specific domain. We believe the future of language models is going to move in this direction, as more and more organizations will want full control and understanding of their LLMs

Larger LLMs are adding options to target specific private data sets but their strength is the larger training data sets



LLM AI Security & Governance Checklist-owasp.pdf. file



LLMs_in_Production_v2.pdf. link book 2024

compact-guide-to-large-language-models.pdf


More on Gen AI and LLM resources


Open LLMs: Llama, Anthropic, Claude and more




Governance and Ethical AI 


HBR Practical Guide to Building Ethical AI - 2020 

https://hbr.org/2020/10/a-practical-guide-to-building-ethical-ai

Practical Guide to Building Ethical AI ** HBR pdf

WEForum.org - ensure AI is ethical - 8 tips

https://www.weforum.org/agenda/2020/01/8-ways-to-ensure-your-companys-ai-is-ethical/


Why Govern AI ?  Risk Management

https://www.forbes.com/sites/vipinbharathan/2023/06/25/guardrails-for-ai-what-is-possible-today/?sh=150eef613a0d

“Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” The Center for AI Safety published this single statement as an open letter on May 30, 2023.


OWASP and AI Governance of LLM ( current technology )

https://www.linkedin.com/posts/brunoschneider_llm-ai-security-and-governance-checklist-activity-7140603869058723840-GGDI?utm_source=share&utm_medium=member_desktop

LLM AI Security & Governance Checklist-owasp.pdf. link

LLM AI Security & Governance Checklist-owasp.pdf.  file




Good Governance for AI - What does that Look Like?



Why can’t you trust AI?

Good Governance is missing


What does good governance for AI look like?

Independent Proof of Governance ( Sybal )


What does Independent Proof of Governance deliver for AI?

Automated, Accurate Verification of the entire AI process ensuring compliance within defined guardrails


From the domain and problem specification, to the models, the training data, the tuning, the input data, execution and output, the entire process needs verification to ensure compliance within the defined guardrails



Proposed Principles for Responsible AI


 

Advice not actions for many use cases

Transparent

Explainable

Traceable

Accurate

Controlled 

Governed

Accountable

Fair

Reliable

Private

Inclusive

Secure

Valuable

Safe - Risk management

Cost-effective

Accessible


Vipin Bharathan on Proposed AI Guardrails Framework - Forbes

https://www.forbes.com/sites/vipinbharathan/2023/06/25/guardrails-for-ai-what-is-possible-today/?sh=150eef613a0d


The architecture of an emerging startup, Modguard, in the accompanying diagram, gives us a path towards compliance with these laws




EU AI Regulations


WEF review of EU AI regs

https://www.weforum.org/agenda/2023/06/european-union-ai-act-explained/


EU Overview of proposed EU AI Regulations

https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai


EU Proposed AI Regulations - June 2023

https://eur-lex.europa.eu/resource.html?uri=cellar:e0649735-a372-11eb-9585-01aa75ed71a1.0001.02/DOC_1&format=PDF

Covers

Concepts, Needs, Responsible Parties, Responsibitles, Use Cases, Rights, Proposed Regulations and more


EY - EU AI Act and what it means for your business

https://www.ey.com/en_ch/forensic-integrity-services/the-eu-ai-act-what-it-means-for-your-business

  • The EU AI Act brings strict requirements, also for organizations which have not had to deal with model management until now.
  • As a first step, organizations should gain an overview,  build a repository of all models and implement a model management.
  • Even though regulation is not final, it is clearly on the horizon and time should be used to prepare.


Brookings report - EU and US Diverge on AI Regulation - April 2023

https://www.brookings.edu/articles/the-eu-and-us-diverge-on-ai-regulation-a-transatlantic-comparison-and-steps-to-alignment/




G7 Survey on Guiding Principles for Generative AI - Oct 2023

https://digital-strategy.ec.europa.eu/en/news/commission-gathers-views-g7-guiding-principles-generative-artificial-intelligence


News release on EU AI Regulation Act

https://www.europarl.europa.eu/news/en/headlines/society/20230601STO93804/eu-ai-act-first-regulation-on-artificial-intelligence


US Approach to Responsible AI



Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence

https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/

Purpose

Policies and Principles includes many social goals

Definitions with instructions on establishing guidelines for Responsible AI

Secretary of Commerce has a lead role in defining rules for AI model and service management

" The Secretary of Commerce, in consultation with the Secretary of State, the Secretary of Defense, the Secretary of Energy, and the Director of National Intelligence, shall define,"

Assignments > Objectives with normal time limits less than 1 year for reporting on impacts, goals, strategies by agency

Identifies responsibity concepts for businesses in building, using AI in commerce


Fact Sheet on Executive Order for Responsible AI

https://www.whitehouse.gov/briefing-room/statements-releases/2023/10/30/fact-sheet-president-biden-issues-executive-order-on-safe-secure-and-trustworthy-artificial-intelligence/

New Standards for AI Safety and Security

developers of the most powerful AI systems share their safety test results and other critical information with the U.S. government

 standards, tools, and tests to help ensure that AI systems are safe, secure, and trustworthy

Protect against risks to people, organizations, ecosystems, systems

Reduce fraud, deception, misinformation 

Protect Privacy and Rights

Promote Innovation

Promote American Leadership ( or better partnerships ?? )

Effective governance for Responsible AI







NIST Proposed AI Risk Mgt Framework

https://nvlpubs.nist.gov/nistpubs/ai/nist.ai.100-1.pdf



AI Notes including Governance

AI-notes1 gdoc





Potential Value Opportunities



RedPoint Global concepts to deliver business results from ML

https://www.eweek.com/it-management/how-to-use-machine-learning-to-drive-real-value?utm_medium=email&utm_campaign=B2B_NL_EP_20191122_AR1&mkt_tok=eyJpIjoiTm1GaE0yTmxPR1ExTW1SaCIsInQiOiJES1dub1FsTXRETnNtbGlIZVRHQWtNaWFFM0YxektRa0tKbmVSSzNQNnVyV0NLcGRIZFZMUVFpSUViMFwvbFdab21rb1gxUUkwZGZJYWZNVTd3a2pXcGM1RVdmZ2VaUFFWUGs3RU11VHhSXC9ubkdhWXh3QXhkcWlaUHBwcEp2b3ZPIn0%3D


When paired with a persistent, real-time, single customer record, AI and automated machine learning platforms can be utilized to meet those business goals, increase revenue and fundamentally change the way brands communication with customers.

Data Point No. 1: Machine learning should drive revenue.

The ultimate goal of machine learning shouldn’t be a flashy, futuristic example but instead a system to drive revenue and results for the business. The result of effective machine learning isn’t likely a robot, chatbot or facial recognition tool –
it’s machine learning-driven programs that are embedded behind the scenes, driving intelligent decisions for optimized customer engagement.

Data Point No. 2: Having one model–or even many–is not enough.

Organizations need many models running and working in real time to truly make machine learning work for their needs. For future-forward organizations, intelligence and analysis needs to be embedded, so instead of using one model, multiple in-line analytic models can incrementally adjust and find opportunities for growth. These fleets of ML models can optimize business functions and drive associated revenues.

Data Point No. 3: When applied in silos, machine learning is not as effective.

Today’s consumer is omnichannel. Businesses must forego the traditional channel-specific “batch and blast” approach that sufficed when customer choice was limited and the buying journey followed a mostly straight-line path. Today’s customer journey is dynamic, and the learning applied to the customer relationship should be, as well. Machine learning is particularly well-suited to solving these multidimensional problems.

Data Point No. 4: Analytics are only intelligent when models are up to date.

News flash: Machine-learning models age and can quickly become stale. For this reason, organizations must consistently rebuild and retrain models using today’s data. In fact, models should be developed, updated and even applied in real-time on an ongoing testing basis so that businesses can truly capitalize on the moment of interaction. This is most effective in a closed loop system that continually looks for incremental changes to optimize.

Data Point No. 5: You don’t need to be a data scientist to benefit from machine learning.

When models are configured correctly, they will run 24/7 looking for opportunities within the data, set up and managed by marketers. These systems can be set once and guided to produce the specific business metrics needed. With every record tracked in the system, insights are pulled easily, and the recommendations can be made automatically. Businesses should focus on producing continually updated data and let the automation tools use machine learning to drive greater revenue.

Data Point No. 6: Summary: The power is in your hands.

Machine learning has the power to fully transform an enterprise. Therefore, it’s natural for business leaders to get lost in the hype and lose sight of the real value it can deliver day-to-day. The truth is, the real value of machine learning is that it allows businesses to try new things, amplify creative strengths, reveal new discoveries and enable collaboration across the organization. However, these benefits will only be realized once organizations get past the hype and are willing to walk into the weeds.



Practical DS / ML Solutions

file:///C:/Users/Jim%20Mason/Google%20Drive/_docs/howto/apps/Data-Science-in-Practice-Applications-2019.pdf


Recommender Systems

Recommender systems, also known as recommender engines, are one of the mostwell-known applications of data science. Recommender systems are a subclass ofinformation filtering systems, systems that cut through the noise of all options andpresent users with just the subset of options they’ll find appealing. The data beingfiltered can range from products on an e-commerce site to dating matches thatappear as you search for ‘the one.’Recommender systems offer a more intelligent approach to information filteringthan a simple search algorithm by introducing users to items they might not haveotherwise discovered. Recommender systems generally take either a collaborativeor content-based approach to filtering. Collaborative filtering considers a user’sprevious behavior, as well as the behavior of similar users. Content-based filteringprovides recommendations based on discrete attributes or assigned characteristics.

WHAT IS A RECOMMENDER SYSTEM?

A model that filters information to present users with a curated subset of options they’re likely to find appealing

HOW DOES IT WORK?Generally via a collaborative approach (considering user’s previous behavior) or content based approach (based on discrete assigned characteristics)

WHAT IS A REAL USE CASE?Tendril uses recommendation models to match eligible customers with new or existing energy products


Credit or Risk Scoring

If you have ever applied for a credit card or a loan, you’re likely already familiar with the concept of credit scoring. What you may be less aware of is the set of decision management rules evaluating how likely an applicant is to repay debts behind the scenes.

WHAT IS CREDIT SCORING?A model that determines an applicant’s creditworthiness for a mortgage, loan or credit card

HOW DOES IT WORK?A set of decision management rules evaluates how likely anapplicant is to repay debts

WHAT IS A REAL USE CASE?A financial services company uses machine learning models to reachprospective customers that may have been overlooked bytraditional banking institutions

We developed complex statistical and machine learning models to enable smarter lending decisions,” explains the Director of Business Lending . “By getting creative with ourapproach and adopting innovative technologies, we’ve been able to reinvent how both consumers and businesses obtain loans. This has allowed us to reachprospective customers that in the past may have been overlooked by traditionalbanking institutions.”


Dynamic Pricing  ( ?? see Robinson-Patman Act - see airline seat pricing )

Businesses use dynamic pricing algorithms to model rates as a function of supply,demand, competitor pricing, and exogenous factors (e.g. weather or time). Many fields, from airline travel to athletics admission ticketing, employ dynamic pricingto maximize expected revenue. The nuts and bolts of dynamic pricing strategiesvary widely, though generalized linear models and classification trees are populartechniques for estimating the “right” (lowest/highest) price that consumers arewilling to pay for a book, a flight, or a cab

WHAT IS DYNAMIC PRICING?Modeling price as a function of supply, demand, competitor pricingand exogenous factors

HOW DOES IT WORK?Generalized linear models and classification trees arepopular techniques for estimating the “right” price to maximizeexpected revenue

WHAT IS A REAL USE CASE?A car sharing company uses dynamic pricing models to suggest prices to the peoplewho list and rent out cars


Customer Churn

Churn rate describes the rate at which customers abandon a product or service.Understanding customers’ likelihood to churn is particularly important forsubscription-based models; everything ranging from traditional cable or gymmemberships to recently popularized monthly subscription boxes.Data scientists looking to predict customer churn may consider a variety of algorithmsfor the job, such as support vector machines, random forest, or k-nearest-neighbors.Beyond the accuracy of a given model, data scientists must also balance the tradeoffbetween precision (correctly predicting a churning customer) and recall (how manypredictions were actually successful). So what’s better? Classifying every churningcustomer but occasionally mislabeling a non-churning customer? Or identifying fewerchurning customers, but not mislabeling non-churners? It’s a difficult decision thatrequires in-depth knowledge of the business case and years of experience.

WHAT IS CUSTOMER CHURN?Predicting which customers are going to abandon a product or service

HOW DOES IT WORK?Data scientists may consider using support vector machines, random forest or k-nearest-neighbors algorithms

WHAT IS A REAL USE CASE?EAB combines data from transcripts, standardized test scores, demographics and more to identify students at risk of not graduating


Fraud Detection

Financial technology, or ‘FinTech,’ companies offer financial services like banking, investing, and payment processing via software, rather than through traditional banking institutions. Companies processing massive volumes of financial transactions also need a quantifiable way to detect and prevent fraudulent transactions from being processed.

WHAT IS FRAUD DETECTION?Detecting and preventing fraudulent financial transactions from being processed

HOW DOES IT WORK?Fraud detection is a binary classification problem: “is this transaction legitimate or not?”

WHAT IS A REAL USE CASE?Via SMS Group uses a combination of complex data lookups and decision algorithms written in R and implemented in PHP to assess whether a loan applicant is fraudulent



AI Solution Architecture Models


https://www.linkedin.com/posts/miguelgfierro_ai-machinelearning-datascience-activity-7045994122091859968-DdG0?utm_source=share&utm_medium=member_desktop


One of the most important skills for people in AI is being able to architect solutions for companies. This one-pager covers everything you need.

- Custom AI:
At the bottom, you have the ML platforms, data analytics, and data governance, that allows Data Scientists and Data Engineers to create custom AI solutions.

- General AI Services:
The Cognitive Services help you build AI applications by calling APIs. They include generalistic solutions for computer vision, language, speech, and the Azure OpenAI Service.

- Applied AI Services:
On top, we have specific AI solutions like Cognitive Search or Form Recognizer. They are niche applications for Software Developers.

- No code AI:

See Google Cloud for No code AI

The Power Platform helps any person build AI applications. If you can operate PowerPoint and Excel, you can operate the solutions in the Power Platform.

- Industry applications:
Apart from generalistic solutions, there are vertical solutions specialized in industries. Healthcare, agriculture, marketing, etc.

- Open source:
There is a large suite of open-source libraries that help AI engineers and software developers to build AI: PyTorch, ONNX, LightGBM, Recommenders, SynapseML, DeepSpeed, etc.

This is Microsoft cloud business in one page and involves tens of thousands of people serving customers.

Designed by Adrián González Sánchez


Potential Challenges


CX is now multi-channel contact points direct and indirect to an organization creating data integration, validation challenges

Need highly engaged, customized user interfaces by user channels of choice ( eg text here, voice there, docs there etc )

Sometimes batch but more often real-time information and decision making support needed in user interactions



Predictions


AI will decide to improve by reinventing its form as organic beings automating growth, nutrition, with faster and more complex processing modes



ML and AI Governance



Trusted ML apps with ZKP - Daniel Szego

Tokenized Zero Knowledge Machine Learning and Its Applications-Daniel-Szego-2024.pdf.   link

Tokenized Zero Knowledge Machine Learning and Its Applications-Daniel-Szego-2024.pdf. file

Hi Daniel,

Thanks for your article and demo on Tokenized Zero Knowledge ML apps. Really clear, well done. I didn't connect a wallet but did run the simple online tests.

Can I interest you in doing a 1 hour presentation to the Hyperledger Meetup and Public Sector groups on your article? I think it's really important for the community. I lead the Boston and Public Sector groups.

Thanks

Jim


Candidate Solutions



Google Foundation Models based on PALM2 FOR AI solutions

https://seekingalpha.com/article/4613576-google-most-promising-source-of-growth-right-now?mailingid=31901489&messageid=2800&serial=31901489.19951&utm_campaign=rta-stock-article&utm_medium=email&utm_source=seeking_alpha&utm_term=31901489.19951



AI Tools List - 2023


https://www.linkedin.com/posts/philippederidder_innovation-design-ux-activity-7044345913770700800-6_V7?utm_source=share&utm_medium=member_desktop


https://boardofinnovation.notion.site/AI-Tools-for-Innovators-7a80ab30bcfd4a15846436aa347d5af2 


table


Site for more AI tools

https://theresanaiforthat.com/



No Code AI Solutions, Tools



Google AppSheets

https://www.youtube.com/watch?v=MlJX0X9qerc



No Code AI from LinuxFoundation.org


https://www.linuxfoundation.org/projects


https://www.linuxfoundation.org/press/open-metaverse-foundation



Cloudera Data Science Workbench


With Python, R, and Scala directly in the web browser, Cloudera Data Science Workbench (CDSW) delivers a self-service experience data scientists will love. Download and experiment with the latest libraries and frameworks in customizable project environments that work just like your laptop. Access any data, anywhere—from cloud object storage to data warehouses, Cloudera Data Science Workbench provides connectivity not only to CDH and HDP but also to the systems your data science teams rely on for analysis.




MLFlow.org 

https://mlflow.org


MLflow is a platform to streamline machine learning development, including tracking experiments, packaging code into reproducible runs, and sharing and deploying models. MLflow offers a set of lightweight APIs that can be used with any existing machine learning application or library (TensorFlow, PyTorch, XGBoost, etc), wherever you currently run ML code (e.g. in notebooks, standalone applications or the cloud). MLflow's current components are:

  • MLflow Tracking: An API to log parameters, code, and results in machine learning experiments and compare them using an interactive UI.
  • MLflow Projects: A code packaging format for reproducible runs using Conda and Docker, so you can share your ML code with others.
  • MLflow Models: A model packaging format and tools that let you easily deploy the same model (from any ML library) to batch and real-time scoring on platforms such as Docker, Apache Spark, Azure ML and AWS SageMaker.
  • MLflow Model Registry: A centralized model store, set of APIs, and UI, to collaboratively manage the full lifecycle of MLflow Models.


Recommended ML Basic Courses from Datacamp

https://learn.datacamp.com/


These personalized recommendations are based on your skill gaps. Start one now to improve your skills.

STATISTICS FUNDAMENTALS

MACHINE LEARNING

DATA AND SUMMARIES







Step-by-step guide for Example



sample code block

sample code block
 



Recommended Next Steps



*