Tuesday, June 25, 2024

 Challenges in storing drone data.

Unlike traditional data architectures involving an online transaction processing data system, where atomicity, consistency, isolation and durability guarantees enable proper inventory registration and calculations, and an online analytical processing system with its reporting, temporal aggregation and analysis capabilities, the lines between transactions and analysis blur for near real-time data analysis of drones as both the inventory and the associated processing must continuously adapt. It could probably be compared to data and event pipelines built for large-scale commercial applications such as AirBnb and with the potential to become real-time processing by eliminating the cost for network latency and storage access. 

Flights for drones remain undisturbed whether stationary or linear until the next update from the controller or local flight path change determination. Clearly, the capabilities of the drone units might vary from fleet to fleet. Individual drone, degree of freedom, motion capabilities, and variety of non-flight actions that the unit can take must need to be differentiated so they can be used selectively. When the entire fleet moves the same as a single unit in the fleet, there are far fewer updates to the data stored for the drones than otherwise. With updates varying from few to large scale, rare to frequent, the data and the events generated must be handled for any volume and rate. At all processing modules, virtualizations that cover the variety possible from the model and type of drone units mandate consistency in the data types used to represent them, so that the interface remains clean with just the right levers and validations for the associated concerns. A unified Api is not just an evolutionary step for drone data management but also a necessity from the start. It might be customary to build data pipeline on Spark and Scala that aids bookkeeping and double-entries for tallying so that the accounting information can then be segregated. Drone data and events have a lot of similarity to edge computing and streaming data pipelines and stores. The need for these approaches must be balanced by the hard performance goals for regular routines.

Cloud databases such as Azure Cosmos DB provide such a general purpose and easily programmable starter stores which can scale and keep up its amazing response times for requests. They remain starter stores because there is little forethought to the specializations in routines needed with data access and storage of drone data and metadata. That said, location information that is rapidly updated for drones can be continuously updated and maintained in these cloud databases. There is convenience in adding various dimensions to the same store as warehouses have taught and such a schema works well for drones as well. But as noted earlier, specializations in components may mandate hybrid data architectures that might not all fit nicely within a single product even if the product were a starting point.

In conclusion, drones data architectures must be articulated with the full palette of storage options, microservices, shared architectures, events, pipelines and automation that may become just as big as Master Data Management systems are today. The good news is that this needs to be done only once for multi-tenant systems.


Monday, June 24, 2024

 A vector database and search for positioning drones involves the following steps:

The first step would be to install all the required packages and libraries. We use Python in this sample:

import warnings 

warnings.filterwarnings(‘ignore’) 

from datasets import load_dataset 

from pinecone import Pinecone, ServerlessSpec 

from DLAIUtils import Utils 

import DLAIUtils  

import os 

import time 

import torch 

From tqdm.auto import tqdm 

We assume the elements are mapped as embeddings in a 384-dimensional dense vector space. 

A sample query would appear like this: 

query = `what is node nearest this element?` 

xq = model.encode(query) 

xq.shape 

(384,) 

The next step is to set up the Pinecone vector database to upsert embeddings into it. These database index vectors make search and retrieval easy by comparing values and finding those that are most like one-another 

utils = Utils() 

PINECONE_API_KEY = utils.get_pinecone_api_key() 

if INDEX_NAME in [index.name for index in pinecone.list_indexes()]: 

       pinecone.delete_index(INDEX_NAME) 

print(INDEX_NAME) 

pinecone.create_index(name=INDEX_NAME, dimension=model.get_sentence_embedding_dimension(), metric=’cosine’,spec=ServerlessSpec(cloud=’aws’, region=’us-west-2’)) 

index = pinecone.Index(INDEX_NAME) 

print(index) 

Then, the next step is to create embeddings for all the elements in the sample space and upsert them to Pinecone. 

batch_size=200 

vector_limit=10000 

elements=element[:vector_limit] 

import json 

for i in tqdm(range(0, len(elements), batch_size)): 

        i_end = min(i+batch_size, len(elements)) 

        ids = [str(x) for x in range(i, i_end)] 

        metadata = [{‘text’: text} for text in elements[i:i_end]] 

        xc = model.encode(elements[i:i_end]) 

        records = zip(ids, xc, metadata) 

        index.upsert(vectors=records) 

index.describe_index_stats() 

Then the query can be run on the embeddings and the top matches can be returned. 

def run_query(query): 

        embedding = model.encode(query).tolist() 

        results = index.query(top_k=10, vector=embedding, include_metadata=True, include_value) 

        for result in results[‘matches’]: 

                print(f”{round(result[‘score’], 2)}: {result[‘metadata’][‘node’]}”) 

run_query(“what is node nearest this element?”) 

With this, the embeddings-based search over elements is ready. In Azure, cosmos DB offers a similar semantic search and works as a similar vector database. 

The following code outlines the steps using Azure AI Search 

# configure the vector store settings, vector name is in the index of the search

endpoint: str = "<AzureSearchEndpoint>"

key: str = "<AzureSearchKey>"

index_name: str = "<VectorName>"

credential = AzureKeyCredential(key)

client = SearchClient(endpoint=endpoint,

                      index_name=index_name,

                      credential=credential)


# create embeddings 

embeddings: AzureOpenAIEmbeddings = AzureOpenAIEmbeddings(

    azure_deployment=azure_deployment,

    openai_api_version=azure_openai_api_version,

    azure_endpoint=azure_endpoint,

    api_key=azure_openai_api_key,

)

# create vector store

vector_store = AzureSearch(

    azure_search_endpoint=endpoint,

    azure_search_key=key,

    index_name=index_name,

    embedding_function=embeddings.embed_query,

)

# create a query

docs = vector_store.similarity_search(

    query=userQuery,

    k=3,

    search_type="similarity",

)

collections.insert_many(docs)

reference: https://github.com/ravibeta/Node-Element-Predictions


Sunday, June 23, 2024

 Some of fleet management data science algorithms are captured via a comparison table of well-known data mining algorithms as follows:

Data Mining Algorithms Description Use case

Classification algorithms This is useful for finding similar groups based on discrete variables

It is used for true/false binary classification. Multiple label classifications are also supported. There are many techniques, but the data should have either distinct regions on a scatter plot with their own centroids or if it is hard to tell, scan breadth first for the neighbors within a given radius forming trees or leaves if they fall short.

Useful for categorization of fleet path changes beyond the nomenclature. Primary use case is to see clusters of service request that match based on features. By translating to a vector space and assessing the quality of cluster with a sum of square of errors, it is easy to analyze large number of changes as belonging to specific clusters for management perspective.

Regression algorithms This is very useful to calculate a linear relationship between a dependent and independent variable, and then use that relationship for prediction. Fleet path changes demonstrate elongated scatter plots in specific categories. Even when the path changes come demanding different formations in the same category, the reorientation times are bounded and can be plotted along the timeline. One of the best advantages of linear regression is the prediction about time as an independent variable. When the data point has many factors contributing to their occurrence, a linear regression gives an immediate ability to predict where the next occurrence may happen. This is far easier to do than coming up with a model that behaves like a good fit for all the data points.

Segmentation algorithms A segmentation algorithm divides data into groups or clusters or items that have similar properties. Path change stimuli segmentation based on fleet path change feature set is a very common application of this algorithm. It helps prioritize the response to certain stimuli.

Association algorithms This is used for finding correlations between different attributes in a data set Association data mining allows these users to see helpful messages such as “stimulii who caused a path change for this fleet type also caused a path change for this other fleet formation”

Sequence Analysis Algorithms This is used for finding groups via paths in sequences. A Sequence Clustering algorithm is like a clustering algorithm mentioned above but instead of finding groups based on similar attributes, it finds groups based on similar paths in a sequence.  A sequence is a series of events. For example, a series of web clicks by a user is a sequence. It can also be compared to the IDs of any sortable data maintained in a separate table. Usually, there is support for a sequence column. The sequence data has a nested table that contains a sequence ID which can be any sortable data type. This is very useful to find sequences of fleet path changes opened across customers. Generally, a transit failure could result in a cascading failure across the transport network. This sort of sequence determination in a data driven manner helps find new sequences and target them actively even suggesting the same to the stimulii who cause ath changes to the fleet formations so that they can be better prepared for failures across relays.


Sequence Analysis also helps with interactive formation changes as described here.


Outliers Mining Algorithm Outliers are the rows that are most dissimilar. Given a relation R(A1, A2, ..., An), and a similarity function between rows of R, find rows in R which are dissimilar to most point in R. The objective is to maximize dissimilarity function in with a constraint on the number of outliers or significant outliers if given. 

The choices for similarity measures between rows include distance functions such as Euclidean, Manhattan, string-edits, graph-distance etc and L2 metrics. The choices for aggregate dissimilarity measures is the distance of K nearest neighbors, density of neighborhood outside the expected range and the attribute differences with nearby neighbors The steps to determine outliers can be listed as: 1. Cluster regular via K-means, 2.  Compute distance of each tuple in R to nearest cluster center and 3. choose top-K rows, or those with scores outside the expected range. Finding outliers is sometimes humanly impossible because the number of path changes can be quite high. Outliers are important to discover new strategies to encompass them. If there are numerous outliers, they will significantly increase costs. If they were not, then the patterns help identify efficiencies.

Decision tree This is probably one of the most heavily used and easy to visualize mining algorithms. The decision tree is both a classification and a regression tree. A function divides the rows into two datasets based on the value of a specific column. The two list of rows that are returned are such that one set matches the criteria for the split while the other does not. When the attribute to be chosen is clear, this works well. A Decision Tree algorithm uses the attributes of the external stimulii to make a prediction such as the reorientation time on a next path change. The ease of visualization of split at each level helps throw light on the importance of those attributes.  This information becomes useful to prune the tree and to draw the tree

Logistic Regression This is a form of regression that supports binary outcomes. It uses statistical measures, is highly flexible, takes any kind of input and supports different analytical tasks. This regression folds the effects of extreme values and evaluates several factors that affect a pair of outcomes. Path changes based on stimulii category can be used to predict the likelihood of a path change from a category of stimulii. It can also be used for finding repetitions in requests 

Neural Network This is a widely used method for machine learning involving neurons that have one or more gates for input and output. Each neuron assigns a weight usually based on probability for each feature and the weights are normalized across resulting in a weighted matrix that articulates the underlying model in the training dataset. Then it can be used with a test data set to predict the outcome probability. Neurons are organized in layers and each layer is independent of the other and can be stacked so they take the output of one as the input to the other. Widely used for SoftMax classifier in NLP associated with fleet path changes. Since descriptions of stimulii, fleet formation changes, path adjustments and adjustment time to modified path and formation captured by spatial and temporal are conformant to narratives with metric-like quantizations, Natural Language Processing could become a significant part of the data mining and ML portfolio

Naïve Bayes algorithm This is probably the most straightforward statistical probability-based data mining algorithm compared to others.

The probability is a mere fraction of interesting cases to total cases. Bayes probability is conditional probability which adjusts the probability based on the premise. This is widely used for cases where conditions apply, especially binary conditions such as with or without. If the input variables are independent, their states can be calculated as probabilities, and if there is at least a predictable output, this algorithm can be applied. The simplicity of computing states by counting for class using each input variable and then displaying those states against those variables for a give value, makes this algorithm easy to visualize, debug and use as a predictor.  

Plugin Algorithms Several algorithms get customized to the domain they are applied to resulting in unconventional or new algorithms. For example, a hybrid approach on association clustering can benefit determining relevant associations when the matrix is quite large and has a large tail of irrelevant associations from the cartesian product. In such cases, clustering could be done prior to association to determine the key items prior to this market-basket analysis. Fleet path changes are notoriously susceptible to apply with variations even when pertaining to the same category. These path changes do not have pre-populated properties from a template, and spatial and temporal changes can vary drastically along one or both. Using a hybrid approach, it is possible to preprocess these path changes with clustering before analyzing such as with association clustering. 

Simultaneous classifiers and regions-of-interest regressors Neural nets algorithms typically involve a classifier for use with the tensors or vectors. But regions-of-interest regressors provide bounding-box localizations. This form of layering allows incremental semantic improvements to the underlying raw data. Fleet path changes are time-series data and as more and more are applied, specific time ranges become as important as the semantic classification of the origin of path changes and their descriptions. Using this technique, underlying issues can be discovered as tied to internal or external factors. The determination of root cause behind a handful of path changes is valuable information.


Collaborative filtering Recommendations include suggestions for a knowledge base, or to find model service requests. To make a recommendation, first a group sharing similar taste is found and then the preferences of the group is used to make a ranked list of suggestions. This technique is called collaborative filtering. A common data structure that helps with keeping track of people and their preferences is a nested dictionary. This dictionary could use a quantitative ranking say on a scale of 1 to 5 to denote the preferences of the people in the selected group.  To find similar people to form a group, we use some form of a similarity score. One way to calculate this score is to plot the items that the people have ranked in common and use them as axes in a chart. Then the people who are close together on the chart can form a group. Several approaches mentioned earlier provide a perspective to solving a problem. This is different from those in that opinions from multiple participants or sensors in a stimuli creation or recognition agens are taken to determine the best set of fleet formation or path changes to recommend.

Collaborative Filtering via Item-based filtering This filtering is like the previous except that it was for user-based approach, and this is for item-based approach. It is significantly faster than the user-based approach but requires the storage for an item similarity table. There are certain filtering cases where divulging which stimuli/sensors go with what formation/path change, is helpful to the fleet manager or participants. At other times, it is preferable to use item/flight path-based similarity. Similarity scores are computed in both cases. All other considerations being same, item-based approach is better for sparse dataset. Both stimuli-based and item-based approach perform similarly for the dense dataset.

Hierarchical clustering Although classification algorithms vary quite a lot, hierarchical algorithm stands out and is called out separately in this category. It creates a dendrogram where the nodes are arranged in a hierarchy.  Specific domain-based ontology in the form of dendrogram can be quite helpful to mining algorithms. 

NLP algorithms Popular NLP algorithms like BERT can be used towards text mining. NLP models come very useful for processing flight path commentary and associated artifacts in the fleet flight management.

Algorithm Implementations:

https://jsfiddle.net/g2snw4da/

https://jsfiddle.net/jmd62ap3/

https://jsfiddle.net/hqs4kxrf/ 

https://jsfiddle.net/xdqyt89a/

#codingexercise https://1drv.ms/w/s!Ashlm-Nw-wnWhPAI9qa_UY0gWf8ZPA?e=PRMYxU


Saturday, June 22, 2024

 

This is a summary of the book titled “Glad We Met – The Art and Science of 1:1 Meetings” written by Steven G. Rogelberg and published by Oxford UP in 2024. The workplace involves plenty of 1:1 meetings and almost half of them do not achieve the desired results. Drawing on extensive research, the author provides a framework on setting up, conducting, and following through on one-on-one meetings. Since career advancement depends on performance evaluation by manager for his or her reports, the author encourages managers to ask the right questions, foster engagement, illuminating each person’s progress. It works both ways for the manager to educate their reports and for their own leadership journey.

These one-on-one meetings do benefit from a framework, argues the author, and those between the manager and direct reports already come with an agenda. Weekly sessions are helpful to the managers and the meeting locations and questions to ask must be planned. Staying positive, sharing mutual priorities, and covering new material, asking for feedback, and saying thank you are all part of it. Regularly conducting these sessions gives more practice to both parties.

One-on-one meetings are crucial for team members and organizations, as they address their priorities, goals, problems, productivity, and employee development. With about a billion business meetings daily, 20% to 50% of these sessions could cost $1.25 billion daily. However, participants often report suboptimal results. Managers can improve their one-on-one meetings to gain a better return on time and money. These meetings strengthen ties within teams and organizations, supplement performance appraisals, and fuel communication between direct reports and managers. To maximize the benefits of one-on-one meetings, create an agenda using the "listings" approach, with the employee covering their list first and the team leader going down. This approach covers immediate work issues and long-range topics, such as career growth and development.

One-on-one meetings are crucial for managerial success, team success, and employee learning and engagement. They promote diversity and inclusion, strengthen relationships, and produce better outcomes. Before meetings, provide context for the topics and ask usual questions. Establish a routine for meetings and explain that they represent the manager's decision to prioritize employees' needs. Stay open-minded and explain shared objectives.

Hold weekly sessions, especially with remote employees, to avoid micromanagement. Choose a schedule that aligns with your needs and preferences, giving your directs some agency. If employees operate from the same office, consider deferring to their preferences.

Plan the location and questions for the meetings, including the office, direct report's office, or outdoor locations. Involve your employee in planning the setting and direct the conversation. The quality of questions asked will determine the quality of dialogue.

Effective one-on-one meetings are essential for team success, fostering better outcomes, strengthening relationships, and promoting diversity and inclusion. Focus on building connections with your employees, their engagement, setting priorities, giving feedback, and fostering career growth and development. Avoid asking personal questions or gossip and maintain a cheerful outlook. Take notes, cover new ground, and ask for feedback. Work through tactical and personal issues, ask for candid feedback, and implement "five key behaviors" to improve your performance. Both parties should feel free to ask for help, and the meeting should end on time. Wrap up the meeting and record important takeaways. Follow up on all commitments made during the meeting.

One-on-one meetings can occur between managers and their direct reports, or with employees meeting individually with their managers' manager or a higher-up executive. Regular one-on-one sessions help ensure your success as a leader, as they provide valuable insights, foster relationships, and help you make better decisions. The Chinese proverb "If you want happiness for an hour, take a nap, go fishing, inherit a fortune, but if you want happiness for a lifetime, help somebody" suggests that fostering relationships and helping others can lead to long-term happiness.

Previous book articles: BookSummary111.docx

Friday, June 21, 2024

 This is a summary of the book titled “The Digital Coaching Revolution – how to support employee development with Coaching Tech” written by Anna Tavis and Woody Woodward. Technology not only augments professional coaching services but also provide a platform for coaches to embrace the potential of data analysis and digital delivery. The authors provide a comprehensive review of digital coaching in corporate environments, sports, healthcare and so on. Digital coaching platforms are maturing towards full autonomy. They help to deliver tailored interventions. Holistic coaching improves individual well-being and productivity. AIs and chatbots have forced coaches to re-imagine their role. Digital coaching has potential to become universities with proper credentials for students. It can help improve diversity related outcomes.

Coaching is a collaborative process that helps individuals and organizations achieve their potential through affective, cognitive, skill-based, and results-based growth. It involves assessing a client's initial needs, agreeing to a schedule, writing a contract, conducting coaching sessions, measuring outcomes, and providing feedback. Although academic research on coaching is still in its infancy, a meta-analysis of scientific studies suggests it improves work satisfaction and well-being more than training. Professional coaching has its roots in Socratic questioning and has evolved with the rise of digital coaching, which uses technology to enhance coaching practices. Digital coaching platforms use machine learning to match coaches and coachees, offer group chats and Google Docs for remote communication, and provide dashboard tools for performance tracking. The multi-billion dollar coaching industry is expected to continue growing and disrupt the $360 billion L&D industry. Digital coaching platforms are maturing toward full autonomy, with four distinct phases: emergent, expansion, maturity, and optimization.

Digital coaches and platforms are increasingly using data to deliver personalized, scalable, and accessible approaches for their clients. This data allows providers to collect empirical evidence demonstrating the value of coaching interventions, such as the ROI Calculator released by CoachHub in 2023. Digital coaching platforms are also experimenting with sentiment analysis and biometrics to guide their services. However, coaches must balance the use of data with relational interactions to drive individual and team performance. Holistic coaching can improve both individuals' well-being and business productivity, as it addresses rising rates of pessimism, anxiety, and depression among workers. Digital health and well-being platforms, such as BetterUp, EZRA, and CoachHub, offer coaching services that cover physical health, mental health, work-life balance, and personal development. As AI and chatbots continue to develop, coaches must re-imagine their role, using AI tools to recommend targeted learning resources and streamline feedback. Integrating AI and chatbots into digital coaching platforms makes personalized learning accessible to a broader audience.

Digital coaching tools can enhance diversity-related outcomes by helping companies achieve their DEIB (Diversity, Equity, Inclusion, and Belonging) goals. These tools can be customized and applied to both upper management and junior employees. BetterUp focuses on the belonging aspect of DEIB, ensuring all employees have access to a coach. As coaching education moves from private companies to universities, it becomes more critical to ensure coaches have appropriate education, training, and credentials. Professional coaching associations offer recognized credentials based on their own standards. Universities are stepping up their coaching education opportunities to advance research in coaching technology and train a new generation of practitioners.


Thursday, June 20, 2024

 This is a continuation of articles on IaC shortcomings and resolutions. As a recap of several articles, this section brings to light what has worked and what hasn’t across a diverse and complex set of deployments. 

1. Cloud is our friend in its ability to provide management free, elastic, performant and highly available resources that can be reached from around the world. Transition from on-premises while requiring migration and modernization concerns are not insurmountable and the final state as cloud native deployments is very appealing in the long run.

2. Resources transition from preview to general acceptance as various cloud services become more mainstream. Since they are developed and made generally available independent of each other, many mature and lower the users concerns in their usage. Leveraging these built-in features over customizations holds us in good stead as it is maintenance free.

3. Anything managed is good including networks, storage and compute if the resource or set of resources to be deployed give that option. Taking the example of an Azure Machine Learning Workspace over an Azure Databricks instance, storage accounts, key vaults, managed virtual network are tightly integrated with the managed option and eschews rediscovery by hands-on configuration.

4. Complex deployments have a significant number of configuration parameters that can quickly get out of hand and require a large test matrix. Capturing them in Infrastructure-as-code with source control is helpful to both repeatable deployments as well as curating best practices.

5. The final state of the resources on their deployment must meet all the criteria so instead of working through the order of various steps and getting lost in the choices, it is better to work backwards from the final state once that has gained deliberation and buy-ins from stakeholders.

6. Zonal redundancy and regional disaster recovery are aspects that must be considered as early as design and implementation. They are not an afterthought and must be judiciously chosen to conserver resources and costs without detriment to the business continuity

7. Most workspaces and resources with public ip connectivity grow feet in the form of private endpoints to gain private plane connectivity through virtual networks so that the connectivity and traffic are secured and there is less attack surface from the ubiquitous internet. Designating subnets for providing addresses to private endpoints from various resources is good practice. Only when private connectivity interferes with usage of resources via workspace notebooks or management views and requires jump hosts or Bastions, then both public and private connectivity might be required. For public connectivity alone, simple firewall rules to enumerate the source and destinations can thwart many attacks that might require otherwise have required costly setup like Defender

8. Cost is always a concern as they have a tendency to balloon with usage and keeping them nanageable requires attention to features that are used and those that aren’t and the process must be repeated at setup as well as an ongoing basis.

9. Just like built-in features, most resources come with diagnostics, troubleshooting and documentation as well as support, so leveraging them avoids lengthy investigations.

10. Naming convention is important to capture the variations possible to resources and do very weill with prefixes and suffixes that have well-known meanings.  

Wednesday, June 19, 2024

 

This is a summary of the book titled “The Digital Coaching Revolution – how to support employee development with Coaching Tech” written by Anna Tavis and Woody Woodward. Technology not only augments professional coaching services but also provide a platform for coaches to embrace the potential of data analysis and digital delivery. The authors provide a comprehensive review of digital coaching in corporate environments, sports, healthcare and so on. Digital coaching platforms are maturing towards full autonomy. They help to deliver tailored interventions. Holistic coaching improves individual well-being and productivity. AIs and chatbots have forced coaches to re-imagine their role. Digital coaching has potential to become universities with proper credentials for students. It can help improve diversity related outcomes.

Coaching is a collaborative process that helps individuals and organizations achieve their potential through affective, cognitive, skill-based, and results-based growth. It involves assessing a client's initial needs, agreeing to a schedule, writing a contract, conducting coaching sessions, measuring outcomes, and providing feedback. Although academic research on coaching is still in its infancy, a meta-analysis of scientific studies suggests it improves work satisfaction and well-being more than training. Professional coaching has its roots in Socratic questioning and has evolved with the rise of digital coaching, which uses technology to enhance coaching practices. Digital coaching platforms use machine learning to match coaches and coachees, offer group chats and Google Docs for remote communication, and provide dashboard tools for performance tracking. The multi-billion dollar coaching industry is expected to continue growing and disrupt the $360 billion L&D industry. Digital coaching platforms are maturing toward full autonomy, with four distinct phases: emergent, expansion, maturity, and optimization.

Digital coaches and platforms are increasingly using data to deliver personalized, scalable, and accessible approaches for their clients. This data allows providers to collect empirical evidence demonstrating the value of coaching interventions, such as the ROI Calculator released by CoachHub in 2023. Digital coaching platforms are also experimenting with sentiment analysis and biometrics to guide their services. However, coaches must balance the use of data with relational interactions to drive individual and team performance. Holistic coaching can improve both individuals' well-being and business productivity, as it addresses rising rates of pessimism, anxiety, and depression among workers. Digital health and well-being platforms, such as BetterUp, EZRA, and CoachHub, offer coaching services that cover physical health, mental health, work-life balance, and personal development. As AI and chatbots continue to develop, coaches must re-imagine their role, using AI tools to recommend targeted learning resources and streamline feedback. Integrating AI and chatbots into digital coaching platforms makes personalized learning accessible to a broader audience.

Digital coaching tools can enhance diversity-related outcomes by helping companies achieve their DEIB (Diversity, Equity, Inclusion, and Belonging) goals. These tools can be customized and applied to both upper management and junior employees. BetterUp focuses on the belonging aspect of DEIB, ensuring all employees have access to a coach. As coaching education moves from private companies to universities, it becomes more critical to ensure coaches have appropriate education, training, and credentials. Professional coaching associations offer recognized credentials based on their own standards. Universities are stepping up their coaching education opportunities to advance research in coaching technology and train a new generation of practitioners.