Artificial Intelligence at Visier: FAQs

Last Updated: March 7, 2024

Trust
Security
Privacy
Compliance
Documents
Code of Conduct
Trust Assurance FAQs
AI FAQs
Bias Prevention

Table of Contents

Overview

Visier has been leveraging artificial intelligence (AI) technologies for over a decade to power our people analytics platform and deliver predictions throughout our suite of Visier solutions. The following Q&A is intended to helpt customers understand how Visier is currently employing AI technologies.

Visier’s AI Digital Assistant - Vee

Visier uses generative AI in the form of a digital assistant designed to help users navigate the scope of their Visier People data. We call this digital assistant Vee.

Vee lets anyone in an organization ask questions about people, work, and their impact on productivity and receive immediate, accurate, and secure answers from their organization’s data.

  1. How does Vee work?
    Vee interacts with users through a chat interface. Users ask questions in natural language and Vee answers based on the people data that the user’s organization has loaded to Visier People. To learn more about how Vee’s user experience functions, see Vee’s product page here.
  2. What is Vee’s purpose in relation to Visier People?
    Vee serves as a translation layer between a user's natural language question and Visier People’s query language. Rather than a user needing to know the ‘language’ Visier People speaks (i.e. input prompts, filters and metadata), Vee allows a user to simply ask natural, conversational questions, which Vee then translates for Visier People to surface results. Vee thereby improves the user experience of Visier People by allowing an easier interaction with the data already loaded by the customer into Visier People.
  3. What are the steps in Vee’s architecture and data flow when interacting with Visier People?
    • The user asks Vee a question in natural language.
    • Vee requests translation of the natural language questions from a partnering LLM into prompts understood by Visier People.
    • Vee shares those prompts with Visier People
    • Visier People provides an answer to the prompts
    • Those answers are then:
      1. Surfaced back to the user directly through the Vee interface (e.g. in the form of a supporting visual, graphic, or text); or
      2. Returned to the LLM to be translated back into a natural language response to the initial query. This response is then surfaced through the Vee interface to the user as a chat-based response.
  4. Does Visier sanitize the question or prompt submitted by the user?
    Visier does not sanitize the question or prompt submitted by the user.
  5. Does Vee keep user questions after a chat session is complete?
    Vee clears all data associated with a session as soon as the user closes or refreshes their session. Information is not stored beyond the immediate context of the interaction.

LLMs

  1. Why does Vee partner with LLMs?
    Vee uses the natural language capabilities of LLMs to effectively translate a natural language input into prompts that Visier People understands. Vee partners with third-party Large Language Models (LLMs): GPT-3.5 Turbo and GPT-4, both hosted within Microsoft Azure, to assist in this translation. Where the summary feature is selected, Vee will also leverage the LLM to translate a natural language output back to the user in direct response to a user’s chat input, creating a more conversational and contextual experience.
  2. What data is sent to the LLMs?
    The LLMs will process the user’s original input query, data contained in the related chat session, and, when the summary capability is enabled, the output of Visier People in response to a user’s input such to provide a natural language response. This allows Vee to effectively serve as a translation layer between a user's natural language question and Visier People’s query language.
  3. Does the LLM keep this data?
    The LLMs retain this data only for so long as the conversation thread/Vee session remains.

Training

  1. Does Vee train on customer data?
    No, Vee does not train on customer data.
  2. How is Vee trained? What is Blueprint?
    Vee is taught Visier’s proprietary Blueprint, which is pre-built content made available in each Visier People module purchased by a customer; Vee must understand the pre-built content Visier has developed in order to answer user questions related to it. Blueprint is best understood as structured metadata that contextualizes user questions and maps to potential sources of output within a customer’s unique deployment of Visier People. This mapping allows Visier People to generate a relevant response to a specific user input in the context of that customer’s deployment of Visier.

Bias

  1. What about bias?
    Visier products are built with bias mitigation in mind. For example, Visier uses LLMs that leverage bias mitigation techniques such as specific data acquisition and pre-processing, algorithmic fairness and transparency, and monitoring and evaluation. For more information, see Visier’s Bias Prevention and Transparency Statement here.
  2. How does Visier address concerns around bias and discrimination in generative AI?
    Vee’s functionality is scoped to the context of Visier People only. It is taught Visier’s proprietary Blueprint, which has been specifically curated to address employee populations in the same manner as Visier People. Vee’s training data is not generic, and aggregate data is not used to train Vee.

Transparency

  1. How does Vee offer users full transparency?
    Users are able to see how Vee constructed its query to Visier People, and have transparency into the surfaced output. Specifically users can view the prompts, metrics and filters translated by Vee from their natural language input. Vee also allows for human intervention and correction through its interface in the event a user feels Vee has misinterpreted the user’s input. Users can upvote or downvote specific responses, which is subsequently tracked, investigated, and incorporated into future Vee updates.

Risk Management

  1. How does Vee mitigate common risks associated with using generative AI?
    Common generative AI risks are addressed by ensuring Vee is tied to the purpose and function of Visier People as follows:
    • Vee’s scope does not extend beyond Visier People
    • Vee follows the same security policies and procedures applicable to other Visier products. See our Customer Data Safeguards Policy for more information.
    • Vee provides only factual answers available to be surfaced from Visier People. It does not guess or hallucinate.
    • Vee is particularly well-equipped to answer people analytics questions because it has been trained on Visier’s Blueprint which has been specifically curated to address employee populations in the same manner as Visier People (i.e. Vee’s training data is not generic).
  2. What steps have been taken to ensure Vee’s accuracy?
    Vee allows users to upvote or downvote specific responses. This interaction from users is tracked, investigated, and incorporated into future Vee updates, ensuring continuous improvement and maintaining a high overall degree of accuracy and effectiveness.
  3. How does Vee handle uncertainty?
    Vee contains a clarification workflow - as soon as Vee is no longer certain of what is being asked for, it will prompt the user for additional clarification or input.
  4. How does Visier balance experimentation and innovation with risk?
    Experimentation and innovation are balanced with risk through careful planning, assessment, and mitigation strategies:
    • Prototyping and testing new ideas on a smaller scale prior to full implementation allows for early identification of risks and challenges, so that the approach can be refined and improved before widespread deployment.
    • An incremental approach to innovation is used, allowing for continuous feedback, adjustment, and refinement.
    • Ongoing monitoring and evaluation are crucial to track progress, identify emerging risks, and assess the effectiveness of risk mitigation measures.
    • Engaging stakeholders, including users, customers, and internal teams, is essential for understanding concerns, gathering feedback, and building support for innovative initiatives.

Customer Choice

  1. What choices can customers make about Vee?
    Customers can choose to enable Vee. Customers can also elect to only provide the Vee experience to specific users or groups of users within their organization. Customers can choose to enable the Vee summary feature.
  2. Which users have access to Vee?
    Access to Vee is based on customer controlled access permissions and profiles associated with an individual user. These profiles are established as part of your onboarding to Visier People, with Vee adhering to the same permissions for all Visier People access as selected by customer administrators during the onboarding process (or as subsequently modified by customer).

Responsible AI Development

  1. How does Visier ensure Vee is ethically developed?
    Visier builds its AI programs, including Vee, in accordance with the following guiding principles:

    • We respect the evolving guidance of legislative authorities globally, including without limitation the Blueprint for an AI Bill of Rights (US), Responsible use of artificial intelligence (AI) (Canada), and the European Commission's proposed EU Regulatory framework for AI (EU).
    • We believe in responsible, measured development, over innovation at all costs.
    • We ascribe to high levels of transparency, accountability, and explainability.
    • We value continued human oversight with appropriate checks and balances on AI autonomy.
    • We prioritize data security and limit the sharing and persisting of data.
    • We recognize, understand, and address inherent flaws in AI.
    • We are committed to continuing to learn, to evolve, and to reevaluate with each new development.
  2. Does Visier have internal AI expertise or dedicated resources for responsible AI development and implementation?

    Visier has created an internal AI Taskforce to oversee activities in AI governance strategy and ensure cross-collaboration when it comes to AI development initiatives and the use of AI technologies within Visier.

  3. How does Visier ensure responsible AI governance?

    Visier is dedicated to maintaining a safe and responsible approach to our use of AI technologies as we leverage them towards better business outcomes. Our AI governance model is an extension of our core value of ethical and responsible use of data. We adhere to principles intended to (i) foster measured development in consideration of the latest standards; (ii) actively monitor and adapt to compliance requirements based on emerging regulations; and (iii) prioritize transparency and accountability in AI development by working to understand and address risks as they materialize.

Visier’s AI Insights for Executives – Vee Boards

Vee Boards are AI-augmented insight boards that enable users to provide any leader with personalized analysis and guidance. Since they’re powered by Vee, Visier's generative AI assistant, these analyses go far beyond what traditional dashboards can do. More than just a summary, Vee Boards illuminate what’s meaningful in the data quickly, and how leaders should best respond.

Vee Boards is available with a subscription to Vee by Visier®.

  1. How do Vee Boards work?

Vee Boards are AI-augmented analyses. To create one, users can first create an analysis and then turn on AI Explanations which allows Vee to analyse the data and deliver simple, clear and personalized insights to end users.

Visier experts have developed a series of Vee Boards specifically designed to address the top concerns of specific executives and executive groups, such as Talent Density and People Cost. Users can use these prebuilt Vee Boards or build their own.

For more information on Vee Boards please visit our product page.

  1. What are the steps in Vee Boards’ architecture and data flow when interacting with Visier People?

    • An analysis is created and shared using Visier’s regular data and content permissions.
    • Once the analysis is delivered to an end user, they can initiate an AI Explanation experience.
    • Once initiated, this process of explaining the most significant changes through normal statistical processes will run.
    • To generate the AI Explanations, Vee passes the data within the analysis and a prompt to the LLM using the regular Vee infrastructure (as described above). This differs from asking Vee a question, where only the question goes to the LLM.
    • Vee’s generative AI capabilities then return a structured explanation of the changes and results that are most significant. These are presented as a written explanation alongside the data within the analysis.
    • If a user changes their filters and seeks to understand the analysis relative to a different population, the explanation will detect this change and ask the user if they want to re-run the explanation based on the new data.
  2. Are AI Explanations kept after a session is complete?
    All data associated with a session is cleared as soon as the user closes or refreshes their session. Information is not stored within the workflow beyond the immediate context of the interaction. The explanation output can be downloaded to a PowerPoint or PDF output, much like any other written content that can be downloaded from Visier.

LLMs

  1. Why are LLMs used as part of Vee Boards?
    LLMs are used to generate meaningful, concise and personalized explanations of the data and insights contained within an authored analysis. The same third-party LLMs used for Vee are also used for Vee Boards.

  2. What data is sent to the LLMs?
    The LLMs will process the input query (containing the data within the analysis and a prompt) to detect the most significant changes or results that are contained within the analysis. The combination of these inputs and the capabilities of the LLM are what make the creation of personalized explanations at scale possible.
  3. Is the data safe?
    Yes. Visier prioritizes data privacy by ensuring that the LLM only receives the necessary information to generate a natural language response. Additionally, our LLM partners are SOC 2 certified for their GPT and LLM services and have committed to not persisting or using the data sent by Vee for training purposes.
  4. Does the LLM keep this data?
    No. The LLMs only have access to this data for the duration of the request.

Training

  1. Is customer data used to train Vee Boards?
    No, Visier does not use Customer Data to train any generative AI models or technologies within the SaaS Services.

Bias

  1. What about bias?
    Visier products are built with bias mitigation in mind. For example, Visier uses LLMs that leverage bias mitigation techniques such as specific data acquisition and pre-processing, algorithmic fairness and transparency, and monitoring and evaluation. For more information, see Visier’s Bias Prevention and Transparency Statement here
  2. How does Visier address concerns around bias and discrimination in generative AI?
    The functionality of Vee and Vee Boards is scoped to the context of Visier People only. Any insights delivered by Vee Boards are backed by data loaded by a customer into Visier People.

Transparency

  1. How are users of Vee Boards offered full transparency?
    Users can read and review all of the written commentary and supporting data that is presented on a Vee Board. The user has full access to the charts, filters and supporting data to ensure that what is “explained” by the LLM is supported by specific elements of data.

Risk Management

  1. How are common risks associated with using generative AI mitigated with Vee Boards?
    Common generative AI risks are addressed by ensuring Vee Boards is tied to the purpose and function of Visier People as follows:
    • The scope of Vee Boards does not extend beyond Visier People and is directly tied to the security permissions and access of each individual user.
    • Security policies and procedures applicable to other Visier products also apply to Vee Boards. See our Customer Data Safeguards Policy for more information.
    • Vee Boards provides factual information available to be surfaced from Visier People. It does not guess or hallucinate, nor will it return unfavorable or derogatory information.
  2. What steps have been taken to ensure the accuracy of Vee Boards?
    Users are permitted to upvote or downvote specific responses. This interaction from users is tracked, investigated, and incorporated into future updates, ensuring continuous improvement and maintaining a high overall degree of accuracy and effectiveness.
  3. How does Visier balance experimentation and innovation with risk?
    Experimentation and innovation are balanced with risk through careful planning, assessment, and mitigation strategies:
    • Prototyping and testing new ideas on a smaller scale prior to full implementation allows for early identification of risks and challenges, so that the approach can be refined and improved before widespread deployment.
    • An incremental approach to innovation is used, allowing for continuous feedback, adjustment, and refinement.
    • Ongoing monitoring and evaluation are crucial to track progress, identify emerging risks, and assess the effectiveness of risk mitigation measures.
    • Engaging stakeholders, including users, customers, and internal teams, is essential for understanding concerns, gathering feedback, and building support for innovative initiatives.

Customer Choice

  1. Who can use this feature?
    In order to use this feature, Customers must have purchased Vee and opted into AI features. Individual users must have permission capability for Vee.
  2. What choices can customers make about Vee Boards?
    Customers can choose to only provide the Vee Boards experience to specific users or groups of users within their organization, enabling Vee Boards as follows:

    o If the user has Vee capability, and has access to the Vee Board content, they will get the AI Explanation experience.

    o If the user does not have Vee capability, but has access to the Vee Board content, they will get the content without the AI Explanation experience.

  3. Which users have access to Vee Boards?
    Access to Vee Boards is based on customer-controlled access permissions and profiles associated with an individual user. These profiles are established as part of customer onboarding to Visier People, with Vee Boards adhering to the same permissions for all Visier People access as selected by customer administrators during the onboarding process (or as subsequently modified by customer).

Responsible AI Development

  1. How does Visier ensure its generative AI models and technologies are ethically developed?
    Visier builds its AI programs, including Vee and Vee Boards, in accordance with the following guiding principles:
    • We respect the evolving guidance of legislative authorities globally, including without limitation the Blueprint for an AI Bill of Rights (US), Responsible use of artificial intelligence (AI) (Canada), and the European Commission's proposed EU Regulatory framework for AI (EU).
    • We believe in responsible, measured development, over innovation at all costs.
    • We ascribe to high levels of transparency, accountability, and explainability.
    • We value continued human oversight with appropriate checks and balances on AI autonomy.
    • We prioritize data security and limit the sharing and persisting of data.
    • We recognize, understand, and address inherent flaws in AI.
    • We are committed to continuing to learn, to evolve, and to reevaluate with each new development.
  2. Does Visier have internal AI expertise or dedicated resources for responsible AI development and implementation?
    Visier has created an internal AI Taskforce to oversee activities in AI governance strategy and ensure cross-collaboration when it comes to AI development initiatives and the use of AI technologies within Visier.
  3. How does Visier ensure responsible AI governance?
    Visier is dedicated to maintaining a safe and responsible approach to our use of AI technologies as we leverage them towards better business outcomes. Our AI governance model is an extension of our core value of ethical and responsible use of data. We adhere to principles intended to (i) foster measured development in consideration of the latest standards; (ii) actively monitor and adapt to compliance requirements based on emerging regulations; and (iii) prioritize transparency and accountability in AI development by working to understand and address risks as they materialize.

Visier People Machine Learning

Visier leverages machine learning and predictive modeling as part of its Visier People analytics solution to surface results from a customer’s data.

  1. How does Visier People use machine learning?
    Machine learning provides a predictive capability using elements of a customer’s specific people and business data to drive individualized predictive models. These models assist users to identify trends, such as employees who are, for example, likely to resign, to be promoted, and to change jobs.
  2. How does Visier People make predictions?
    Visier People uses a random forest machine learning technique to predict how likely a future event is to occur. The predictive model builds a series of decisions and computes an employee’s expected path along them, averaging historical event likelihoods to generate a result.
  3. How do users interact with Visier People’s predictive capabilities?
    Visier uses Random Forest Models to create predictions and surface the insights to end users. Predictions can be viewed in the aggregate (for example, to see how many employees from a certain role are predicted to resign in the future), as well as at the employee level (for example, to see the specific employees predicted to resign and the top contributing factors driving the prediction). Visier also includes advanced clustering models to identify similar employees across different populations, based on certain attributes or events.
    Visier People offers three different avenues to help users interpret and understand the proper course of action:
    1. Metric guides with guidance - every metric has a useful guide associated with it the user can access from within the application. Guides include not only the definition of the metric, usage and the calculations but also additional suggestions for what the user should focus on. For example, when reviewing resignations you should look at the metric and consider diversity impacts or whether high performers are affected more than poor performers. These suggestions help the user draw the appropriate conclusions and determine where to explore next.
    2. Drivers chart - Visier's unique drivers chart helps understand what factors are most influencing a metric. Looking at absenteeism, for example, the chart will identify the top attributes of employees consuming the most time off and identify which factors are most influential. In this way the user can understand what elements to pursue further to help take the appropriate corrective actions. Is it driven by the new hire onboarding program? Are men and women using time-off equally? Are there age components to it that should be reviewed?
    3. Risk of exit top 5 attributes - Visier's risk of exit predictive algorithm surfaces the attributes causing attrition within your organization and can predict, with a detailed score, who is likely to resign next. The visual surfaces the most compelling attributes for each person at risk so the user can better understand what factors are causing the risk and put in place programs to address it.
  4. Visier People provides a validation metric for each predictive model that lets you measure how close the number of actual exits, promotions, and internal moves were to the predicted values inside the application. You can directly verify using the data of your organization alone and report on whether a higher prediction likelihood resulted in a higher rate of actual events. Predictive success is calculated by taking the predictions for employees at one instant in time and then measuring the actual event rate of these employees in the following validation period of one year. The predictive success measure is defined as the actual event rate of the employees with the highest predicted likelihood divided by the overall event rate in the organization.
    Customers are able to configure these models to use the data attributes they choose, i.e. tenure, direct manager, age, etc. This allows customers to configure the models to be specific to their needs, improve accuracy and reduce biases as needed. Changes to the data attributes being used by the model will regenerate the Random Forest Model and generate new predictions.
  5. What kind of control do customers have over Visier People predictions?
    Customers can control the attributes and metrics that are included as inputs in Visier People. Customers can also choose to turn predictive capabilities on or off in their own individual deployment of Visier.
  6. How much data is needed to generate a predictive model?
    Predictive models are most accurate when 36 months of historical data is available.
  7. How frequently does Visier People undergo model updates?
    Visier People undergoes model updates at least annually. Customer administrators also have the ability to control data attributes included in predictive models generally on a weekly refresh schedule.
  8. How is customer data used in machine learning models retained?
    Visier retains customer data throughout the contracted relationship with each customer. Customer data is deleted no more than 30 days after the agreement is over, in accordance with Visier’s Customer Data Safeguards Policy. While in-term, Customers select their own employee records for inclusion or exclusion. Once an employee record is erased, the machine learning models will no longer leverage that record as of its next scheduled refresh (which could take up to a week depending on how close this occurs to the model refresh every Sunday).
  9. How does Visier test and verify there is no bias in our models?
    Visier People is not intended to filter customer data to detect bias or discrimination. Since customers control the data attributes that are included in their models, it is up to each customer to confirm that the source data is free from bias.