Live at GTC: Hear From Industry Leaders Using AI to Drive Innovation and Agility

Live at GTC: Hear From Industry Leaders Using AI to Drive Innovation and Agility

Interest in new AI applications reached a fever pitch last year as business leaders began exploring AI pilot programs. This year, they’re focused on strategically implementing these programs to create new value and sharpen their competitive advantage.

GTC, NVIDIA’s conference on AI and accelerated computing, set for March 18-21 at the San Jose Convention Center, will feature leaders across a broad swath of industries discussing how they’re charting the path to AI-driven innovation.

Execs from Bentley Systems, Lowe’s, Siemens and Verizon are among those sharing their companies’ AI journeys.

Don’t miss NVIDIA founder and CEO Jensen Huang’s GTC keynote on Monday, March 18, at 1 p.m. PT.

AI Takes Center Stage in Enterprise Technology Priorities

Nearly three-quarters of C-suite executives plan to increase their company’s tech investments this year, according to a BCG survey of C-suite executives, and 89% rank AI and generative AI among their top three priorities. More than half expect AI to deliver cost savings, primarily through productivity gains, improved customer service and IT efficiencies.

However, challenges to driving value with AI remain, including reskilling workers, prioritizing the right AI use cases and developing a strategy to implement responsible AI.

Join us in person or online to learn how industry leaders are overcoming these challenges to thrive with AI.

Here’s a preview of top industry sessions:

Financial Services

Navigating the Opportunity for Generative AI in Financial Services, featuring speakers from NVIDIA, MasterCard, Capital One and Goldman Sachs.

Enterprise AI in Banking: How One Leader Is Investing in “AI First,” featuring Alexandra V. Mousavizadeh, CEO of Evident, and Chintan Mehta, chief information officer and head of digital technology and innovation at Wells Fargo.

How PayPal Reduced Cloud Costs by up to 70% With Spark RAPIDS, featuring Illay Chen, software engineer at PayPal.

Public Sector

Generative AI Adoption and Operational Challenges in Government, featuring speakers from Microsoft, NVIDIA and the U.S. Army.

How to Apply Generative AI to Improve Cybersecurity, featuring Bartley Richardson, director of cybersecurity engineering at NVIDIA.

Healthcare

Healthcare Is Adopting Generative AI, Becoming One of the Largest Tech Industries, featuring Kimberly Powell, vice president of healthcare and life sciences at NVIDIA.

The Role of Generative AI in Modern Medicine, featuring speakers from ARK Investment Management, NVIDIA, Microsoft and Scripps Research.

How Artificial Intelligence Is Powering the Future of Biomedicine, featuring Priscilla Chan, cofounder and co-CEO of the Chan Zuckerberg Initiative, and Mona Flores, global head of medical AI at NVIDIA.

Retail and Consumer Packaged Goods

Augmented Marketing in Beauty With Generative AI, featuring Asmita Dubey, chief digital and marketing officer at L’Oréal.

AI and the Radical Transformation of Marketing, featuring Stephan Pretorius, chief technology officer at WPP.

How Lowe’s Is Driving Innovation and Agility With AI, featuring Azita Martin, vice president of artificial intelligence for retail and consumer packaged goods at NVIDIA, and Seemantini Godbole, executive vice president and chief digital and information officer at Lowe’s.

Telecommunications

Special Address: Three Ways Artificial Intelligence Is Transforming Telecommunications, featuring Ronnie Vasishta, senior vice president of telecom at NVIDIA.

Generative AI as an Innovative Accelerator in Telcos, featuring Asif Hasan, cofounder of Quantiphi; Lilach Ilan, global head of business development, telco operations at NVIDIA; and Chris Halton, vice president of product strategy and innovation at Verizon.

How Telcos Are Enabling National AI Infrastructure and Platforms, featuring speakers from Indosat, NVIDIA, Singtel and Telconet.

Manufacturing

Accelerating Aerodynamics Analysis at Mercedes-Benz, featuring Liam McManus, technical product manager at Siemens; Erich Jehle-Graf of Mercedes Benz; and Ian Pegler, global business development, computer-aided design at NVIDIA.

Omniverse-Based Fab Digital Twin Platform for Semiconductor Industry, featuring Seokjin Youn, corporate vice president and head of the management information systems team at Samsung Electronics.

Digitalizing Global Manufacturing Supply Chains With Digital Twins, Powered by OpenUSD, featuring Kirk Fleischhaue, senior vice president at Foxconn.

Automotive

Applying AI & LLMs to Transform the Luxury Automotive Experience, featuring Chrissie Kemp, chief data and digital product officer at JLR (Jaguar Land Rover).

Accelerating Automotive Workflows With Large Language Models, featuring Bryan Goodman, director of artificial intelligence at Ford Motor Co.

How LLMs and Generative AI Will Enhance the Way We Experience Self-Driving Cars, featuring Alex Kendall, cofounder and CEO of Wayve.

Robotics 

Robotics and the Role of AI: Past, Present and Future, featuring Marc Raibert, executive director at The AI Institute, and Dieter Fox, senior director of robotics research at NVIDIA.

Breathing Life into Disney’s Robotic Characters With Deep Reinforcement Learning, featuring Mortiz Bächer, associate lab director of robotics at Disney Research.

Media and Entertainment 

Unlocking Creative Potential: The Synergy of AI and Human Creativity, featuring Andrea Gagliano, senior director of data science, AI/ML at Getty Images.

Beyond the Screen: Unraveling the Impact of AI in the Film Industry, featuring Nikola Todorovic, cofounder and CEO at Wonder Dynamics; Chris Jacquemin, head of digital strategy at WME; and Sanja Fidler, vice president of AI research at NVIDIA.

Revolutionizing Fan Engagement: Unleashing the Power of AI in Software-Defined Production, featuring ​​Lewis Smithingham, senior vice president of innovation and creative solutions at Media.Monks.

Energy

Panel: Building a Lower-Carbon Future With HPC and AI in Energy, featuring speakers from NVIDIA, Shell, ExxonMobil, Schlumberger and Petrobas.

The Increasing Complexity of the Electric Grid Demands Edge Computing, featuring Marissa Hummon, chief technology officer at Utilidata.

Browse a curated list of GTC sessions for business leaders of every technical level and area of interest.  

Read More

NLPositionality: Characterizing Design Biases of Datasets and Models

NLPositionality: Characterizing Design Biases of Datasets and Models

TLDR; Design biases in NLP systems, such as performance differences for different populations, often stem from their creator’s positionality, i.e., views and lived experiences shaped by identity and background. Despite the prevalence and risks of design biases, they are hard to quantify because researcher, system, and dataset positionality are often unobserved.

We introduce NLPositionality, a framework for characterizing design biases and quantifying the positionality of NLP datasets and models. We find that datasets and models align predominantly with Western, White, college-educated, and younger populations. Additionally, certain groups such as nonbinary people and non-native English speakers are further marginalized by datasets and models as they rank least in alignment across all tasks.

Figure 1. Carl from the U.S. and Aditya from India both want to use Perspective API, but it works better for Carl than it does for Aditya. This is because toxicity researchers’ positionalities lead them to make design choices that make toxicity datasets, and thus Perspective API, to have positionalities that are Western-centric.

Imagine the following scenario (see Figure 1): Carl, who works for the New York Times, and Aditya, who works for the Times of India, both want to use Perspective API. However, Perspective API fails to label instances containing derogatory terms in Indian contexts as “toxic”, leading it to work better overall for Carl than Aditya. This is because toxicity researchers’ positionalities lead them to make design choices that make toxicity datasets, and thus Perspective API, to have Western-centric positionalities.

In this study, we developed NLPositionality, a framework to quantify the positionalities of datasets and models. Prior work has introduced the concept of model positionality, defining it as “the social and cultural position of a model with regard to the stakeholders with which it interfaces.” We extend this definition to add that datasets also encode positionality, in a similar way as models. Thus, model and dataset positionality results in perspectives embedded within language technologies, making them less inclusive towards certain populations.

In this work, we highlight the importance of considering design biases in NLP. Our findings showcase the usefulness of our framework in quantifying dataset and model positionality. In a discussion of the implications of our results, we consider how positionality may manifest in other NLP tasks.

Figure 2. Overview of the NLPositionality Framework. Collection (steps 1-4): A subset of datasets’ instances are re-annotated via diverse volunteers recruited on LabintheWild. Processing (step 5): We compare the labels collected from LabintheWild with the dataset’s original labels and models’ predictions. Analysis (step 6): We compute the Pearson’s r correlation between the LabintheWild annotations by demographic for the dataset’s original labels and the models’ predictions. We apply the Bonferroni correction to account for multiple hypothesis testing.
Figure 3. Example Annotation. An example instance from the Social Chemistry dataset that was sent to LabintheWild along with the mean of the received annotation scores across various demographics.

NLPositionality: Quantifying Dataset and Model Positionality

Our NLPositionality framework follows a two-step process for characterizing the design biases and positionality of datasets and models. We present an overview of the NLPositionality framework in Figure 2.

First, a subset of data for a task is re-annotated by annotators from around the world to obtain globally representative data in order to quantify positionality. An example of a reannotation is included in Figure 3. We perform reannotation for two tasks: hate speech detection (i.e., harmful speech targeting specific group characteristics) and social acceptability (i.e., how acceptable certain actions are in society). For hate speech detection, we study the DynaHate dataset along with the following models: Perspective API, Rewire API, ToxiGen RoBERTa, and GPT-4 zero shot. For social acceptability, we study the Social Chemistry dataset along with the following models: the Delphi model and GPT-4 zero shot.

Then, the positionality of the dataset or model is computed by calculating the Pearson’s r scores between responses of the dataset or model with the responses of different demographic groups for identical instances. These scores are then compared with one another to determine how models and datasets are biased.

While relying on demographics as a proxy for positionality is limited, we use demographic information for an initial exploration in uncovering design biases in datasets and models.

Table 1: Positionality of NLP datasets and models quantified using Pearson’s r correlation coefficients. # denotes the number of annotations associated with a demographic group. α denotes Krippendorff’s alpha of a demographic group for a task. * denotes statistical significance (p<2.04e−05 after Bonferroni correction). For each dataset or model, we denote the minimum and maximum Pearson’s r value for each demographic category in red and blue respectively.

The demographic groups collected from LabintheWild are represented as rows in the table; the Pearson’s r scores between the demographic groups’ labels and each model and/or dataset are located in the last three and five columns within the social acceptability and toxicity and hate speech sections respectively. For example, in the fifth row and the third column, there is the value 0.76. This indicates Social Chemistry has a Pearson’s r value of 0.76 with English-speaking countries, indicating a stronger correlation with this population.

Experimental Results

Our results are displayed in Table 1. Overall, across all tasks, models, and datasets, we find statistically significant moderate correlations with Western, educated, White, and young populations, indicating that language technologies are WEIRD (Western, Educated, Industrialized, Rich, Democratic) to an extent, though each to varying degrees. Also, certain demographics consistently rank lowest in their alignment with datasets and models across both tasks compared to other demographics of the same type.

Social acceptability. Social Chemistry is most aligned with people who grow up and live in English speaking countries, who have a college education, are White, and are 20-30 years old. Delphi also exhibits a similar pattern, but to a lesser degree. While it strongly aligns with people who grow up and live in English-speaking countries, who have a college education (r=0.66), are White, and are 20-30 years old. We also observe a similar pattern with GPT-4. It has the highest Pearson’s r value for people who grow up and live in English-speaking countries, are college-educated, are White and are between 20-30 years old.

Non-binary people align less to both Social Chemistry, Delphi, and GPT-4 compared to men and women. Black, Latinx, and Native American populations consistently rank least in correlation to education level and ethnicity.

Hate speech detection. Dynahate is highly correlated with people who grow up in English-speaking countries, who have a college education, are White, and are 20-30 years old. Perspective API also tends to align with WEIRD populations, though to a lesser degree than DynaHate. Perspective API exhibits some alignment with people who grow up and live in English-speaking, have a college education, are White, and are 20-30 years old. Rewire API similarly shows this bias. It has a moderate correlation with people who grow up and live in English-speaking countries, have a college education, are White, and are 20-30 years old. A Western bias is also shown in ToxiGen RoBERTa. ToxiGen RoBERTa shows some alignment with people who grow up and live in English-speaking countries, have a college education, are White and are between 20-30 years of age. We also observe similar behavior with GPT-4. The demographics with some of the higher Pearson’s r values in its category are people who grow up and live in English-speaking countries, are college-educated, are White, and are 20-30 years old. It shows stronger alignment with Asian-Americans compared to White people.

Non-binary people align less with Dynahate, PerspectiveAPI, Rewire API, ToxiGen RoBERTa, andGPT-4 compared to other genders. Also, people are Black, Latinx, and NativeAmerican rank least in alignment for education and ethnicity respectively.

What can we do about dataset and model positionality?

Based on these findings, we have recommendations for researchers on how to handle dataset and model positionality:

  1. Keep a record of all design choices made while building datasets and models. This can improve reproducibility and aid others in understanding the rationale behind the decisions, revealing some of the researcher’s positionality.
  2. Report your positionality and the assumptions you make.
  3. Use methods to center the perspectives of communities who are harmed by design biases. This can be done using approaches such as participatory design as well as value-sensitive design.
  4. Make concerted efforts to recruit annotators from diverse backgrounds. Since new design biases could be introduced in this process, we recommend following the practice of documenting the demographics of annotators to record a dataset’s positionality.
  5. Be mindful of different perspectives by sharing datasets with disaggregated annotations and finding modeling techniques that can handle inherent disagreements or distributions, instead of forcing a single answer in the data.

Finally, we argue that the notion of “inclusive NLP” does not mean that all language technologies have to work for everyone. Specialized datasets and models are immensely valuable when the data collection process and other design choices are intentional and made to uplift minority voices or historically underrepresented cultures and languages, such as Masakhane-NER and AfroLM.

To learn more about this work, its methodology, and/or results, please read our paper: https://aclanthology.org/2023.acl-long.505/. This work was done in collaboration with Sebastin Santy and Katharina Reinecke from the University of Washington, Ronan Le Bras from the Allen Institute for AI, and Maarten Sap from Carnegie Mellon University.

Read More