Gen AI Pilot Post Org-Sci report

Do those that said yes to (1) uncertainty; (2) lack of incentives; and (3) leadership has principles impact what kind of actions are they taking if any?

What types of challenges have you or your team faced in regards to using gen AI responsibly?

http://courses.atlas.illinois.edu/spring2016/STAT/STAT200/RProgramming/RegressionFactors.html

action sig oddsratio ci_low(2.5) ci_high(97.5)
The text of the action The significance of the test
(*** = p < .001, NS = Not significant)
Odds ratio. See note below about interpretation. Lower limit of confidence interval for odds ratio Upper limit of confidence interval for odds ratio
Click here for interpretation help

Interpretation: A value in the oddsratio column indicates that participants were who said “Yes” to the text in the tab header were X times more likely to engage in the behavior in the action column.

Example:

Tab: (1) Said “Yes” to “There is a lack of clarity on what ‘responsibility’ looks like”

action sig oddsratio ci_low(2.5) ci_high(97.5)
Want a grilled cheese *** 2.01 1.75 2.75

Participants who said that there is a a lack of clarity on what ‘responsibility’ looks like are 2 times more likely to want a grilled cheese than those who said there is not a lack of clarity on what ‘responsibility’ looks like.

(1) Said “Yes” to “There is a lack of clarity on what ‘responsibility’ looks like”

text odds asterix lower_ci upper_ci
Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility 1.2759 NS 0.7391 2.241
Take ethical / responsible AI trainings 0.9437 NS 0.5528 1.621
Conduct audits of the gen AI tools 1.3616 NS 0.7326 2.654
Conduct fairness / bias testing 2.8708 *** 1.4790 6.041
Consider data privacy implications and take actions to protect data privacy 1.1837 NS 0.6946 2.031
Ask about the data or model to understand potential limitations or issues it has 1.8000 *** 1.0215 3.270
Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors 1.3991 NS 0.6450 3.385
Use explainability methods (e.g., to enable team members to better understand and probe the model) 1.2565 NS 0.7030 2.319
12 Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others) 1.7077 NS 0.9360 3.258
No actions taken (to my knowledge) 1.1335 NS 0.5869 2.315

(2) Said “Yes” to “There is a lack of incentives”

text odds asterix lower_ci upper_ci
Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility 1.9282
1.0869 3.437
Take ethical / responsible AI trainings 2.5414
1.4271 4.605
Conduct audits of the gen AI tools 3.9472
2.1667 7.223
Conduct fairness / bias testing 1.8115
0.9961 3.260
Consider data privacy implications and take actions to protect data privacy 1.5720 NS 0.8888 2.805
Ask about the data or model to understand potential limitations or issues it has 2.6868
1.5095 4.842
Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors 2.3065
1.1051 4.654
Use explainability methods (e.g., to enable team members to better understand and probe the model) 1.7748 NS 0.9764 3.192
12 Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others) 1.5881 NS 0.8705 2.858
No actions taken (to my knowledge) 0.3182
0.1069 0.765

(3) Said “Leadership that has expressed commitment to responsible AI”

text odds asterix lower_ci upper_ci
Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility 2.8907 *** 1.7993 4.6841
Take ethical / responsible AI trainings 3.8870 *** 2.4085 6.3492
Conduct audits of the gen AI tools 3.7284 *** 2.1812 6.4921
Conduct fairness / bias testing 2.3777 *** 1.4445 3.9425
Consider data privacy implications and take actions to protect data privacy 1.8737 *** 1.1818 2.9865
Ask about the data or model to understand potential limitations or issues it has 2.5422 *** 1.5828 4.1128
Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors 2.2222 *** 1.1599 4.3345
Use explainability methods (e.g., to enable team members to better understand and probe the model) 2.3029 *** 1.4017 3.8092
12 Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others) 3.3064 *** 2.0004 5.5314
No actions taken (to my knowledge) 0.2876 *** 0.1398 0.5517

Contingency tables

## $`Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           46           25
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         137           95
## 
## $`Take ethical / responsible AI trainings`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           40           31
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         134           98
## 
## $`Conduct audits of the gen AI tools`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           56           15
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         170           62
## 
## $`Conduct fairness / bias testing`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           60           11
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         152           80
## 
## $`Consider data privacy implications and take actions to protect data privacy`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           40           31
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         121          111
## 
## $`Ask about the data or model to understand potential limitations or issues it has`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           51           20
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         136           96
## 
## $`Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           63            8
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         197           35
## 
## $`Use explainability methods (e.g., to enable team members to better understand and probe the model)`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           52           19
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         159           73
## 
## $`Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others)`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           55           16
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         155           77
## 
## $`No actions taken (to my knowledge)`
## # A tibble: 2 × 3
##   text_challenges_clarity                                                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                                                <int>        <int>
## 1 Said 'No' to 'There is a lack of clarity on what 'responsibility' looks like'           58           13
## 2 Said 'Yes' to 'There is a lack of clarity on what 'responsibility' looks like'         185           47
## 
## $`Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          155           89
## 2 Said 'Yes'  to 'There is a lack of incentives'          28           31
## 
## $`Take ethical / responsible AI trainings`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          151           93
## 2 Said 'Yes'  to 'There is a lack of incentives'          23           36
## 
## $`Conduct audits of the gen AI tools`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          196           48
## 2 Said 'Yes'  to 'There is a lack of incentives'          30           29
## 
## $`Conduct fairness / bias testing`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          177           67
## 2 Said 'Yes'  to 'There is a lack of incentives'          35           24
## 
## $`Consider data privacy implications and take actions to protect data privacy`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          135          109
## 2 Said 'Yes'  to 'There is a lack of incentives'          26           33
## 
## $`Ask about the data or model to understand potential limitations or issues it has`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          162           82
## 2 Said 'Yes'  to 'There is a lack of incentives'          25           34
## 
## $`Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          215           29
## 2 Said 'Yes'  to 'There is a lack of incentives'          45           14
## 
## $`Use explainability methods (e.g., to enable team members to better understand and probe the model)`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          176           68
## 2 Said 'Yes'  to 'There is a lack of incentives'          35           24
## 
## $`Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others)`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          174           70
## 2 Said 'Yes'  to 'There is a lack of incentives'          36           23
## 
## $`No actions taken (to my knowledge)`
## # A tibble: 2 × 3
##   text_orghave_incentives                        `Said 'No'` `Said 'Yes'`
##   <fct>                                                <int>        <int>
## 1 Said 'No'  to 'There is a lack of incentives'          189           55
## 2 Said 'Yes'  to 'There is a lack of incentives'          54            5
## 
## $`Work with colleagues who have job responsibilities are related to supporting AI fairness or responsibility`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          126           52
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          57           68
## 
## $`Take ethical / responsible AI trainings`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          126           52
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          48           77
## 
## $`Conduct audits of the gen AI tools`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          151           27
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          75           50
## 
## $`Conduct fairness / bias testing`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          138           40
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          74           51
## 
## $`Consider data privacy implications and take actions to protect data privacy`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          106           72
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          55           70
## 
## $`Ask about the data or model to understand potential limitations or issues it has`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          126           52
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          61           64
## 
## $`Conduct adversarial testing or red teaming (pretend to be a bad actor to test how a product we develop could be used for harm by bad actors`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          160           18
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'         100           25
## 
## $`Use explainability methods (e.g., to enable team members to better understand and probe the model)`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          137           41
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          74           51
## 
## $`Build transparency approaches (e.g., use documentation that can make dataset and model decisions transparent to others)`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          142           36
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'          68           57
## 
## $`No actions taken (to my knowledge)`
## # A tibble: 2 × 3
##   text_orghave_leadership                                                     `Said 'No'` `Said 'Yes'`
##   <fct>                                                                             <int>        <int>
## 1 Said 'No'  to 'Leadership that has expressed commitment to responsible AI'          130           48
## 2 Said 'Yes'  to 'Leadership that has expressed commitment to responsible AI'         113           12