學門類別
政大
哈佛
- General Management
- Marketing
- Entrepreneurship
- International Business
- Accounting
- Finance
- Operations Management
- Strategy
- Human Resource Management
- Social Enterprise
- Business Ethics
- Organizational Behavior
- Information Technology
- Negotiation
- Business & Government Relations
- Service Management
- Sales
- Economics
- Teaching & the Case Method
最新個案
- Leadership Imperatives in an AI World
- Vodafone Idea Merger - Unpacking IS Integration Strategies
- Predicting the Future Impacts of AI: McLuhan’s Tetrad Framework
- Snapchat’s Dilemma: Growth or Financial Sustainability
- V21 Landmarks Pvt. Ltd: Scaling Newer Heights in Real Estate Entrepreneurship
- Did I Just Cross the Line and Harass a Colleague?
- Winsol: An Opportunity For Solar Expansion
- Porsche Drive (B): Vehicle Subscription Strategy
- Porsche Drive (A) and (B): Student Spreadsheet
- TNT Assignment: Financial Ratio Code Cracker
-
Nudge Users to Catch Generative AI Errors
Having a human in the loop is critical to mitigating the risks of generative AI errors and biases. But humans are also vulnerable to errors and biases and may trust artificial intelligence either too much or not enough. Findings from a field experiment by MIT and Accenture suggest that targeted friction in the form of labels that flag potential errors and omissions can direct users' attention to content that should be given closer inspection without sacrificing efficiency.