
Neha Bhargava
Neha is the head of Know Your Customer (KYC) for Goldman Sachs Singapore and the deputy head of KYC in the Asia Pacific. Neha has over 16 years of experience at Goldman Sachs with more than 10 years in the financial crime compliance and anti-money laundering space.
Neha is responsible for client identification, enhanced due diligence, strategic enhancements, and regulatory audits, and provides advice for KYC-related issues in the region. Before joining the Compliance team, Neha was the Head of the Prime Brokerage cash management operations team.
Neha holds a Bachelor of Commerce degree from Bangalore University.
Getting one or two AI models into production is very different to running an entire enterprise or product on AI, and as AI is scaled, problems can (and often do) scale too.
- Standardizing how you build and operationalize models
- Focusing teams where they’re strongest
- Introducing MLOps and establishing best practices and tools to facilitate rapid, safe, and efficient development and operationalization of AI
Failure to adequately explain model development, working and outcome inherently invites both regulatory and customer scrutiny, especially when things go wrong.
- The extent to which customers need to know how and why a particular outcome has been reached
- Do you need to understand black box models and if so, why?
- Where is explainability a luxury and where is it absolute necessity
- Lessons learned from failures and how explainability could have helped
Significant amounts of data are required for AI in both training and operation, and it is vital to ensure both data quality and compliance. In an age where individuals are being granted ever more control over their data, there are new and emerging challenges post GDPR. This panel will explore:
- How to ensure both high quality and compliant data
- Potential solutions for using highly sensitive data
Opportunities for data collaboration and partnership
Even without intentionally prejudice data or development practices, AI can produce inequitable results. How can organizations ensure they are mitigating bias at all levels and reducing the risk of reputational, societal and regulatory harm?
AI has the potential to unlock significant long-term value but trust needs to be felt both by the company deploying the AI and the customers that will experience it and its outcomes. This discussion will explore the success and failures organizations have faced with:
- Effectively building trust and conveying the net benefits for all parties
- Human circuit breaks as a safety mechanism
- Leveraging trust to create more desirable business outcomes
- The definition of trust in the context of AI – is it in the development process, the outcome or both?