Explainable Artificial Intelligence (XAI)

University - Computer Science - Artificial Intelligence
Match Rate
?
Stage
Ph.D
Duration
48 month
Fund
35000 USD
Fee
125 USD
Deadline
Dec 12, 2024
Start
Sep 01, 2024

About the position

The field of Artificial Intelligence (AI) has witnessed tremendous growth, with AI models achieving remarkable performance in diverse tasks. However, the opaque nature of these models raises concerns about explainability and interpretability. This research project delves into Explainable AI (XAI) techniques, aiming to shed light on how AI models arrive at their decisions. The successful candidate will explore various XAI approaches, including model-agnostic methods and model-specific techniques. Model-agnostic methods, such as SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations), provide explanations for individual predictions without relying on the internal workings of the model. On the other hand, model-specific techniques leverage knowledge about the model's architecture to generate explanations. The candidate will be involved in:

Exam Requirements

TOEFL
Listening
25
Reading
25
Writing
20
Speaking
25
Overall
90

Research Interests

  • • Assist in developing and implementing methods to make AI models more transparent and interpretable.
  • • Collect and analyze data related to XAI models.
  • • Contribute to the preparation of research papers and presentations.
  • • Participate in lab meetings and discussions related to XAI research.
  • • Participate in lab meetings and discussions related to XAI research.
  • • In-depth understanding of machine learning algorithms, particularly deep learning models.
  • • Familiarity with interpretable machine learning frameworks like SHAP and LIME.
  • • Experience with scientific computing libraries like NumPy and SciPy.
  • • Strong mathematical foundation in probability and statistics.
  • Work alongside Professor Sarah Thomas on a research project exploring Explainable Artificial Intelligence (XAI) techniques.

Duties and Responsibilities

  • • Strong foundation in computer science principles
  • • Experience working with AI frameworks like Tenso
  • • Programming skills in Python.
  • • Excellent analytical and problem-solving abiliti
  • • Strong written and verbal communication skills.

Other Requirements, Additional Info

Stanford University offers a vibrant academic community with access to state-of-the-art computing facilities. The successful candidate will have the opportunity to collaborate with leading researchers in the field of AI and participate in academic conferences and workshops.
Share position
Professor

Dr. Sarah Thomas

Professor
Bio

Sarah Thomas is a Professor of Computer Science at Stanford University. She received her PhD in Computer Science from MIT in 2004. Her research interests include machine learning, artificial intelligence, computer vision, robotics, and natural language processing. She has published over 100 papers in top academic conferences and journals. She has also received several awards for his research, including the ACM SIGKDD Innovation Award, the NSF CAREER Award, and the Google Faculty Research Award. Professor Thomas is a passionate teacher and has won numerous teaching awards. She is also an active member of the computer science community and has served on several program committees and editorial boards. She is a Fellow of the Association for the Advancement of Artificial Intelligence (AAAI) and the Institute of Electrical and Electronics Engineers (IEEE).