AI4LIFE-GROUP
The AI4LIFE group at Harvard is led by Hima Lakkaraju. We study interpretability, fairness, privacy, and reliability of AI and ML models.
Popular repositories Loading
-
LLM_Explainer
LLM_Explainer PublicCode for paper: Are Large Language Models Post Hoc Explainers?
-
med-safety-bench
med-safety-bench PublicMedSafetyBench: Evaluating and Improving the Medical Safety of LLMs [NeurIPS 2024]
Repositories
Showing 10 of 28 repositories
- med-safety-bench Public
MedSafetyBench: Evaluating and Improving the Medical Safety of LLMs [NeurIPS 2024]
AI4LIFE-GROUP/med-safety-bench’s past year of commit activity - sae_robustness Public
AI4LIFE-GROUP/sae_robustness’s past year of commit activity - RLHF_Trust Public
AI4LIFE-GROUP/RLHF_Trust’s past year of commit activity - interp_interv Public
Code for "Towards Unifying Interpretability and Control: Evaluation via Intervention"
AI4LIFE-GROUP/interp_interv’s past year of commit activity - average-case-robustness Public
Characterizing Data Point Vulnerability via Average-Case Robustness, UAI 2024
AI4LIFE-GROUP/average-case-robustness’s past year of commit activity
People
This organization has no public members. You must be a member to see who’s a part of this organization.
Top languages
Loading…
Most used topics
Loading…