Improving trust and compliance in language models through accurate source attribution.
― 6 min read
Cutting edge science explained simply
Improving trust and compliance in language models through accurate source attribution.
― 6 min read
This study reveals LLMs' preference for their own output over human-written content.
― 3 min read
Study reveals difficulties for humans and AI in recognizing each other.
― 6 min read
Examining how language influences gender views through biases in AI models.
― 3 min read
CEIPA helps uncover vulnerabilities in large language models to enhance their safety.
― 6 min read
Examining the risks and safety measures for LLM-powered robots.
― 7 min read
Examining the impact of LLMs on social stereotyping and ways to improve outcomes.
― 5 min read
Study reveals surprising risks of sharing sensitive data with chatbots.
― 7 min read
Examining differential privacy and NTK regression to protect user data in AI.
― 6 min read
Exploring differential privacy to protect sensitive information in AI applications.
― 5 min read
Honest Computing emphasizes trust and ethics in data handling.
― 9 min read
A framework enabling data owners to check unauthorized use of their data in ML.
― 7 min read
This method enhances privacy while improving model accuracy in AI.
― 6 min read
Exploring techniques for effectively removing data from AI models.
― 7 min read
Examining the societal and environmental effects of AI foundation models.
― 6 min read
Examining differential privacy in natural language processing for better data protection.
― 7 min read
Examining how AI tools may reinforce biases in resume screening.
― 5 min read
CrossWalk method improves fairness in node embeddings while considering sensitive information.
― 5 min read
Deceptive AI explanations can sway beliefs, complicating the fight against false information.
― 5 min read
This paper examines the effectiveness of current machine unlearning verification methods.
― 6 min read
Examining how T2Is portray caste identities and reinforce stereotypes.
― 6 min read
Study reveals potential leaks of personal identity information by VLMs.
― 6 min read
Examining the role of contestable AI for fairness and accountability.
― 6 min read
A new framework improves alignment of language models with user preferences.
― 7 min read
Research explores how to estimate statistics while respecting user privacy needs.
― 5 min read
A new dataset highlights biases in speech models based on gender and age.
― 7 min read
AI chatbots may create false memories, raising concerns for legal settings.
― 5 min read
New methods enhance face recognition while addressing privacy concerns.
― 6 min read
A method to improve deepfake detection by reducing data biases.
― 5 min read
Analyzing how language models handle social biases in opinions.
― 5 min read
A new method reduces biases in image and text models effectively.
― 6 min read
A new method improves detection of harmful prompts in language models.
― 6 min read
A look at the key differences between Explainable AI and Interpretable AI.
― 7 min read
Exploring certified unlearning and its role in data privacy protection.
― 6 min read
Examining societal bias in Chinese language models and search engines.
― 4 min read
A new approach to improve fairness and accuracy in facial recognition.
― 4 min read
Examining how social identities affect biases in language models.
― 5 min read
Study reveals vulnerabilities in AI models due to backdoor attacks.
― 5 min read
A new dataset aims to improve the understanding of gender bias in AI.
― 5 min read
Examining how liability and insurance can mitigate AI risks.
― 5 min read