Conference

NeurIPS Code of Ethics

· · 7 min read · 28 views

NeurIPS Code of Ethics Preamble The Code of Ethics aims to guide the NeurIPS community towards higher standards of ethical conduct as it pertains to elements of research ethics and the broader societal and environmental impact of research submitted to NeurIPS. It outlines conference expectations about the ethical practices that must be adopted by the submitting authors, members of the program and organizing committees. The Code of Ethics complements the NeurIPS Code of Conduct , which focuses on professional conduct and research integrity issues, including plagiarism, fraud and reproducibility concerns. The points described below also inform the NeurIPS Submission Checklist, which outlines more concrete communication requirements. Potential Harms Caused by the Research Process Research involving human subjects or participants: Fair Wages: all human research subjects or participants must receive appropriate compensation. If you make use of crowdsourcing or contract work for a particular task as part of your research project,  you must respect the minimum hourly rate in the region where the work is carried out. Research involving human participants: if the research presented involves direct interactions between the researchers and human participants or between a technical system and human participants, authors are required to follow existing protocols in their institutions (e.g. human subject research accreditation, IRB) and go through the relevant process. In cases when no formal process exists, they can undergo an equivalent informal process (e.g. via their peers or an internal ethics review). Data-related concerns: The points listed below apply to all datasets used for submissions, both for publicly available data and internal datasets. Privacy: Datasets should minimize the exposure of any personally identifiable information, unless informed consent from those individuals is provided to do so. Consent: Any paper that chooses to create a dataset with real data of real people should ask for the explicit consent of participants, or explain why they were unable to do so. Deprecated datasets: Authors should take care to confirm with dataset creators that a dataset is still available for use. Datasets taken down by the original author (ie. deemed obsolete, or otherwise discontinued), should no longer be used, unless it is for the purposes of audit or critical assessment. For some indication of known depreciated datasets, please refer to the NeurIPS list of deprecated datasets. Copyright and Fair Use: While the norms of fair use and copyright in machine learning research are still evolving, authors must respect the terms of datasets that have defined licenses (e.g. CC 4.0, MIT, etc). Representative evaluation practice: When collecting new datasets or making decisions about which datasets to use, authors should assess and communicate the degree to which their datasets are representative of their intended population. Claims of diverse or universal representation should be substantiated by concrete evidence or examples. Societal Impact and Potential Harmful Consequences Authors should transparently communicate the known or anticipated consequences of research: for instance via the paper checklist or a separate section in a submission. The following specific areas are of particular concern: Safety: Contributors should consider whether there are foreseeable situations in which their technology can be used to harm, injure or kill people through its direct application, side effects, or potential misuse. We do not accept research whose primary goal is to increase the lethality of weapons systems. Security: Researchers should consider whether there is a risk that applications could open security vulnerabilities or cause serious accidents when deployed in real world environments. If this is the case, they should take concrete steps to recommend or implement ways to protect against such security risks. Discrimination: Researchers should consider whether the technology they developed can be used to discriminate, exclude, or otherwise negatively impact people, including impacts on the provision of services such as healthcare, education or access to credit. Surveillance: Researchers should consult on local laws or legislation before collecting or analyzing any bulk surveillance data. Surveillance should not be used to predict protected categories, or be used in any way to endanger individual well-being. Deception & Harassment: Researchers should communicate about whether their approach could be used to facilitate deceptive interactions that would cause harm such as theft, fraud, or harassment, and whether it could be used to impersonate public figures and influence political processes, or as a tool to promote hate speech or abuse. Environment: Researchers should consider whether their research is going to negatively impact the environment by, e.g., promoting fossil fuel extraction, increasing societal consumption or producing substantial amounts of greenhouse gasses. Human Rights: We prohibit circulation of any research work that builds upon or facilitates illegal activity, and we strongly discourage any work that could be used to deny people rights to privacy, speech, health, liberty, security, legal personhood, or freedom of conscience or religion. Bias and fairness: Contributors should consider any suspected biases or limitations to the scope of performance of models or the contents of datasets and inspect these to ascertain whether they encode, contain or exacerbate bias against people of a certain gender, race, sexuality, or other protected characteristics. Impact Mitigation Measures We propose some reflection and actions taken to mitigate potential harmful consequences from the research project. Data and model documentation: Researchers should communicate the details of the dataset or the model as part of their submissions via structured templates. Data and model licenses: If releasing data or models, authors should also provide licenses for them. These should include the intended use and limitations of these artifacts, in order to prevent misuse or inappropriate use. Secure and privacy-preserving data storage & distribution : Authors should leverage privacy protocols, encryption and anonymization to reduce the risk of data leakage or theft. Stronger measures should be employed for more sensitive data (e.g., biometric or medical data). Responsible release and publication strategy: Models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, e.g. by requiring that users adhere to a code of conduct to access the model. Authors of papers exposing a security vulnerability in a system should follow the responsible disclosure procedures of the system owners. Allowing access to research artifacts: When releasing research artifacts, it is important to make accessible the information required to understand these artifacts (e.g. the code, execution environment versions, weights, and hyperparameters of systems) to enable external scrutiny and auditing. Disclose essential elements for reproducibility: Any work submitted to NeurIPS should be accompanied by the information sufficient for the reproduction of results described. This can include the code, data, model weights, and/or a description of the computational resources needed to train the proposed model or validate the results. Ensure legal compliance: Ensure adequate awareness of regional legal requirements. This can be done, for instance, by consulting with law school clinics specializing in intellectual property and technology issues. Additional information is required from authors where legal compliance could not be met due to human rights violations (e.g. freedom of expression, the right to work and education, bodily autonomy, etc.). Violations Violations to the Code of Ethics should be reported to hotline@neurips.cc . NeurIPS reserves the right to reject the presentation of scientific works that violate the Code of Ethics. Notice that conference contributors are also obliged to adhere to additional ethical codes or review requirements arising from other stakeholders such as funders and research institutions. Further reading UNDERSTANDING LICENSES Towards Standardization of Data Licenses: The Montreal Data License Behavioral Use Licensing for Responsible AI Choose an open source license MODEL AND DATA DOCUMENTATION TEMPLATES Model Cards for Model Reporting Datasheets for Datasets Using AI Factsheets for AI Governance ML Lifecycle Documentation Practices SOCIETAL IMPACT Safety: Key Concepts in AI Safety: An Overview Security: SoK: Security and Privacy in Machine Learning Discrimination: Bias in algorithms – Artificial intelligence and discrimination ; What about fairness, bias and discrimination? Surveillance: The Human Right to Privacy in the Digital Age Deception & Harassment: Generative Language Models and Automated Influence Operations: Emerging Threats and Potential Mitigations Environment: Quantifying the Carbon Emissions of Machine Learning Human Rights: Technology and Rights | Human Rights Watch Bias and fairness: Fairness and machine learning Dual use problem: Dual use of artificial-intelligence-powered drug discovery Data Enrichment: Responsible Sourcing for Data Enrichment Synthetic Media: PAI’s Responsible Practices for Synthetic Media RELATED ENDEAVORS ACM Code of Ethics ACL Ethics FAQ ICLR Code of Ethics Responsible Conduct of Research Training RELATED RESEARCH COMMUNITIES IEEE SaTML 2023 ACM FAccT Aies Conference FORC 2022 Eaamo Successful Page Load NeurIPS uses cookies for essential functions only. We do not sell your personal information. Our Privacy Policy » Accept

Executive Summary

The NeurIPS Code of Ethics outlines conference expectations for ethical practices in research, focusing on research ethics, societal impact, and environmental considerations. It emphasizes the importance of fair wages, informed consent, data protection, and responsible use of datasets. While the code provides a comprehensive framework for researchers, it may face challenges in enforcing compliance and navigating evolving norms of fair use and copyright in machine learning research. The code's emphasis on transparency and accountability is commendable, but its implementation and effectiveness will depend on the NeurIPS community's commitment to upholding these standards.

Key Points

  • Fair wages must be provided to human research subjects or participants
  • Informed consent is required for research involving human participants or datasets containing personally identifiable information
  • Datasets should minimize personally identifiable information and respect copyright and fair use terms

Merits

Strength in Emphasizing Transparency

The code promotes transparency and accountability in research practices, which is essential for maintaining public trust and ensuring the integrity of research findings.

Comprehensive Framework

The code addresses various aspects of research ethics, including human subjects, data protection, and environmental considerations, providing a comprehensive framework for researchers to follow.

Demerits

Enforcement Challenges

The code may face challenges in enforcing compliance, particularly in cases where researchers may not be aware of or choose to disregard the guidelines.

Evolving Norms of Fair Use and Copyright

The code's emphasis on fair use and copyright may be influenced by evolving norms in machine learning research, which could lead to inconsistencies and challenges in implementation.

Expert Commentary

The NeurIPS Code of Ethics is a critical step towards establishing a culture of transparency and accountability in the research community. However, its effectiveness will depend on the NeurIPS community's commitment to upholding these standards. To address the challenges of enforcement, the code could benefit from more explicit guidelines and consequences for non-compliance. Furthermore, ongoing dialogue and education are necessary to ensure that researchers understand the evolving norms of fair use and copyright in machine learning research.

Recommendations

  • Develop more explicit guidelines and consequences for non-compliance to enhance enforcement of the Code of Ethics.
  • Establish a more robust framework for overseeing research ethics and ensuring accountability in the research community, involving policy makers and institutions.

Sources

Related Articles