Maikel Leon*
Department of Business Technology, Miami Herbert Business School, University of Miami, USA
*Corresponding author:Maikel Leon, Department of Business Technology, Miami Herbert Business School, University of Miami, Miami, Florida, USA
Submission: June 19, 2025;Published: July 10, 2025
ISSN: 2576-8840 Volume 22 Issue 1
Machine Learning (ML) is revolutionizing how scientists predict, synthesize, and validate new materials from super-strong high-entropy alloys to next-generation battery cathodes. Ensuring that this progress remains fair, transparent, and environmentally responsible is therefore central to the future of sustainable materials R&D. This paper examines the normative foundations guiding the accountable creation and deployment of ML models. Because ML rapidly shapes healthcare, finance, and public policy, upholding applications that advance transparency and social benefit is crucial. We articulate ten core principles: accuracy, bias, accessibility, security, privacy, transparency, accountability, human oversight, sustainability, and harm avoidance and demonstrate pathways to incorporate them so that ML systems strengthen social well-being rather than undermine it. Blending theoretical perspectives with real-world illustrations, we propose actionable best practices that foster trust and responsible progress in ML while sketching out emerging research frontiers.
Keywords:Machine Learning; Computational materials science; Materials informatics; Bias; Transparency; Data privacy; Sustainability; Human centric AI
Machine Learning (ML) has reshaped core aspects of modern life, from aiding doctors in diagnostics to helping financial analysts forecast markets using sophisticated, data-driven insights that are nearly unimaginable. In materials science, ML accelerates the discovery of catalysts, polymers, and structural alloys by predicting properties before costly laboratory synthesis. However, alongside these benefits, there are pressing ethical questions regarding fairness, accountability, and the broader societal im- pact. Individuals who design and implement ML systems including researchers, corporate innovators, and policymakers face the challenge of ensuring these technologies serve the public interest rather than amplifying biases or infringing on privacy [1].
One might note that in high-stakes areas (e.g., credit decisions, law enforcement), unexamined ML tools can introduce serious complications [2]. For instance, predictive policing that heavily relies on historical data might inadvertently direct more police scrutiny toward already over-policed neighborhoods, fueling a harmful cycle. This quandary highlights the need for ethical guidelines that strike a balance between technical robustness and social equity.
Transparent ML is essential in engineering and material design because critical decisions like selecting alloys for aircraft wings or optimizing battery chemistries must be traceable from data to outcome. Engineers begin by verifying that input measurements (e.g., composition, temperature, stress) genuinely influence the model, then proceed to interpreting how the algorithm prioritizes competing objectives, such as strength, cost, and sustainability. Clear explanations help detect biases, reveal hidden correlations that violate physical laws, and ensure safety margins are not eroded by overfitting. Regulators and clients also demand evidence that models comply with industry standards, while researchers need interpretable insights to guide new experiments rather than treating the model as a black box. Ultimately, transparency fosters trust, accelerates iteration cycles, and enables the reliable integration of AI with domain knowledge, thereby bridging the gap between theoretical predictions and practical manufacturing constraints.
Meanwhile, industries such as targeted advertising and social media have also undergone swift ML- driven changes, sometimes testing the boundaries of consent and data governance [3]. Recommender systems can magnify controversial or divisive content under the simple goal of maximizing engagement [4,5]. Regulators often struggle to keep pace with this swift evolution, leaving gaps in accountability [6]. Thus, the widening gap between ”what can be built” and ”what ought to be built” underlines the urgent need for concrete ethical frameworks.
In addition, the operational scale of contemporary ML ranging from wearable health monitors to nationwide credit bureaus means that a single line of erroneous code can propagate globally within minutes. A 2024 World Economic Forum survey ranked ”unintended consequences of AI systems” among the top five technological risks to global stability. This reinforces the argument that ethics cannot remain an afterthought bolted onto finished products; instead, ethical reasoning must permeate every phase of the ML lifecycle, from data acquisition and model design to monitoring and retirement.
To address these challenges, we explore how core principles from privacy and bias mitigation to energy sustainability can shape more responsible uses of ML [7,8]. As ML penetrates more aspects of everyday life, the spectrum of decision making tasks handed over to automated systems grows. This trend raises further questions about legal responsibility and public trust. Researchers have begun investigating frameworks for ”explainable AI,” which aim to demystify complex models for non-technical audiences. Yet, even with growing attention, a critical gap persists between theoretically sound ethical guidelines and their consistent application in real world settings. Bridging this gap demands continuous dialogue among tech developers, regulators, and end-users, ensuring that innovations in ML genuinely align with human values, societal norms, and environmental constraints.
The remainder of this paper proceeds as follows. Section 2 distills the ten foundational ethical principles that underpin trustworthy machine-learning practice and pairs each principle with concrete safeguards. Section 3 examines emblematic failure cases from biased hiring algorithms to privacy eroding surveillance and discusses the technical and policy mitigations that followed. Building on these lessons, Section 4 synthesizes cross-sector “best practices” and high- lights the organizational incentives that make them stick. Recognizing the climate cost of large-scale computation, Section 5 turns to sustainability, surveying recent advances in energy efficient model design and low-carbon infrastructure. Section 6 presents the case for truly multidisciplinary project teams, outlining governance structures that keep legal, social, and domain expertise informed. Section 7 offers concluding reflections on the evolving ethical landscape and sketches promising directions for future research and regulation. These sections provide a blueprint for translating high-level principles into practical machine-learning applications and policies.
Key ethical principles in machine learning
Adhering to ethical principles in ML is vital for developing trustworthy and inclusive technologies. These principles reduce the likelihood of harm, such as deepening inequality or compromising personal freedoms, and help build public confidence. Here we briefly outline ten key considerations and provide at least one immediate safeguard for each.
Accuracy: Models must display consistent and reliable performance, especially in sensitive do- mains such as credit underwriting or medical diagnoses. Failing to manage error rates can lead to systemic problems, harming those already at a disadvantage [9]. Recommended practice: run continual calibration against hold-out streams [10].
Bias: Historically imbalanced datasets can perpetuate discrimination unless bias mitigation is made a routine part of system development and implementation. Regular audits and corrections are crucial [11]. Techniques such as counterfactual fairness checks are gaining traction.
Accessibility: ML-based applications should be usable by diverse people, including those from underrepresented communities. Neglecting inclusivity risks widening digital and social divides [12]. Universal-design guidelines and multilingual interfaces are low-hanging fruit.
Security: ML solutions must remain well- protected from data breaches or malicious manipulation as they embed themselves into critical infrastructure. Recent advances in robust aggregation help defend federated models from poisoning.
Privacy: Organizations using personal data for ML must respect privacy rules (such as GDPR) and avoid data misuse [13]. Differential privacy calibration provides mathematical guarantees of bounded leakage.
Transparency: Explaining, at least in broad terms, how ML models arrive at their decisions enhances accountability and user trust [14]. Model cards and decision logs are emerging as standard artifacts.
Accountability: Those who create and deploy ML must take responsibility for the results, ensuring channels for redress in cases of demonstrable harm [15]. Service level agreements can encode ethical KPIs.
Human oversight: Retaining people ”in the loop” for ethically charged decisions reduces the risk of purely automated errors [4]. Escalation protocols that surface uncertainty to human reviewers are proven safeguards.
Sustainability: Because advanced ML can require immense computational power, adopting eco-efficient design from hardware choices to overall carbon footprints is increasingly necessary [9]. Energy dashboards now let teams schedule training during lowcarbon grid hours.
Harm avoidance: A rigorous testing culture is essential to uncover potential negative con- sequences, whether physical, financial, or societal. Red-team exercises emulate worst-case abuse scenarios before release.
We expand on integrating these principles (accuracy, bias, accessibility, security, privacy, transparency, accountability, human oversight, sustainability, and harm avoidance) into real-world ML systems. Along the way, we examine instances where improper uses amplify existing inequalities or endanger personal data. Yet, we also profile examples where well-structured oversight minimized adverse outcomes. Recognizing that ethical ML design requires diverse insights, we emphasize interdisciplinary teams that combine technical, policy, and social expertise. Building on these perspectives, we share lessons and best practices and present a broader framework for continuous evaluation. We also de- vote attention to the ecological implications of large-scale computing, advocating for ”greener” ML design. Finally, we reiterate the importance of ethical vigilance throughout the entire ML lifecycle and highlight promising research and policy innovation directions.
Each of these ten principles can intersect in com- plex ways. For instance, a highly accurate model might deliver lower performance for a specific demographic if bias was inadvertently introduced during data collection. At the same time, improving accessibility can sometimes raise questions about privacy, as broader data sharing might introduce new security risks. Understanding these interdependencies under- scores the importance of addressing ethical concerns holistically, acknowledging trade offs while seeking balanced solutions. For example, a data collection initiative aimed at improving fairness for underrepresented groups must also incorporate robust data- handling policies to safeguard privacy [16].
Navigating ethical trade-offs
Some ethical principles naturally conflict for instance, increasing transparency might require revealing model internals that could compromise user privacy or intellectual property. Similarly, enhancing accessibility through broader data-sharing may inadvertently weaken security controls. Balancing these tensions requires prioritizing context-specific values and applying mitigation strategies such as selective disclosure, layered permissions, or multi-level model abstraction. Ultimately, these trade-offs are not fully resolvable by technology alone; they demand participatory processes where stakeholders co-determine acceptable risk boundaries [9].
One helpful approach is the principle of ”contextual proportionality”: for each ethical conflict, assess the relative severity, reversibility, and affected population. For example, when deciding whether to prioritize explainability or IP protection in materials discovery, a firm might favor transparency if the ML model’s output influences public infrastructure safety but lean toward secrecy for proprietary consumer products. Contextual trade-off matrices and stakeholder scoring tools can formalize this decision-making.
Improper uses and mitigations
Despite growing awareness of AI ethics, lapses still occur in many real-world deployments. Below, we explore settings where flawed ML design or misuse has directly led to problematic outcomes, reminding us why careful regulation and practical governance matter.
Discriminatory algorithms in hiring processes: A significant technology firm discovered that its candidate-screening algorithm systematically over- looked qualified female applicants because it was trained on historical hiring data that was heavily skewed toward men. This situation showed how ignoring demographic biases in legacy data can perpetuate real-world discrimination. It underscores the importance of thorough bias screening and, when needed, rebalancing or ”de-biasing” training sets.
Even after the firm acknowledged the issue, course-correcting the algorithm was not trivial. It involved retraining on a data set carefully adjusted to represent a gender-balanced candidate pool. Additional layers of oversight were instituted, including a manual review of critical junctures in the hiring process. This example highlights that addressing bias in ML requires more than shortterm fixes; it often necessitates revisiting organizational practices surrounding data collection, labeling, and performance benchmarks. A year later, follow-up studies found that departments using the corrected model increased female representation in technical roles by eight percentage points, emphasizing the tangible benefits of algorithmic remediation efforts [13].
Surveillance overreach: Facial recognition technology is increasingly used by law enforcement, yet it raises valid concerns about civil liberties and unequal profiling [5,10]. While intended for public safety, these tools can become intrusive or unfair without well-crafted policies, ongoing audits, and clearly defined boundaries. For example, in one smaller American city, residents discovered that cameras captured facial images in their neighborhoods and referenced them without public discussion, provoking lively debates over privacy rights.
The fallout from such hidden surveillance pro- grams often includes erosion of public trust in local authorities. Once the media exposes unapproved or inadequately supervised implementations, the com- munity can become wary of all subsequent technology solutions, even those with clear benefits. This highlights the importance of incorporating transparency and community engagement into the initial design and procurement phases, rather than treating them as an afterthought. Cities like Amsterdam now mandate ”algorithmic impact assessments” before deployment, demonstrating a governance blueprint that other municipalities can replicate.
Manipulation through social media: Many social media algorithms prioritize content that sparks emotional reactions, boosting engagement metrics but sometimes intensifying misinformation or polarization [4]. Especially during elections, these echo chamber effects can mislead voters and sow dis- trust in democratic institutions. Solving this issue means rethinking how engagement is measured so that factual content is not overshadowed by sensational items that generate more clicks.
Proposed solutions include allowing users to view chronologically ordered feeds or letting them weigh the importance of different content categories themselves. This user-centric approach does not necessarily eliminate misinformation, but it can dilute its rapid spread. Some platforms have experimented with reducing the distribution of problematic content while displaying authoritative sources more prominently, indicating that minor algorithmic modifications can have a significant social impact.
Automated healthcare decisions: Incomplete or skewed medical data can yield flawed diagnostic models that fail to predict outcomes for specific populations accurately [8]. If an ML system prescribes a suboptimal treatment path for patients with less common health profiles, trust in these systems quickly erodes. One might imagine a scenario in which a rural clinic’s ML-driven tool struggles with local demographic data, missing crucial indicators that differ from national averages. Such gaps underscore the importance of continuous model performance verification across diverse demographic segments.
In healthcare contexts, the cost of errors can be much higher than in other applications. A misdiagnosis or delayed diagnosis can directly endanger patients’ lives. Consequently, health organizations often implement rigorous validation procedures, including offline simulations and pilot programs, before fully integrating ML-driven decisions into patient care. This method can mitigate biases that only become apparent when real-world variables, such as local diet patterns or cultural differences in symptom reporting, come into play.
Opaque predictions in materials engineering: In materials science, ML models are increasingly used to predict properties such as fatigue life, corrosion resistance, or thermal conductivity; yet, their internal logic often remains inscrutable. A battery manufacturer, for instance, adopted a graph neural network to recommend new cathode chemistries but later discovered that the model implicitly favored element combinations common in the training set, overlooking rarer dopants that laboratory researchers had shown to improve cycle stability by 15%. Because the algorithm’s feature importances were not exposed, engineers only recognized the oversight after pilot cells underperformed in accelerated aging tests. Similar pitfalls arise in structural alloys, where data scarcity for extreme temperature regimes leads to models that extrapolate poorly, thereby risking unexpected embrittlement. Addressing these challenges requires directly integrating domain constraints such as phase diagrams or mechanistic failure models into ML pipelines and reporting uncertainty bounds alongside predictions. Several aerospace firms now mandate “materials model cards” that document data sources, physics- based assumptions, and validation gaps before an algorithm can influence design decisions, demonstrating how transparency safeguards safety margins and innovation tempo [14].
Notably, several organizations and regulatory bodies are standardizing practices around ML transparency in materials science. ASTM International’s emerging standards for digital data and model reliability in material property prediction (e.g., ASTM E3200) provide a technical scaffold for evaluating algorithmic outputs. Similarly, the U.S. Materials Genome Initiative encourages data-sharing protocols and reproducibility benchmarks to support trustworthy ML integration into design workflows. Model cards in this domain are evolving to include uncertainty quantification (UQ), explainability layers, and cross-validation routines across processing property performance spaces [15]. Technical augmentations, such as embedding physical priors (e.g., phase stability constraints) directly into ML architectures, have also shown promise in avoiding unphysical predictions during extrapolation. These practices are increasingly codified in aerospace and automotive industries, where data-driven design carries safety- critical implications.
Analysis of implications: When ML is misapplied, it can violate civil rights, worsen inequalities, and spark legitimate legal or public relations risks for organizations. The ensuing damage to trust, whether from customers, citizens, or investors, can be longlasting. Additionally, data protection laws, such as GDPR, have steep penalties for privacy breaches. These consequences underscore the importance of ethical design, not only for moral or social reasons, but also for maintaining compliance and a positive reputation.
Addressing these ongoing challenges requires collaboration among governments, industry, community advocates, and academics. Each group provides insights that can influence the overall design, implementation, and evaluation of an ML solution. Both clear transparency around data handling and ac- countability for algorithmic outcomes are fundamental. Additionally, enabling meaningful participation from communities that have been historically side-lined by advanced technologies can help ensure fair outcomes.
In particular, we highlight the following:
a) Regulatory frameworks: Formulating adaptive laws that
respond to rapid ML breakthroughs re- quires policymakers,
engineers, and legal experts to work together [17].
b) Accountability and transparency: Public statements on
how data is sourced, and decisions are made can discourage
misuse [18].
c) Education and public awareness: Nonprofits and
universities play a central role in offering training and resources,
helping everyday citizens identify potential ML abuses [19].
d) Stakeholder involvement: Civil society groups,
policymakers, and local communities must be at the table to
shape decisions that could affect them the most.
One key advantage of involving a broad coalition of stakeholders is the early detection of potential harm. Community representatives can flag issues that data scientists or policymakers may overlook, such as how certain cultural groups interpret data collection or how well user interfaces accommodate people with disabilities. Governments can support funding and regulations that incentivize inclusive design, and professional associations can develop certifications that signal ethical adherence to ML products. Over time, such collective efforts can raise the baseline for re- sponsible or trustworthy ML [20].
Lessons learned and best practices
Examining the evolution of ML deployments whether they soared or stumbled helps us formulate guidelines for future endeavors. This involves distilling lessons from major successes and scrutinizing where things went wrong, as well as what could have been done earlier to prevent issues. Inclusive processes that integrate multiple viewpoints, from designers to community activists, can identify and address ethical pitfalls before they become significant crises.
We also see that context is key. When ML is used in journalism or social media, it faces constraints and oversight demands different from those of health- care or finance. Thus, guidelines should be tailored to each sector’s unique legal and operational factors, leading to more relevant and practical solutions [21]. Across case studies, two meta-lessons recur: (i) documentation is destiny projects with thorough experiment tracking recover faster from errors; (ii) incentives matter teams whose performance evaluations include ethical KPIs exhibit measurably lower incident rates.
Well-known hiring platforms like LinkedIn or Hire Vue have
publicly affirmed their efforts to track and fix bias, showing that
these checks can be woven into real-time hiring workflows:
a) Best practice: Integrate fairness checks into the pipeline
and remove sensitive features that have no legitimate bearing
on job competence. Conduct regular audits to determine if
specific groups are being disproportionately excluded from
opportunities.
b) Impact: This approach broadens the talent pool, enabling
companies to demonstrate a genuine commitment to diversity
and inclusion.
Local legislation, such as New York City’s Public Oversight
of Surveillance Technology (POST) Act, is an example of how
transparency laws reduce the potential harm from unchecked
surveillance:
a) Best practice: Clearly define the scope of facial recognition
tools, maintain public logs detailing where and why they are
deployed, and invite annual audits by an impartial watchdog.
b) Impact: When communities are informed and consent
is sought, skepticism tends to drop, im- proving trust in the
institution’s commitment to privacy.
Major sites like Facebook and Twitter have experimented with
labeling disputed content, which, while imperfect, demonstrates
that platforms can shape more informed public discourse:
a) Best practice: Promote or label credible content and flag
possible misinformation. Additionally, it allows users to adjust
or even override auto- recommendations, fostering a sense of
personal control.
b) Impact: Such measures can diminish the impact
of sensational falsehoods, fostering a healthier online
environment.
Companies such as FICO and Experian have introduced products
(e.g., Experian Boost) that factor in bill payments, illustrating a step
toward more equitable credit-scoring approaches:
a) Best practice: Incorporate data points like rent or utility
payment histories to create a more holistic snapshot of a
borrower’s reliability. Regularly test the model for disparities
across demographic lines.
b) Impact: This method broadens financial inclusion and
bolsters trust in creditworthiness metrics.
IBM’s Watson Health project has emphasized continuous
updates with broader patient data to refine algorithmic accuracy,
illustrating a real-world push for inclusivity in healthcare ML:
a) Best practice: Utilize comprehensive training data that
encompasses diverse demographics and convene expert panels
that combine medical professionals with data scientists. These
panels should regularly review performance metrics.
b) Impact: Inclusive data and consistent oversight enable
ML-driven healthcare platforms to work more effectively for
diverse patient populations, avoiding systematic misdiagnoses.
Global manufacturers such as Airbus, GE Aviation, and leaders in
the U.S. Materials Genome Initiative now require ”materials model
cards” that document data provenance, physics-based constraints,
and uncertainty ranges before any machine-learning pre- diction
can influence alloy or composite selection:
a) Best practice: Embed domain rules (e.g., phase- diagram
limits, failure-mechanism checks) directly into training
pipelines; publish model cards detailing datasets, feature
importance, and error bars; and mandate small-scale lab
validation before scaling results to production.
b) Impact: These measures cut costly redesigns, accelerate
safety certification, and build trust across R&D, quality, and
regulatory teams while also surfacing novel chemistries sooner
for competitive advantage.
Embedding an ethical culture into ML development requires
more than mere compliance. It demands forward-thinking
strategies that address immediate moral and legal issues and
anticipate future complications.
a) Ethical implications: Bringing potential biases and privacy
pitfalls to light early helps maintain public trust and stable
relationships with key stakeholders.
b) Legal implications: A well-documented approach to
fairness and data protection reduces the risk of lawsuits and
sanctions under laws like the GDPR.
c) Rigorous ethical standards: Treating fairness,
transparency, and privacy as must-haves during development
ensures consistent checks and improvements.
d) Stakeholder collaboration: Actively seeking input
from policymakers, civil rights organizations, and impacted
communities enhances legitimacy and acceptance.
e) Continuous education: Team members who stay updated
on evolving ethical challenges can adapt more quickly,
keeping products in line with shifting regulations and societal
expectations [22].
By treating ethics as a foundation for ML solutions (rather than an afterthought), developers and deployers can avoid negative surprises and build momentum for beneficial technologies.
Emerging ethical performance metrics
In response to rising demands for accountability, scholars and practitioners are proposing measurable indicators of ethical ML performance. Table 1 summarizes emerging metrics designed to translate abstract ethical principles into measurable indicators applicable across various ML applications.
Table 1:Examples of emerging ethical performance metrics in ML.

For example, an ML model used in insurance underwriting may report high overall accuracy but exhibit a demographic parity gap of 18% between income groups. By incorporating this metric into quarterly dashboards, the insurer can initiate targeted rebalancing of its datasets. Similarly, tracking the model card completeness score enables project managers to benchmark transparency practices across different teams and promote best-in-class documentation [23].
Sectoral maturity in ethical ML adoption
The maturity of ethical ML adoption varies across industries. While finance and healthcare have well- established compliance cultures, materials engineering is only beginning to adopt robust ML governance. For instance, while financial regulators mandate stress testing and fairness audits, most materials design tools lack standardized documentation or regulatory oversight. This gap highlights the importance of initiatives like the Materials Genome Initiative, which promotes reproducibility and the use of open-access datasets for validation. As materials science increasingly relies on high-throughput simulations and generative surrogate models, establishing sector-specific standards (e.g., digital traceability of alloy recipes or compliance-ready ”materials passports”) will be vital.
Toward eco-conscious ML: Addressing energy sustainability and environmental risks
Although fairness, accountability, and transparency are common focal points in AI ethics, the high environmental cost of large-scale computing also demands attention. Training large neural networks or other computationally heavy models can consume vast amounts of energy [8,9]. Some big tech companies have even considered building or leasing nuclear facilities, which opens up further discussions around waste disposal and community safety [24].
Green AI research prioritizes efficient model de- sign and code to reduce power usage. Approaches such as model pruning or quantization can preserve effectiveness while lowering computation [16]. Simultaneously, many data centers are transitioning to renewable sources (such as solar, wind, and hydro) to reduce their environmental impact.
Nuclear power provides reliable, low-carbon energy during operation, but it also triggers concerns about radioactive waste handling and the risk of accidents [25]. Organizations leaning toward nuclear solutions must seriously address waste management, security protocols, and public acceptance before implementation.
Model distillation and transfer learning innovations allow systems to perform robustly using fewer computational resources. Smaller businesses often benefit from these strategies because they can run top tier ML without expensive data center setups, and the planet benefits from lowered overall energy consumption [11,13].
A lifecycle assessment reminds us of that GPUs accounts for only half of an AI system’s carbon foot- print; manufacturing, cooling, and disposal complete the picture [26]. Robust reporting standards can encourage greener procurement choices.
Examining environmental impact from start to finish (hardware manufacturing to software disposal) helps identify less obvious hotspots of carbon usage [26]. Partnering with hardware vendors can improve transparency about energy consumption and raw material sourcing. Making carbon footprints publicly available also incentivizes the adoption of more efficient infrastructure.
As climate legislation becomes more stringent worldwide, aligning ML with green energy is ethically desirable and strategically savvy [6]. Firms that in- vest early in sustainability stand out to customers and investors seeking a cleaner future.
Bringing sustainability into ML is both an ecological
commitment and a practical business strategy:
a) Transparent energy reporting: Publish metrics on data
center usage, including energy mix and emissions [12].
b) Collaborative green alliances: Partner with environmental
organizations to test more efficient cooling systems or nextgeneration
renewable options [14].
c) Incentivizing Sustainable Architectures: Encourage or
require model optimization to reduce computational intensity
[27].
d) International Standards Alignment: Work toward
international benchmarks, harmonizing local ML goals with
global climate objectives [28].
The importance of multidisciplinary teams in machine learning
Multidisciplinary teams are essential for addressing a wide array of challenges. While data scientists and software developers provide technical expertise, collaboration with legal scholars, ethicists, sociologists, and domain experts offers broader perspectives to help identify issues that purely technical viewpoints might overlook. This section examines how various skill sets contribute to the development of responsible and effective ML projects.
Bridging technical and domain expertise: Many ML projects must incorporate knowledge specific to an industry or application area. When designing a healthcare model, for example, partnering with physicians or clinical researchers can help identify meaningful variables, patient outcomes, and safety thresholds [24].
This collaborative approach:
a) Ensures that important domain factors are not overlooked,
b) Clarifies which metrics are truly relevant for patient care,
c) Aligns modeling strategies with regulatory standards in
healthcare.
Combining expert medical input with data-driven methods makes the resulting models more likely to reflect real-world conditions, ultimately improving patient outcomes and user trust.
Interdisciplinary exchange helps minimize the risk of misinterpretation, where numerical results or confidence scores might be taken at face value without considering social or contextual factors [24]. Data scientists can explain the level of uncertainty in the data, while domain experts highlight nuances that might not be obvious from a purely statistical standpoint. Collaborative discussions also foster healthy skepticism about the model assumptions, reducing the likelihood of overreliance on algorithmic outputs [29].
Ethicists, legal advisors, and social scientists play a critical role
by raising early warnings about potential ethical dilemmas. These
may include:
a) Privacy breaches in handling sensitive data,
b) Based outcomes that disadvantage certain groups,
c) Questions about the fairness of automated decisions [6].
By engaging such experts at the project’s inception, organizations can anticipate how an ML model may affect different stakeholders, thereby addressing problems before they escalate into reputational or legal crises. In sectors like public policy or climate modeling, incorrect interpretations of predictive outputs can lead to severe consequences [23]. Subject-matter experts can help verify model findings by comparing them to historical data, theoretical expectations, or established domain-specific benchmarks. They can also guide sensitivity analyses to determine how small changes in input variables might affect outcomes, enhancing confidence in the model’s reliability and robustness.
Strengthening governance and accountability: Clear governance frameworks are critical for maintaining accountability and prioritizing ethical considerations.
Multidisciplinary teams can define:
a) Who is authorized to audit model decisions and
performance?
b) How often should these audits take place?
c) What remediation steps are needed if models produce
harmful or biased results?
d) How to document the rationale behind key model design
choices?
By establishing these roles and processes, organizations create structures that encourage transparency and continual improvement in their ML initiatives.
Long-term organizational and societal benefits: When ethical thinking and diverse expertise are integrated into a project’s foundation, organizations are more likely to gain the trust of their customers, regulators, and the public.
Over time, this trust can translate into:
a) Competitive advantage through a reputation for social
responsibility,
b) Reduced regulatory risks by proactively adhering to or
even surpassing legal requirements,
c) Greater willingness from stakeholders to engage with and
adopt new technologies.
In this way, a multidisciplinary approach does more than prevent problems; it fosters a culture of responsible innovation that can yield lasting benefits for the organization and the broader community [30].
ML technology is increasingly woven into our daily lives, affecting decisions in sectors as diverse as fi- nance, health, public safety, engineering, and material design. Consequently, the ethical concerns that arise from these deployments are not theoretical. They have real effects on everyday people and, in the case of engineered systems, on the reliability and safety of critical infrastructure [4,10].
Unchecked biases within ML can reinforce in- equality or restrict opportunities, while data breaches and privacy violations can corrode public faith. In materials engineering, for example, a model that systematically overlooks minority-owned suppliers or undervalues sustainable dopants may limit innovation and exacerbate environmental impacts. Addition- ally, ignoring energy efficiency in large-scale computing harms sustainability, particularly when high- fidelity simulations of alloys or composite structures consume vast computational resources [8,9]. Because of these high stakes, robust ethical guidelines incorporating fairness, transparency, accountability, and ecological responsibility are vital for harnessing ML’s potential benefits without inflicting unjust harm.
Regulatory bodies and industry consortia are working to develop frameworks that keep pace with the rapid innovation in ML. Standards organizations, such as ASTM International, are now exploring “materials model cards” that document data provenance and uncertainty bounds, mirroring similar moves in medicine and finance. Still, it remains the responsibility of practitioners and organizations to operationalize ethics in practical ways. We can push the field toward a future that honors human rights, incorporates safety factors, and respects environmental limits by embedding considerations such as bias auditing, broad stakeholder engagement, and energy-aware design into each phase of the ML lifecycle [19].
The convergence of ML with edge devices, quantum hardware, additive manufacturing, and digital twin platforms will introduce novel ethical puzzles for society and the engineering domain. Proactive dialogue among technologists, materials scientists, regulators, and the public will be indispensable for navigating these frontiers responsibly.
A variety of ML ethics frameworks exist, including the EU’s Ethics Guidelines for Trustworthy AI, IEEE’s Ethically Aligned Design, and the U.S. NIST AI Risk Management Framework [22]. While these share common themes (e.g., transparency, fairness, human control), they differ in enforceability and do- main specificity. The EU’s framework emphasizes rights-based approaches suitable for public policy; IEEE’s model is more philosophy-driven and applicable for early-stage design; NIST offers a modular, risk-based approach more aligned with industrial set- tings. Harmonizing insights from these frameworks can yield a more adaptable and cross-sectoral ethical playbook. Table 2 provides a comparative overview of leading ML ethics frameworks, illustrating how their philosophical underpinnings and regulatory orientations influence their suitability for different sectors and deployment stages.
Table 2:Comparison of major ML ethics frameworks.

While the proposed best practices and ethical principles are broadly applicable, challenges arise when generalizing them across regions. Legal frameworks differ e.g., GDPR in Europe vs. sectoral laws in the U.S., which affects how transparency and privacy are implemented. Additionally, resource constraints in lowincome regions may limit the ability to audit or retrain ML models regularly. Cultural values also influence perceptions of fairness and autonomy. Future efforts should tailor ethical ML design to local contexts through participatory design, while pushing for globally aligned minimum standards [31].
To address generalization issues, future research should invest in participatory case studies conducted within diverse legal and cultural ecosystems. For example, assessing the implementation of bias audits in ML systems deployed in Brazil, South Africa, or India can reveal how social norms and data infrastructure shape outcomes. Developing region-specific ethical toolkits that draw on global principles but integrate local realities is essential for inclusive ML innovation.
As ML technologies evolve and permeate new do- mains,
ongoing research and multi-sector collaboration become even
more crucial. Below are several avenues for expanded study and
development:
a) Various ethical codes exist across different verticals
healthcare, autonomous vehicles, and finance but a crossindustry
comparison could illuminate overlapping best
practices and over-looked gaps [6].
b) Long-term studies that measure how faithfully
organizations adhere to ethical principles and the resulting
real-world outcomes (for example, patient health gains or
reduced lending disparities) can help refine existing guidelines
[14].
c) Developing or refining sector-specific indicators like
credit-score fairness indexes or data-center carbon usage
metrics will permit more transparent communication of ethical
performance [15].
d) Novel research might generate accurate energy usage and
emissions predictions for various ML architectures, guiding
policymakers and industry innovators [9].
e) Flexible and inclusive governance structures that gather
insights from government, academia, private industry, and civil
society will be vital to handle ML’s rapid transformations [24].
f) Widening the circle of input from traditionally
underrepresented groups can reduce algorithmic harm and
promote ML solutions more aligned with societal needs [28].
g) Human-AI teaming: Investigating optimal hand- off
protocols between automated agents and hu- man experts
remain an under-researched yet high-impact frontier [32].
h) Engineering and material design: Establishing
standardized “materials model cards” that document data
provenance, physics-based constraints, and uncertainty ranges
can improve traceability and safety in alloy and composite selection
[1].
i) Digital-twin integration: Exploring how ML surrogates
can couple with finite-element and phase-field simulations
to accelerate materials discovery while preserving physical
fidelity offers a promising research trajectory [33].
© 2025 © Maikel Leon. This is an open access article distributed under the terms of the Creative Commons Attribution License , which permits unrestricted use, distribution, and build upon your work non-commercially.
a Creative Commons Attribution 4.0 International License. Based on a work at www.crimsonpublishers.com.
Best viewed in