aster.cloud aster.cloud
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
aster.cloud aster.cloud
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
AI robot
  • Design
  • Platforms

An AI Dilemma: How To Implement Generative AI Tools Safely And Ethically

  • aster_cloud
  • August 23, 2023
  • 4 minute read

Artificial intelligence is being used in all sorts of ways, from chatbots and virtual assistants to self-driving cars, and 97% of business owners believe that ChatGPT will help their business. But with any new technology, there are concerns about safety and ethics – and it’s no different with AI. 

Some business leaders have recently called for a six-month pause on the development of new models more powerful than GPT-4, warning of “profound risks to society and humanity.” With the introduction of the Biden Administration roadmap to promote responsible innovation and focus investment in AI research and development, it’s clear that these risks must be properly mitigated to ensure that safety and the public good remain at the center of all innovation.


Partner with aster.cloud
for your next big idea.
Let us know here.



From our partners:

CITI.IO :: Business. Institutions. Society. Global Political Economy.
CYBERPOGO.COM :: For the Arts, Sciences, and Technology.
DADAHACKS.COM :: Parenting For The Rest Of Us.
ZEDISTA.COM :: Entertainment. Sports. Culture. Escape.
TAKUMAKU.COM :: For The Hearth And Home.
ASTER.CLOUD :: From The Cloud And Beyond.
LIWAIWAI.COM :: Intelligence, Inside and Outside.
GLOBALCLOUDPLATFORMS.COM :: For The World's Computing Needs.
FIREGULAMAN.COM :: For The Fire In The Belly Of The Coder.
ASTERCASTER.COM :: Supra Astra. Beyond The Stars.
BARTDAY.COM :: Prosperity For Everyone.

For companies looking to adopt AI on an enterprise level, there is hesitation on the longevity and safety of new generative AI tools, which poses a necessary question – is all AI bad? What ethical concerns do we need to be aware of? 

As we work to identify answers to these questions, there are tangible steps that can be taken to avoid risking possible ethical dilemmas brought on by data bias. Companies using generative AI must be cognizant of the potential damage that bias can cause and, while large language models (LLMs) are useful, they rely on large sets of data that must be reliable and unbiased. 

Ethical challenges of AI

While ChatGPT and other new AI-generated tools are tempting, and the opportunities seem endless, integrating them into existing products without caution and careful review can reinforce existing stereotypes and discriminatory practices. These generative AI models rely on large sets of data to form their reasonings and explanations, and if those data sets are flawed, biases will be reflected in the responses and work it produces.

Read More  Oracle Introduces Exadata Cloud Service X8M

Data bias used to train these tools can lead to catastrophic results, which is one of the many reasons why an ethical code must be developed and enforced among organizations creating, adopting and integrating these tools into existing products and platforms. For instance, a study by two researchers at the University of Washington found that ChatGPT perpetuates gender stereotypes for occupations across several different spoken languages.

How to harness the benefits of AI – without causing harm

Avoid bias

The most obvious step in creating AI tools that do not suffer bias is to ensure that the data on which the AI is trained does not have bias. This is at odds with models that are trained on the public internet; there is no way to ensure that data pulled randomly from the internet can be free of bias (and, in fact, virtually guarantees that bias will exist). However, when targeting very specific use cases, you can limit the input data and in turn, vet the training data for bias. 

Choose use cases wisely

When deciding whether or not to use AI in a particular use case, think about whether and how AI might be affected by bias. You may find use cases that are much less likely to suffer from bias (for example, in my industry, generating Kubernetes YAML from an English description of a deployment topology) than others (for example, writing a job description for an engineering position, which could accidentally introduce gendered pronouns indicating bias).  

Protect user privacy

We are more aware than ever of how data is being used – think about the number of times a day you get asked about “cookies” on a website. AI and language models represent yet another way that data can be used, and just like with waves of innovation that preceded this one, we need to ensure that we are protecting data privacy. 

Read More  Extended Retention For Custom And Prometheus Metrics In Cloud Monitoring

If you are planning on using user-submitted content as part of your training dataset, you must at least notify your users that their data can be used in that way. And ideally, you would allow users to opt-out of having their data used in training.  

Be transparent about how AI is being used

While ChatGPT can be a useful tool and ease many monotonous and routine tasks, it is crucial to be transparent about AI usage – both internally and externally. A thorough understanding of not only how the work was created but also the data set that was used to inform the work is required to ensure proper fact-checking and bias-reducing actions can be taken. 

Transparency can help police any bias and build trust with users by openly sharing information to inform employees, customers and users’ decisions based on their comfort and encouraging a two-way dialogue about the use of such tools.

Large-Language Models

LLMs are a powerful interactive tool for implementing ChatGPT and other generative AI tools. The best part? They can be trained on private and personalized data sets and models, mitigating many of the ethical issues that may arise in other use cases.

Enterprise companies looking to adopt generative AI can use LLMs to build AI-driven chatbots ranging from technical support portals to blog post generators. However, a disclaimer is needed here – like all code, whether it was written by a colleague, copied from Stack Overflow or generated by an LLM, it must be carefully reviewed and tested before put into production.

Read More  Carrefour Belgium: Driving A Seamless Digital Experience With SAP On Google Cloud

While it is important for the industry as a whole to take steps to ensure ethical models are being enforced, companies themselves must also take on the responsibility of reducing bias when implementing new generative AI tools. As the generative AI landscape continues to evolve and new models are introduced to the market, companies should keep a close eye on not only how these new models can benefit their organizations, but also the broader impacts of implementing these technologies on a larger scale.

By: Dan Ciruli, VP of Product at D2iQ
Originally publish at Cloud Native Computing Foundation

Source: cyberpogo.com


For enquiries, product placements, sponsorships, and collaborations, connect with us at [email protected]. We'd love to hear from you!

Our humans need coffee too! Your support is highly appreciated, thank you!

aster_cloud

Related Topics
  • AI
  • Artificial Intelligence
  • Ethics
  • Generative AI
  • Large Language Models
  • LLM
  • Responsible AI
You May Also Like
View Post
  • Data
  • Engineering
  • Platforms
  • Solutions

How ‘Anything Is Possible’ Automated Data Pipelines With BigQuery And Windsor.ai

  • September 27, 2023
View Post
  • Computing
  • Platforms

Oracle CloudWorld 2023: 6 Key Takeaways From The Big Annual Event

  • September 25, 2023
View Post
  • Platforms

Combining AI With A Trusted Data Approach On IBM Power To Fuel Business Outcomes

  • September 21, 2023
View Post
  • Computing
  • Platforms

Start Your Ubuntu Confidential VM With Intel® TDX On Google Cloud

  • September 20, 2023
Microsoft and Adobe
View Post
  • Platforms
  • Solutions

Microsoft And Adobe Partner To Deliver Cost Savings And Business Benefits

  • September 20, 2023
View Post
  • Platforms
  • Technology

Huawei Releases Data Center 2030, Leading Innovation and Development of New Data Centers

  • September 20, 2023
View Post
  • Multi-Cloud
  • Platforms
  • Public Cloud

Oracle Expands Distributed Cloud Offerings to Help Organizations Innovate Anywhere

  • September 20, 2023
Volkswagen
View Post
  • Multi-Cloud
  • Platforms
  • Technology

Volkswagen Races Toward Next-Gen Automotive Manufacturing Leadership With Google Cloud And T-Systems

  • September 19, 2023

Stay Connected!
LATEST
  • 1
    How ‘Anything Is Possible’ Automated Data Pipelines With BigQuery And Windsor.ai
    • September 27, 2023
  • 2
    Oracle CloudWorld 2023: 6 Key Takeaways From The Big Annual Event
    • September 25, 2023
  • 3
    Nvidia H100 Tensor Core GPUs Come To Oracle Cloud
    • September 24, 2023
  • 4
    Combining AI With A Trusted Data Approach On IBM Power To Fuel Business Outcomes
    • September 21, 2023
  • 5
    Start Your Ubuntu Confidential VM With Intel® TDX On Google Cloud
    • September 20, 2023
  • Microsoft and Adobe 6
    Microsoft And Adobe Partner To Deliver Cost Savings And Business Benefits
    • September 20, 2023
  • Coffee | Laptop | Notebook | Work 7
    First HP Work Relationship Index Shows Majority of People Worldwide Have an Unhealthy Relationship with Work
    • September 20, 2023
  • 8
    Oracle Expands Distributed Cloud Offerings to Help Organizations Innovate Anywhere
    • September 20, 2023
  • 9
    Huawei Connect 2023: Accelerating Intelligence For Shared Success
    • September 20, 2023
  • 10
    Huawei Releases Data Center 2030, Leading Innovation and Development of New Data Centers
    • September 20, 2023
about
Hello World!

We are aster.cloud. We’re created by programmers for programmers.

Our site aims to provide guides, programming tips, reviews, and interesting materials for tech people and those who want to learn in general.

We would like to hear from you.

If you have any feedback, enquiries, or sponsorship request, kindly reach out to us at:

[email protected]
Most Popular
  • Penguin 1
    How To Find And Fix Broken Packages On Linux
    • September 19, 2023
  • Volkswagen 2
    Volkswagen Races Toward Next-Gen Automotive Manufacturing Leadership With Google Cloud And T-Systems
    • September 19, 2023
  • 3
    VMware Scales Multi-Cloud Security With Workforce Identity Federation
    • September 18, 2023
  • Intel Innovation 4
    Intel Innovation 2023
    • September 15, 2023
  • 5
    Microsoft And Oracle Expand Partnership To Deliver Oracle Database Services On Oracle Cloud Infrastructure In Microsoft Azure
    • September 14, 2023
  • /
  • Technology
  • Tools
  • About
  • Contact Us

Input your search keywords and press Enter.