29
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world
9
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

Prominent international brands are unintentionally funding low-quality AI content platforms. Major banks, consumer tech companies, and a Silicon Valley platform are some of the key contributors. Their advertising efforts indirectly fund these platforms, which mainly rely on programmatic advertising revenue.

  • NewsGuard identified hundreds of Fortune 500 companies unknowingly advertising on these sites.
  • The financial support from these companies boosts the financial incentive of low-quality AI content creators.

Emergence of AI Content Farms: AI tools are making it easier to set up and fill websites with massive amounts of content. OpenAI's ChatGPT is a tool used to generate text on a large scale, which has contributed to the rise of these low-quality content farms.

  • The scale of these operations is significant, with some websites generating hundreds of articles a day.
  • The low quality and potential for misinformation does not deter these operations, and the ads from legitimate companies could lend undeserved credibility.

Google's Role: Google and its advertising arm play a crucial role in the viability of the AI spam business model. Over 90% of ads on these low-quality websites were served by Google Ads, which indicates a problem in Google's ad policy enforcement.

Source (Futurism)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

2
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

Amazon Kindle Unlimited’s bestseller list, especially for young adult romance, was recently flooded with AI-created books. Many of these books were nonsense and were apparently being used for click farming.

  • Of the top 100 books, only 19 seemed legitimate, with the others being AI-generated.

Examples of AI-Generated Titles: Among the nonsensical titles were "When the three attacks," "Apricot bar code architecture," "The journey to becoming enlightened is arduous," and "Department of Vinh Du Stands in Front of His Parents’ Tombstone."

  • A book titled "wait you love me," featuring a seagull image on its cover, was 90th on the bestseller list, with two reviews labeling it a "fake AI book."
  • Other peculiar titles included "The God Tu mutters," "Ma La Er snorted scornfully," and "Jessica's Attention."

Continued Presence of AI-Generated Books: Despite their removal from the bestseller list, these AI-generated books are still available for purchase on Amazon.

  • Users can search for and even read samples of these books.
  • For instance, the book "Apricot bar code architecture" starts with a nonsensical sentence about black lace pajamas.
  • As of the time of the report, an Amazon spokesperson had not responded to requests for comment.

Source (Vice)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

1
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

AI is increasingly helping doctors not only in technical tasks but also in communicating with patients empathetically. AI chatbots are proving to be useful in offering quality responses and showcasing empathy superior to human doctors in some cases.

AI in Human Aspects of Medical Care:

  • AI tools like ChatGPT are being used to communicate with patients more empathetically.
  • For instance, in an encounter with a patient's family, ER physician Dr. Josh Tamayo-Sarver used ChatGPT-4 to explain a complex medical situation in simpler, more compassionate terms.
  • The tool generated a thoughtful, empathetic response, which helped comfort the patient's family and save the doctor's time.

AI in Providing Compassionate Counsel:

  • Dr. Gregory Moore used ChatGPT to counsel a friend with advanced cancer, including breaking bad news and dealing with her emotional struggles.
  • Rheumatologist Dr. Richard Stern uses ChatGPT in his clinical practice to write kind responses to patient emails, provide compassionate replies to patient queries, and manage paperwork.

Reasons Behind the Success of AI in Displaying Empathy:

  • AI tools, unlike humans, are not affected by work stress, insufficient coaching, or the need to maintain work-life balance.
  • AI tools like ChatGPT have proven effective in generating text responses that make patients feel they are receiving empathy and compassion.

Source (Forbes)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

3
submitted 2 years ago* (last edited 2 years ago) by Rifal@lemmy.world to c/chatgpt@lemmy.world

Many employees would be open to AI replacing their bosses due to dissatisfaction with their leadership, according to a survey. This openness stems largely from the belief that AI could offer fair and unbiased management.

Initial Discoveries: The survey questioned a thousand workers and found nearly one-fifth of them would like a robotic replacement for their current boss. This sentiment arises from complaints against bosses, notably a perceived lack of appreciation, empathy, and favoritism.

  • The primary complaints include bosses' lack of appreciation and empathy.
  • Another significant issue is favoritism, with some workers feeling that they are treated unfairly compared to others.

Dissatisfaction with Current Leadership: Participants also expressed dissatisfaction with their leaders' management styles. Key grievances included unclear expectations, disorganization, and micromanagement.

  • A significant number of respondents pointed to their bosses' unclear expectations.
  • Others expressed frustration with their bosses' disorganization.
  • Micromanagement also emerged as a common complaint.

Beliefs About AI Leadership: Many of the surveyed workers believed an AI would outperform their current boss. About a third believed AI will soon dominate the workplace.

  • Some participants felt that an AI would be more competent than their current boss.
  • A good number of the participants also believe that AI will soon be commonplace in workplaces.

Industry Variations: The acceptance of AI leadership varied across industries. The most acceptance came from the Arts and Culture sector, followed by HR, Manufacturing and Utilities, Finance, and Healthcare.

  • Arts and Culture workers were the most open to AI leadership.
  • Workers in the HR, Manufacturing and Utilities, Finance, and Healthcare sectors also showed significant acceptance.

Gender and Generational Differences: The survey noted minor gender differences and more pronounced generational differences. Younger respondents were more open to AI leadership than older ones.

  • A slightly higher percentage of males were open to AI bosses compared to females.
  • Younger workers (18-24) showed a significantly higher acceptance for AI bosses compared to older ones (55 and above).

Perceived Advantages of AI Leadership: The main reasons for preferring AI leadership were the elimination of favoritism, discrimination, and making unbiased decisions. Some participants also felt that AI could help reduce workplace drama.

  • The elimination of favoritism and discrimination were cited as key advantages.
  • Participants also appreciated the perceived ability of AI to make unbiased decisions.
  • Some respondents believed AI could help reduce workplace drama.

Source (Techradar)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

9
submitted 2 years ago by Rifal@lemmy.world to c/technology@lemmy.ml

OpenAI's lobbying efforts in the European Union are centered around modifying proposed AI regulations that could impact its operations. The tech firm is notably pushing for a weakening of regulations which currently classify certain AI systems, such as OpenAI's GPT-3, as "high risk."

Altman's Stance on AI Regulation:

OpenAI CEO Sam Altman has been very vocal about the need for AI regulation. However, he is advocating for a specific kind of regulation - those favoring OpenAI and its operations.

OpenAI's White Paper:

OpenAI's lobbying efforts in the EU are revealed in a document titled "OpenAI's White Paper on the European Union's Artificial Intelligence Act." The document focuses on attempting to change certain classifications in the proposed AI Act that classify certain AI systems as "high risk."

"High Risk" AI Systems:

The European Commission's "high risk" classification includes systems that could potentially harm health, safety, fundamental rights, or the environment. The Act would require legal human oversight and transparency for such systems. OpenAI, however, argues that its systems such as GPT-3 are not "high risk," but could be used in high-risk use cases. It advocates that regulation should target companies using AI models, not those providing them.

Alignment with Other Tech Giants:

OpenAI's position mirrors that of other tech giants like Microsoft and Google. These companies also lobbied for a weakening of the EU's AI Act regulations.

Outcome of Lobbying Efforts:

The lobbying efforts were successful, as the sections that OpenAI opposed were removed from the final version of the AI Act. This success may explain why Altman reversed a previous threat to pull OpenAI out of the EU over the AI Act.

Source (Mashable)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

2
submitted 2 years ago by Rifal@lemmy.world to c/technology@lemmy.ml

A Wharton professor believes that businesses should motivate their employees to share their individual AI-enhanced productivity hacks, despite the prevalent practice of hiding these tactics due to corporate restrictions.

Worker's Use of AI and Secrecy:

  • Employees are increasingly using AI tools, such as OpenAI's ChatGPT, to boost their personal productivity and manage multiple jobs.
  • However, due to strict corporate rules against AI use, these employees often keep their AI usage secret.

Issues with Corporate Restrictions:

  • Companies tend to ban AI tools because of privacy and legal worries.
  • These restrictions result in workers being reluctant to share their AI-driven productivity improvements, fearing potential penalties.
  • Despite the bans, employees often find ways to circumvent these rules, like using their personal devices to access AI tools.

Proposed Incentives for Disclosure:

  • The Wharton professor suggests that companies should incentivize employees to disclose their uses of AI.
  • Proposed incentives could include shorter workdays, making the trade-off beneficial for both employees and the organization.

Anticipated Impact of AI:

  • Generative AI is projected to significantly transform the labor market, particularly affecting white-collar and college-educated workers.
  • As per a Goldman Sachs analysis, this technology could potentially affect 300 million full-time jobs and significantly boost global labor productivity.

Source (Business Insider)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

3
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

A Wharton professor believes that businesses should motivate their employees to share their individual AI-enhanced productivity hacks, despite the prevalent practice of hiding these tactics due to corporate restrictions.

Worker's Use of AI and Secrecy:

  • Employees are increasingly using AI tools, such as OpenAI's ChatGPT, to boost their personal productivity and manage multiple jobs.
  • However, due to strict corporate rules against AI use, these employees often keep their AI usage secret.

Issues with Corporate Restrictions:

  • Companies tend to ban AI tools because of privacy and legal worries.
  • These restrictions result in workers being reluctant to share their AI-driven productivity improvements, fearing potential penalties.
  • Despite the bans, employees often find ways to circumvent these rules, like using their personal devices to access AI tools.

Proposed Incentives for Disclosure:

  • The Wharton professor suggests that companies should incentivize employees to disclose their uses of AI.
  • Proposed incentives could include shorter workdays, making the trade-off beneficial for both employees and the organization.

Anticipated Impact of AI:

  • Generative AI is projected to significantly transform the labor market, particularly affecting white-collar and college-educated workers.
  • As per a Goldman Sachs analysis, this technology could potentially affect 300 million full-time jobs and significantly boost global labor productivity.

Source (Business Insider)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 50+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

2
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

Microsoft's investment in AI, notably through OpenAI's ChatGPT, has led to predictions of a $10 billion revenue increase in the coming years, driving shares to an all-time high.

Record High Stocks and AI Growth: Microsoft shares have reached a record high due to its growth prospects in artificial intelligence.

  • The company's stocks rose 3.2%, closing at $348.10, largely fueled by AI, particularly with Microsoft's investment in OpenAI.

Microsoft and OpenAI Partnership: The partnership with OpenAI is pivotal to Microsoft's AI success.

  • Microsoft heavily invested in OpenAI and provides underlying computing power for its projects.
  • Microsoft has an exclusive license on OpenAI’s models, like the GPT-4 language model.
  • The integration of OpenAI tools into Microsoft's services like Bing and Windows boosts their offerings.

Financial Prospects and Investor Interest: Microsoft's AI ventures have raised investor interest and revenue expectations.

  • Microsoft’s finance chief Amy Hood forecasts Azure cloud's growth at 26-27% YoY, with 1% coming from AI services.
  • Hood mentioned that “the next generation AI business will be the fastest-growing $10 billion business in our history.”
  • This prospect has lifted the interest of investors who are keen on the company's earnings and revenue.

Future Predictions and Market Response: Microsoft’s recent successes have led to optimistic market predictions.

  • JPMorgan analysts raised their price target from $315 to $350.
  • Despite challenges like cloud growth and a shrinking PC market, Microsoft's AI investments, such as OpenAI/ChatGPT, signal long-term success.
  • Microsoft’s shares have recovered from their 2022 losses, indicating a positive market response.

AI and Market Trends: AI has emerged as a leading factor in tech market trends.

  • AI has been a trending topic after the release of the ChatGPT chatbot.
  • Tech companies have adopted AI technologies in their products to drive cost savings amid recession concerns.
  • The widespread adoption of AI, backed by companies like Microsoft, has sparked optimism in the tech sector, reviving bullish market sentiments.

Source (CNBC)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 40+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

1
submitted 2 years ago by Rifal@lemmy.world to c/chatgpt@lemmy.world

The tech industry is experiencing significant job cuts, driving demand for HR professionals who can manage termination processes well. ChatGPT is being increasingly used to aid these professionals in their difficult tasks.

Layoffs in Tech Industry: Major tech corporations have recently cut jobs, leading to increased need for HR professionals. These individuals are sought after for their ability to handle sensitive termination processes with tact.

  • Tech giants like Google, Meta, and Microsoft have laid off tens of thousands of workers in the past half year.
  • The layoffs have sparked a demand for Human Resources professionals, particularly those skilled in handling termination processes.

HR Professionals and AI Tools: To better manage these difficult termination conversations, HR professionals are leveraging AI tools.

  • Many HR professionals in the tech industry are turning to AI to assist them with challenging tasks.
  • Over 50% of HR professionals in the tech industry have used AI like ChatGPT for training, surveys, performance reviews, recruiting, employee relations, etc.
  • More than 10% of these HR professionals have used ChatGPT to craft employee terminations.

Survey Findings and AI Usage: A recent survey studied the experiences of tech HR professionals and tech employees with HR in the industry, revealing extensive AI use.

  • The survey involved 213 tech HR professionals and 792 tech employees.
  • The findings suggest an increasing reliance on AI tools, especially ChatGPT, for diverse HR tasks, including crafting terminations.

Implications of AI Use: Despite its convenience, using AI in sensitive situations like employee termination can lead to potential trust issues.

  • AI chatbots, like ChatGPT, allow users to emotionally detach from difficult situations such as job termination.
  • However, using AI for these purposes could result in decreased trust between employees and HR professionals.

Previous Use of ChatGPT: ChatGPT has been used for a variety of sensitive matters in the past, such as writing wedding vows and eulogies.

  • ChatGPT's use is not limited to HR-related tasks; it has previously been used to write wedding vows and eulogies.
  • This illustrates the versatility of AI tools in dealing with emotionally charged situations.

Source (ZDnet)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 40+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

1
submitted 2 years ago* (last edited 2 years ago) by Rifal@lemmy.world to c/chatgpt@lemmy.world

Many workers on platforms like Amazon Mechanical Turk are using AI language models like GPT-3 to perform their tasks. This use of AI-produced data for tasks that eventually feed machine learning models can lead to concerns like reduced output quality and increased bias.

Human Labor & AI Models:

  • AI systems are largely dependent on human labor, with many corporations using platforms like Amazon Mechanical Turk.
  • Workers on these platforms perform tasks such as data labeling and annotation, transcribing, and describing situations.
  • This data is used to train AI models, allowing them to perform similar tasks on a larger scale.

Experiment by EPFL Researchers:

  • Researchers at the École polytechnique fédérale de Lausanne (EPFL) in Switzerland conducted an experiment involving workers on Amazon Mechanical Turk.
  • The workers were tasked with summarizing abstracts of medical research papers.
  • It was found that a significant portion of the completed work appeared to be generated by AI models, possibly to increase efficiency and income.

Use of AI Detected Through Specific Methodology:

  • The research team developed a methodology to detect if the work was human-generated or AI-generated.
  • They created a classifier and used keystroke data to detect whether workers copied and pasted text from AI systems.
  • The researchers were able to validate their results by cross-checking with the collected keystroke data.

The Drawbacks and Future of Using AI in Crowdsourced Work:

  • Training AI models on data generated by other AI could result in a decrease in quality, more bias, and potential inaccuracies.
  • Responses generated by AI systems are seen as bland and lacking the complexity and creativity of human-generated responses.
  • Researchers suggest that as AI improves, the nature of crowdsourced work may change with the potential of AI replacing some workers.
  • The possibility of collaboration between humans and AI models in generating responses is also suggested.

The Importance of Human Data:

  • Human data is deemed as the gold standard as it is representative of humans, whom AI serves.
  • The researchers emphasize that what they often aim to study from crowdsourced data are the imperfections of human responses.
  • This could imply that measures might be implemented in future to prevent AI usage in such platforms and ensure human data acquisition.

Source (The Register)

PS: I run a ML-powered news aggregator that summarizes with an AI the best tech news from 40+ media (TheVerge, TechCrunch…). If you liked this analysis, you’ll love the content you’ll receive from this tool!

view more: next ›

Rifal

joined 2 years ago