3. Results
In this study, we were interested in GPT-3´s view on benefits and risks of AI contribution in regard to the social dimension of the SDGs, covering nine goals, namely SDG 1 to 7, SDG 11 and SDG 16, including their 58 outcome targets [
1,
7]. GPT-3 provided the following outputs for each of the subheadings, stratified in the specific goal, e.g., for goal 1 and the according outcome targets, i.e., 1.1 to 1.5. Notably, typos and blanks were corrected by the human authors to increase readability, but were summarized in the descriptive analysis section.
3.1. Goal 1. End poverty in all its forms everywhere
Target 1.1: Eradicate Extreme Poverty: AI can be a valuable tool to help reduce poverty, by automatically identifying and analyzing large amounts of data to identify areas where poverty is most prevalent and where the greatest need for aid exists. Additionally, AI can be used to monitor economic trends in developing nations, allowing for proactive responses and early interventions when needed. However, there are risks associated with this technology, including the potential for misuse or misinterpretation of data and algorithms which could lead to inappropriate or unfair decisions.
Target 1.2: Reduce Poverty Proportion: The application of AI in poverty reduction can provide more accurate predictions on how best to allocate resources and create more efficient policies that target poverty reduction. This will enable governments and organizations to better understand the needs of those living in poverty and develop effective strategies to address them. However, if not regulated properly, AI could also create biases or be used as a tool for discrimination against disadvantaged groups.
Target 1.3: Social Protection Systems: By using AI-driven technologies such as machine learning and natural language processing (NLP), governments can quickly process large amounts of social protection data and develop more comprehensive policies tailored specifically to vulnerable populations’ needs. Furthermore, such technologies can provide faster access to benefits for those most in need while reducing bureaucracy costs associated with traditional systems. However, these technologies may also lead to privacy concerns if not used appropriately or securely stored by appropriate regulations.
Target 1.4: Equal Rights Access: AI technology has the potential to increase access to financial services for people living in rural areas or other disadvantaged communities who have limited access due to financial constraints or lack of infrastructure. Additionally, through machine learning algorithms, companies can identify the best products for customers according to their individual needs, thus ensuring equal rights access regardless of location or financial situation. Nonetheless, there is still a risk that algorithms could perpetuate existing biases within society if they are not properly monitored and regulated by stakeholders across different sectors.
Target 1.5: Build Resilience & Reduce Vulnerability: AI-based tools such as predictive modeling can help identify vulnerable populations before a disaster occurs so that aid groups can better plan relief efforts accordingly. Additionally, AI-powered analytics can assess the magnitude of disasters faster than traditional methods so that relief services can be deployed quicker in order to assist those affected by disasters with greater speed and accuracy than before. Despite these potential positive outcomes from leveraging AI in disaster relief operations, there are still risks associated with this technology such as errors in predictions due to inadequate datasets or security breaches resulting from malicious actors infiltrating systems storing potentially sensitive data about vulnerable communities
3.2. Goal 2. End hunger, achieve food security and improved nutrition and promote sustainable agriculture
Target 2.1: End hunger: AI can help to develop and implement strategies for improving access to food, helping to reduce food insecurity. This can involve analyzing data to identify vulnerable populations and track the distribution of resources. It can also be used to optimize crop production by monitoring soil conditions and weather patterns, while reducing water usage and deforestation. However, AI could lead to ethical issues such as exploitation of labor or displacement of agricultural workers due to automated processes.
Target 2.2: End malnutrition: AI can be used to analyze nutrition data at a population level, providing better insight into which nutrient deficiencies occur in specific regions. This information can help inform food assistance programs, improve supply chain management and manage health care services more effectively. AI could also aid in developing new foods with higher nutritional value or improved taste profiles tailored for different populations. However, this could lead to challenges in terms of the accuracy of the data collected and how it is interpreted or misinterpreted by decision makers.
Target 2.3: Increase small-scale food producer incomes: AI can be used to provide financial services tailored for small-scale farmers such as microfinance loans or insurance products that are more accessible than traditional banking services. It can also help identify potential markets for their products and improve production yields through optimized farming techniques based on weather forecasts and soil analysis data. On the other hand, there is a risk that AI may increase the digital divide between those who have access to technology-based solutions versus those who do not have access or the skills to use them effectively.
Target 2.4: Sustainable food production systems: AI can be used in precision agriculture techniques such as automation of irrigation systems or pest control mechanisms that reduce water and chemical inputs while increasing productivity levels. It can also monitor soil quality through remote sensing technologies which will enable early detection of changes in land fertility levels before they become critical problems when combined with other agricultural practices such as crop rotation methods or organic fertilizers usage. Nevertheless, there is a risk that over-reliance on automated solutions might lead to reduced knowledge transfer from experienced farmers who possess important local insights about their environment’s ecology system needs
Target 2.5: Genetic diversity: AI has applications across many areas related to genetic diversity monitoring such as species identification using machine learning algorithms for recognizing images captured from camera traps placed in natural habitats; gene editing with CRISPR technology; genome sequencing technologies; DNA barcoding; robotics for collecting samples from remote locations; and phenotyping using sensors and drones recording plant characteristics like height or leaf area index (LAI). While these tools have great potential for conserving genetic diversity, there are several risks associated with their use including contamination by invasive species due to increased human activity in sensitive ecosystems, inadequate regulation governing their use, genetic pollution caused by mixing wild types with domesticated ones, biopiracy if traditional knowledge isn’t correctly acknowledged during commercialization processes etc.
3.3. Goal 3. Ensure healthy lives and promote well-being for all at all ages
Target 3.1: Reduce global maternal mortality ratio <70/100K live births by 2030: AI can help reduce maternal mortality by using predictive analytics to identify areas where it is highest and provide targeted resources to those regions, as well as provide better access to healthcare through telemedicine and remote monitoring. However, there is a risk of data privacy violation due to the potential for sensitive patient information being collected and stored in AI systems.
Target 3.2: End preventable deaths neonates/under-5s: AI could be used to develop early warning systems for detecting diseases in infants, and allow for better detection of illnesses which can lead to preventable deaths in newborns. On the other hand, this technology may not be accurate enough or may create biases that can lead to wrong diagnoses and unequal access to healthcare services.
Target 3.3: End AIDS, TB, malaria & neglected diseases: AI-powered diagnostics give improved accuracy in diagnosing infectious diseases such as AIDS, tuberculosis and malaria, allowing for earlier diagnosis and treatment which can lead to better outcomes for patients. However, there is a risk of data privacy violations due to the potential for sensitive patient information being collected and stored in AI systems.
Target 3.4: Reduce premature mortality from NCDs: Through predictive analytics, AI can help identify individuals who are at higher risk of developing certain non-communicable diseases (NCDs) such as diabetes or cancer, allowing for earlier diagnosis and treatment which can improve outcomes for patients. On the other hand, there is a risk of creating biased algorithms that may lead to inaccurate diagnoses or unequal access to healthcare services based on race or socioeconomic status.
Target 3.5: Substance abuse prevention & treatment: AI-based technologies such as natural language processing (NLP) can be used to detect drug abuse patterns through analysis of conversations on social media platforms or text messages sent by users who are at risk of substance abuse disorders such as opioid addiction or alcohol misuse. This technology could also be used in online interventions that provide personalized support and advice tailored towards an individual’s needs which could help reduce relapse rates among people with substance use disorders. However, there is a risk of data privacy violations due to the potential for sensitive user information being collected and stored in these systems.
Target 3.6: Halve global road traffic fatalities by 2020: Through advanced driver assistance systems (ADAS), autonomous vehicles powered by AI have the potential to reduce traffic accidents caused by human error - such as driving under influence - leading to fewer fatalities from road traffic accidents globally if widely adopted worldwide. However, relying too heavily on autonomous vehicles could potentially lead to reduced alertness among drivers who become overconfident when relying on automated systems which could cause an increase in road traffic accidents instead of reducing them altogether.
Target 3.7: Universal access sexual/reproductive health care: AI-powered digital health tools such as telemedicine platforms provide improved access to reproductive health care services such as family planning advice or STD testing remotely without needing physical visits at doctor’s offices; helping more people gain access even if they lack physical mobility or live far away from health care facilities available near them; thereby improving their quality of life significantly while also helping reduce unwanted pregnancies–and possibly abortions–globally if widely adopted worldwide. On the other hand, there’s a risk of data privacy violation due collection & storage of patient information into these systems plus potential bias against certain groups since most algorithms currently lack proper safeguards against discrimination based on gender identity & other factors like race/ethnicity etc.
Target 3.8: Universal health coverage incl financial protection: Artificial Intelligence (AI) has been proposed as a way towards providing universal health coverage with its ability process large amounts of medical data quickly & accurately thereby allowing better detection & diagnosis thus preventing unnecessary treatments plus its ability support cost effective decision making through predictive analytics. But on the downside, it has been argued that this technology cannot replace human judgement completely, hence proper checks must be made before implementing any system powered by artificial intelligence so that it does not end up denying essential benefits or treatments unfairly.
Target 3.9: Reduce global death/illness from hazardous chemicals: With advancements in internet connected sensors, machine learning algorithms, drones etc., artificial intelligence can be leveraged upon significantly reduce global death & illness caused due hazardous chemical exposure across multiple industries ranging from oil refineries, manufacturing plants etc. Some benefits include improved compliance with regulations along with real time notifications about any unsafe levels being detected anywhere within an organization’s premises But this technology also comes with risks associated with loss/theft/hacking of confidential environmental data collected via these sensors besides there always exists the possibility that some unethical parties might take advantage off this system either maliciously or unintentionally leading possible irreparable damage.
3.4. Goal 4. Ensure inclusive and equitable quality education and promote lifelong learning opportunities for all
Target 4.1: Ensure Free, Equitable Quality Primary/Secondary Ed: To ensure girls and boys complete free, equitable and quality primary and secondary education by 2030, AI can help streamline the learning process and personalize instruction for each student. Benefits include improved student outcomes and greater efficiency in teaching, while potential risks include bias built into the algorithms or a lack of transparency concerning AI decisions.
Target 4.2: Quality Early Childhood Development: AI can be used to improve early childhood development by providing personalized learning experiences that target individual needs. Benefits include improved educational outcomes for children, while risks include data privacy concerns or an overreliance on automation in lieu of human interaction.
Target 4.3: Equal Access/Affordable/Quality Education: Providing equal access to affordable and quality technical, vocational, and tertiary education is key to ensuring equal opportunity for all genders by 2030. AI can provide personalized instruction tailored to each student’s needs and make educational materials more accessible on various platforms. Benefits include increased access to educational materials regardless of socio-economic status or geographic location; however, potential risks associated with AI-driven instruction could be the development of biased algorithms or unintended consequences due to a lack of understanding among students concerning how algorithms make decisions.
Target 4.4: Increase Relevant Skills for Employment: By using AI-driven technology such as natural language processing (NLP) and machine learning (ML), organizations can quickly assess job candidates’ qualifications and skillsets to match them with suitable employment opportunities. The benefits are numerous, including shorter hiring cycles, better matching between job seekers and employers, as well as cost savings from reducing manual labor in the recruitment process; however, there are also risks associated with this technology such as inaccurate assessments due to bias built into the algorithm or a lack of transparency concerning data use during the assessment process.
Target 4.5: Eliminate Gender Disparities: Increasing access to education for all genders requires eliminating gender disparities in educational attainment levels by 2030. To do this, AI can be leveraged to identify patterns in data that indicate gender disparities in educational opportunities so that steps can be taken to address these issues at their source. Benefits include greater accuracy in identifying disparities than traditional methods; however, there are also risks associated with data privacy concerns or an overreliance on automated solutions rather than human intervention when addressing issues related to gender disparities in education levels worldwide.
Target 4.6: Achieve Literacy & Numeracy: Utilizing AI-driven solutions such as NLP and ML can help facilitate literacy and numeracy achievement rates among youth and adults alike by 2030 through personalized instruction tailored to individual needs based on factors such as skill level or language proficiency level among others. This could lead to improved outcomes for students while also reducing costs associated with traditional teaching methods; however potential risks may involve bias built into the algorithm or unintended consequences due to a lack of understanding among students concerning how algorithms make decisions when it comes to literacy & numeracy achievement levels worldwide.
Target 4.7: Education For Sustainable Development: By leveraging AI tools such as predictive analytics, organizations can create more effective strategies for sustainable development which focus on areas such as human rights, gender equality, promotion of peace & nonviolence etcetera which will benefit society significantly by 2030 if achieved properly. The benefits are numerous including more accurate predictions about sustainable development initiatives leading greater success rates, while potential risks may involve inaccuracies due to bias built into these models or a lack of transparency concerning data use during decision making processes.
3.5. Goal 5. Achieve gender equality and empower all women and girls
Target 5.1: Ending discrimination against women and girls: AI has the potential to reduce gender-based discrimination in the workplace and other areas of life, but it could also be used to reinforce existing biases if not properly monitored.
Target 5.2: Eliminating violence against women and girls: AI can provide valuable insights into patterns of violence, allowing for more effective interventions, but its usage could also lead to increased surveillance and further privacy violations.
Target 5.3: Eliminating harmful practices: AI can be used to identify harmful practices such as child marriage and female genital mutilation, allowing for more targeted interventions, but there is a risk that this data could be misused or abused by those in power.
Target 5.4: Recognizing unpaid care work: AI can help to better understand and value unpaid care work through automated data analysis, but there is a risk that this technology could be deployed to devalue or deny unpaid care work done by certain populations.
Target 5.5: Enhancing women’s leadership: AI can facilitate access to leadership positions for women by automating administrative tasks, creating better visibility of their achievements, and providing mentorship opportunities, however there is a risk that existing biases in training data sets may limit its effectiveness.
Target 5.6: Increasing sexual health access: AI can be used to increase access to sexual health information by providing personalized advice and recommendations based on individual needs, but its usage could lead to unintended consequences such as perpetuating stereotypes about sexuality or stigmatizing certain populations.
3.6. Goal 6. Ensure availability and sustainable management of water and sanitation for all
Target 6.1: By 2030, achieve universal and equitable access to safe and affordable drinking water: AI can help reduce water wastage and ensure the efficient use of resources, leading to cost savings. However, the accuracy of data input is important to ensure the safety of water supply, thus introducing potential risks if there are errors in the data.
Target 6.2: By 2030, achieve access to adequate and equitable sanitation and hygiene: AI can leverage predictive analytics to determine areas where improved sanitation is needed so that resources can be allocated accordingly. However, there is a risk that AI may overlook marginalised communities or fail to accurately identify needs due to lack of data or bias in the algorithms.
Target 6.3: By 2030, improve water quality by reducing pollution: AI can help detect sources of water pollution quickly and accurately which will enable timely interventions for improving water quality. There is a risk that AI-powered solutions may not be as reliable as manual inspections when it comes to detecting uncommon pollutants or contaminants in the water.
Target 6.4: By 2030, substantially increase water-use efficiency: AI-enabled tools such as sensors can provide real-time insights into how much water is being used and identify any wastage or misuse quickly, leading to more efficient use of resources. There is a risk that these tools could lead to privacy concerns if they collect personal data without proper consent or security safeguards in place.
Target 6.5: By 2030, implement integrated water resources management: Using AI technologies such as machine learning and natural language processing (NLP), it will be possible to gain insights on how best to manage different types of water resources across multiple jurisdictions while considering various stakeholders’ interests efficiently. However, this approach may lead to conflicts between stakeholders due to differences in opinion on how best use these resources effectively and sustainably.
Target 6.6: By 2020, protect and restore water-related ecosystems: AI can provide valuable insights into how different ecosystems interact with each other by analyzing large volumes of data from numerous sources quickly and accurately for better decision making for conservation efforts. Nevertheless, there is a risk that AI may be unable to detect subtle changes in ecosystems over time due its reliance on historical data sets which may not reflect current conditions accurately enough.
3.7. Goal 7. Ensure access to affordable, reliable, sustainable and modern energy for all
Target 7.1: Universal Access to Affordable Energy: AI can be used to improve energy forecasting and optimize energy supply networks, making it easier for people to access affordable energy sources. However, AI-enabled energy management systems could potentially lead to price manipulation and increase the digital divide between those who can afford advanced technology and those who cannot.
Target 7.2: Increase Renewable Energy Share: AI can help with renewable energy forecasting and optimization of renewable energy sources, increasing the share of renewables in global energy mix. However, if not carefully monitored, AI-driven decision making can lead to over-investment in some renewable sources while neglecting others that may be more suitable for certain locations.
Target 7.3: Double Global Energy Efficiency Improvement: AI can be used to optimize existing processes by detecting areas of inefficiency and reducing waste associated with them. On the other hand, relying on automation for efficiency improvements might lead to job losses or potential security risks in case of a cyber attack on an automated system.
3.8. Goal 11. Make cities and human settlements inclusive, safe, resilient and sustainable
Target 11.1: By 2030, ensure access for all to adequate, safe and affordable housing and basic services and upgrade slums: AI could be used to identify areas in need of improvement and implement solutions faster, with greater accuracy and at a larger scale than manual labor. However, the risk is that AI may overlook unique local needs due to its focus on efficiency over effectiveness.
Target 11.2: By 2030, provide access to safe, affordable, accessible and sustainable transport systems for all: AI can be used to identify patterns in transportation usage and suggest ways to improve safety through improved navigation systems or automated traffic controls. However, there is a risk of bias in data collection if AI is not designed with diverse user needs in mind.
Target 11.3: By 2030, enhance inclusive and sustainable urbanization and capacity for participatory, integrated and sustainable human settlement planning and management in all countries: AI can help governments make better decisions about urban planning by providing more accurate predictions about population growth or changes in infrastructure demands. However, there is a risk that decision-making will become too reliant on algorithms without proper oversight from stakeholders or experts.
Target 11.4: Strengthen efforts to protect and safeguard the world’s cultural and natural heritage: AI can be used to analyse massive amounts of data on cultural artifacts or ecosystems quickly with unprecedented accuracy. This could help authorities better understand how to preserve these places for future generations. However, there is a risk that AI may overlook important details or ignore the context of certain cultural artifacts due to its focus on data analysis over interpretation.
Target 11.5: By 2030, significantly reduce the number of deaths and the number of people affected by disasters: AI could be used to predict when disasters are likely to occur by analysing data from past events or monitoring changes in atmospheric conditions with high precision sensors. This could help authorities take preventative action before disasters occur. However, there is a risk that this technology may not consider other factors such as human behaviour which could have an impact on disaster prevention efforts.
Target 11.6: By 2030 reduce the adverse per capita environmental impact of cities: AI could be used to monitor air quality or analyse waste management systems efficiently so as to identify problem areas quickly while also suggesting solutions tailored towards individual cities’ unique situations. However, there is a risk that any recommendations made by AI may not take into account social considerations such as equity or justice when it comes developing sustainable practices for cities.
Target 11.7: By 2030 provide universal access green public spaces: AI can be used identify areas suitable for green public spaces based on factors like climate change adaptation potential, pollution levels, biodiversity conservation etc. This would enable city planners design green spaces more efficiently. But this would require lot of local knowledge which might get lost if decision making process becomes too dependent on data generated through algorithms.
3.9. Goal 16. Promote peaceful and inclusive societies for sustainable development, provide access to justice for all and build effective, accountable and inclusive institutions at all levels
Target 16.1: Reduce Violence & Death Rates - AI can help to predict and prevent acts of violence before they occur, as well as improve response time in emergency situations. However, AI may also be used to target specific populations, such as minority groups or individuals who are perceived to be at higher risk for violence or crime.
Target 16.2: End Abuse Against Children - AI can be used to detect potential signs of abuse and alert authorities when necessary. However, there is a risk that AI systems may misinterpret data and lead to false accusations or incorrect decisions about whether a child is being abused.
Target 16.3: Promote Rule of Law - AI can monitor laws and regulations and alert decision-makers if any changes are needed. This could lead to faster responses in legal disputes and better compliance with international laws and treaties. But there is also a risk that biased algorithms could be used to unfairly discriminate against certain groups of people or countries.
Target 16.4: Reduce Illicit Flows - AI can track money flows quickly and accurately, enabling more effective enforcement of anti-money laundering regulations and helping reduce the number of illicit financial flows across borders. But it could also lead to increased surveillance measures that infringe on privacy rights.
Target 16.5: Reduce Corruption & Bribery - AI could help identify suspicious activities related to bribery and corruption by analyzing large datasets for patterns that would otherwise go undetected by human investigators alone. But there is still a risk that corrupt actors may use AI technologies for their own benefit without detection from authorities due to the complexity of these systems.
Target 16.6: Develop Effective Institutions - AI-driven solutions can facilitate processes within institutions, such as streamlining administrative tasks or improving communication between departments for better decision-making capabilities at all levels of government institutions worldwide. However, there is potential for misuse by malicious actors if adequate security measures are not implemented properly in these systems.
Target 16.7: Inclusive Decision Making - Using machine learning algorithms, organizations can analyze large amounts of data quickly to identify patterns in decision making processes which can then inform more inclusive policies in the future by providing more accurate insights into public opinion than traditional methods like surveys could provide alone. But there is also the risk that these same algorithms may contain inherent biases if not developed properly with safeguards against discrimination built into them from the start.
Target 16.8: Participation in Global Governance - By using AI-driven analysis tools, developing countries may have improved access to global governance bodies where their interests will be better represented due to their increased visibility within those organizations. However, this technology must be carefully monitored so as not to give an unfair advantage or disadvantage to any country based on its technological capabilities.
Target 16.9: Legal Identity for All - Applying facial recognition technology powered by artificial intelligence can enable governments around the world provide secure legal identity documents quickly while reducing fraud risks associated with manual verification processes. On the other hand, this technology has been shown to have racial biases which could lead discriminatory practices against certain populations if not implemented properly with strict oversight mechanisms in place.
Target 16.10: Public Access to Info and Fundamental Freedoms: AI can help protect the public’s access to information, such as by monitoring online platforms for hate speech and censorship. But it can also be used to violate fundamental freedoms, such as by profiling individuals based on their digital footprints and targeting them with politically motivated ads. Therefore, it is essential to ensure that AI is used in a responsible way that respects international agreements and national legislation.
3.2. Descriptive analysis
Table 2 presents the results of the descriptive analysis of the queries related to the SDGs and their targets, highlighting the word count.
We found that the Goal 1 outcome was quite precise and in exactly the format we expected the AI outcome. Each target was numbered correctly, the titles were shortened down to three to four words, and each paragraph contained exactly three sentences with an argument with opportunities for AI contribution, a further argument or dimension in a second sentence and third sentence containing risks and potential harm the AI could produce. We observed, that although the amount of three sentences stayed the same for each target, the number of words increase with every target, after the first one: 1.1: 87 words / 1.2: 75 words / 1.3: 79 words / 1.4: 93 words / 1.5: 112 words. We assumed that this happened due to the overall text length and the “presence penalty” parameter setting, which defines “how much to penalize new tokens based on whether they appear in the text so far. Increases the model’s likelihood to talk about new topics”. We chose a setting of 0.5 on a scale of 0 to 2 to avoid too much word duplications. Interestingly, the last text block did not close with a punctuation.
As for SDG 2, we noticed that the target descriptions consisted of less sentences and more words: The AI started with 79 words in 4 sentences for target 2.1, and came down to only 2 complex sentences with 123 words for target 2.5. Target 2.4 contained a punctuation mistake with including a blank character before the punctuation sign “[…] organic fertilizers usage.”. The case of SDG 3 highlighted the AIs capabilities in regards of abbreviations and shortening text. It shortened “less than 70 per 100,000 live births” down to “<70/100K live births”, and listed the abbreviations of TB for tuberculosis, NCD for non-communicable diseases or STD for sexual transmittable diseases. Especially in the later text passages, many punctuation mistakes were added by the AI, one point at the end of sentence was omitted by the AI in 3.9.
The analysis showed that the AI introduced a so far unseen way of answering structure as of SDG 4. The AI used exactly two sentences when answering in regard to all targets, whereas the first one listed the opportunities of AI. The second sentence listed further opportunities and then used a separator word to continue with risks, while in previous outputs the last generated sentence per target was solely dedicated to potential risks. For SDG 5, again, the AI introduced a further way of answering – only one sentence per target including benefits and risks. This resulted in the so far lowest number of words in the total answer, as well as in the lowest mean number of words per target. In SDG 6, the AI put emphasis on the delivery timeframe for the first time. It prefixed all goals with the expected delivery date, kept a two-sentence structure with one for benefits and the other for risks, and only a slight increase of the word count per target.
SDG 7 showcased the so far most precise and shortest sentences with only 27.3 words per sentence. SDG 11 showed a similar target text summarizing scheme than seen in SDG 6 in adding “By 2023” to all targets except to target 11.4. When we checked this against the original texts – the AI was correct, despite all other targets under SDG 11, the original target name for 11.4 did not contain a delivery timeframe. We detected another obvious difference to the SDG 6 summaries: SDG 11 summaries (mean 15.4, SD 4.3) were much longer than all SDG 6 summaries (mean 9.7, SD 2.5). Overall, the summarizing quality of SDG 11 can therefore be considered much weaker than from SDG 6.
In SDG 16 the AI introduced a new way of grouping summary and answers: for the first 7 targets the scheme “numbering[blank]summary[blank]- [blank]answer“ was used. Due to the length of the query, a second query was required to be made with the 10th target alone, and the AI used the “: ” as separator again. Interestingly, in 16.10 for the first time the AI used just one sentence for benefits, but two full sentences for potential risks.
3.3. Analysis of detected patterns
We further analyzed the output patterns, differentiated in the three areas summarization, numbering and summarized title structure, answering structure, and general patterns with longer texts. Firstly, summarization was generally performed very well. Most of the summaries were created with normal case, for some of the summaries GPT-3 used capitalized letters. Typically, it replaced the word “and” with an “&” sign. When all targets of a query used the same delivery timeframe e.g. “By 2030, “, GPT-3 removed it as unnecessary prefix. In some cases, when at least one delivery timeframe was different or not provided, GPT-3 decided to include this differentiation into the summary as well. Still, this led to more mistakenly fully shortened exceptions (e.g. 2.5, SDG 3, 8.5, 8.9, SDG 9, SDG 10, SDG 12) than real results (e.g. 6.6, 11.4). As for, numbering and summarized title structure, GPT-3 usually used the provided numbering and structuring format also for the summarized titles, i.e. “[number][space][title][:][space][answer]”. In exceptional cases it decided to use “[number][optional space][-][space][title][:][space][answer]” (SDG 10), or replaced the “:” with a “-“ like in SDG 16: “[number][space][title][-][space][answer]”. In target 17.19, it mixed up the numbering and wrote “17 17” mistakenly.
Regarding, answering sentence structure, we asked for 3-5 sentences including benefits and risks. Yet, the AI most often answered with the following answering schemes:
- a.
Combined 1 sentence: 1 sentence with benefits and risks
- b.
Classic 2 sentence split: 1 sentence with benefits, 1 sentence with risks
- c.
Modified 2 sentences: 1 sentence with benefits plus 1 sentence with benefits combined with risks
- d.
Classic 3 sentences: 2 sentences with benefits, 1 sentence with risks
- e.
Modified 3 sentences: 2 sentences with benefits, 1 sentence with benefits combined with risks
- f.
Unusual 3 sentences (detected only once): 1 sentence with benefits, plus 2 sentences with risks.
Lastly, when dealing with longer texts, the AI showed a distinct pattern in writing. It typically started with more sentences in the first few paragraphs, but gradually decreased the number of sentences while increasing the number of words per paragraph. This led to longer answers as the number of words increases with each subsequent response, starting from the first paragraph (SDG 2, SDG 9, all four SDG 17 queries). Furthermore, there was an increased amount of punctuation mistakes especially in longer texts, and especially more at the end of the texts.