{"?xml":{"@version":"1.0"},"edm:RDF":{"@xmlns:dc":"http://purl.org/dc/elements/1.1/","@xmlns:edm":"http://www.europeana.eu/schemas/edm/","@xmlns:wgs84_pos":"http://www.w3.org/2003/01/geo/wgs84_pos","@xmlns:foaf":"http://xmlns.com/foaf/0.1/","@xmlns:rdaGr2":"http://rdvocab.info/ElementsGr2","@xmlns:oai":"http://www.openarchives.org/OAI/2.0/","@xmlns:owl":"http://www.w3.org/2002/07/owl#","@xmlns:rdf":"http://www.w3.org/1999/02/22-rdf-syntax-ns#","@xmlns:ore":"http://www.openarchives.org/ore/terms/","@xmlns:skos":"http://www.w3.org/2004/02/skos/core#","@xmlns:dcterms":"http://purl.org/dc/terms/","edm:WebResource":[{"@rdf:about":"http://www.dlib.si/stream/URN:NBN:SI:doc-ID8EZS9P/8a633f54-c6dd-462a-a440-9218786abf96/PDF","dcterms:extent":"351 KB"},{"@rdf:about":"http://www.dlib.si/stream/URN:NBN:SI:doc-ID8EZS9P/c7d76e2b-f64c-4090-bb96-edad83a8510f/TEXT","dcterms:extent":"45 KB"}],"edm:TimeSpan":{"@rdf:about":"2018-2024","edm:begin":{"@xml:lang":"en","#text":"2018"},"edm:end":{"@xml:lang":"en","#text":"2024"}},"edm:ProvidedCHO":{"@rdf:about":"URN:NBN:SI:doc-ID8EZS9P","dcterms:isPartOf":[{"@rdf:resource":"https://www.dlib.si/details/URN:NBN:SI:spr-QVBBDN4R"},{"@xml:lang":"sl","#text":"Analiza (Ljubljana)"}],"dcterms:issued":"2020","dc:creator":"Bregant, Janez","dc:format":[{"@xml:lang":"sl","#text":"številka:1"},{"@xml:lang":"sl","#text":"letnik:24"},{"@xml:lang":"sl","#text":"str. 5-20"}],"dc:identifier":["ISSN:2712-4916","COBISSID_HOST:62996995","URN:URN:NBN:SI:doc-ID8EZS9P"],"dc:language":"sl","dc:publisher":{"@xml:lang":"sl","#text":"Društvo za analitično filozofijo in filozofijo znanosti"},"dc:subject":[{"@xml:lang":"en","#text":"artificial intelligence"},{"@xml:lang":"sl","#text":"človekove pravice"},{"@xml:lang":"en","#text":"ethics"},{"@xml:lang":"sl","#text":"etika"},{"@xml:lang":"en","#text":"human rights"},{"@xml:lang":"en","#text":"machine learning"},{"@xml:lang":"en","#text":"natural learning processing"},{"@xml:lang":"sl","#text":"obdelava naravnega jezika"},{"@xml:lang":"sl","#text":"strojno učenje"},{"@xml:lang":"sl","#text":"umetna inteligenca"}],"dcterms:temporal":{"@rdf:resource":"2018-2024"},"dc:title":{"@xml:lang":"sl","#text":"Umetna inteligenca v praksi| Artificial intelligence at work|"},"dc:description":[{"@xml:lang":"sl","#text":"A recent development of artificial intelligence (AI) and its self-evident use in our everyday life have increased interest of the public and experts for all kinds of artificially made autonomous systems. Since AI is not anymore understood merely as a handy tool but also as an independent agent, the question of its impact on our life in terms of morality emerges naturally. Answers to questions such as \"Is AI responsible for its actions?\", \"What are social, legal and cultural consequences of its decisions?\", \"Should its autonomous models be freely accessible in the market?\" etc., determine to a great extent whether AI will be trustworthy in the future or not. The paper introduces five threats associated with our unbridled use of AI and concludes through the analyses of current developments in the fields related to those threats that the society is increasingly becoming a hostage of the AI industry: intelligent systems are not made in a way that they would, while interacting with people, in any of their evolution stages (design, construction, installation and evaluation) respect human rights nor act according to accepted social values"},{"@xml:lang":"sl","#text":"Nagel razvoj umetne inteligence (UI) v zadnjem obdobju in njena samoumevna uporaba v našem vsakodnevnem življenju sta povečala zanimanje za vse vrste umetno izdelanih avtonomnih sistemov tako na strani javnosti kot stroke. Ker UI ni več razumljena zgolj kot priročno orodje, ampak tudi kot samostojni akter, ne moremo mimo vprašanja, kakšen vpliv ima njena uporaba na naše življenje v moralnem smislu. Odgovori na vprašanja, kot so \"Ali je UI odgovorna za svoja dejanja?\", \"Kaj so družbene, pravne in kulturne posledice njenih odločitev?\", \"Ali naj bi bili avtonomni sistemi prosto dostopni na trgu?\" itn., so tisto, kar v veliki meri določa, ali bomo UI v prihodnosti zaupali ali ne. V članku predstavimo pet nevarnosti, ki nam ob nebrzdani uporabi UI pretijo in s pomočjo analize trenutnega dogajanja na področjih, s katerimi so povezane, zaključimo, da družba vse bolj postaja talec industrije inteligentnih sistemov, ki v nobeni fazi svojega razvoja (oblikovanje, izgradnja, namestitev in ocenjevanje) niso narejeni tako, da bi v interakciji z nami spoštovali človekove pravice in ravnali skladno s sprejetimi družbenimi vrednotami"}],"edm:type":"TEXT","dc:type":[{"@xml:lang":"sl","#text":"znanstveno časopisje"},{"@xml:lang":"en","#text":"journals"},{"@rdf:resource":"http://www.wikidata.org/entity/Q361785"}]},"ore:Aggregation":{"@rdf:about":"http://www.dlib.si/?URN=URN:NBN:SI:doc-ID8EZS9P","edm:aggregatedCHO":{"@rdf:resource":"URN:NBN:SI:doc-ID8EZS9P"},"edm:isShownBy":{"@rdf:resource":"http://www.dlib.si/stream/URN:NBN:SI:doc-ID8EZS9P/8a633f54-c6dd-462a-a440-9218786abf96/PDF"},"edm:rights":{"@rdf:resource":"http://rightsstatements.org/vocab/InC/1.0/"},"edm:provider":"Slovenian National E-content Aggregator","edm:intermediateProvider":{"@xml:lang":"en","#text":"National and University Library of Slovenia"},"edm:dataProvider":{"@xml:lang":"sl","#text":"Društvo za analitično filozofijo in filozofijo znanosti"},"edm:object":{"@rdf:resource":"http://www.dlib.si/streamdb/URN:NBN:SI:doc-ID8EZS9P/maxi/edm"},"edm:isShownAt":{"@rdf:resource":"http://www.dlib.si/details/URN:NBN:SI:doc-ID8EZS9P"}}}}