Besides issues of extracting value from large data sets, Scriffignano believes that the primary challenges for AI and data science are focused around governance and ethics
. This is especially the case when personal information is involved. How can we make sure we’re making responsible use of private information when building large databases and building intelligent models that use that private information?
Part of the reason why there’s increased scrutiny on machine learning models has more to do with issues of privacy and security than on the specific characteristics of those models. Scriffignano makes an interesting point by stating just how troublesome it will be to cater to everyone in terms of AI regulation
with the variation in needs and wants. People are looking for more customization and more rapid model development yet are not willing to compromise on their privacy. Some companies and individuals will benefit from models that use lots of data to create much more precise and accurate predictions, but at the expense of scooping up large amounts of private information. Others might resist the inclusion of their data in those models even if it results in less-accurate models that they might end up depending on. As a result, not everyone will be satisfied by greater expansion of data used to build machine learning models.
Scriffignano believes that Government regulators will need to keep up with evolving technologies if they wish to ensure optimal national security and avoid these privacy-related issues. These laws and regulations will vary heavily in different regions of the world, and as such, the notion of ethics might not even be consistent across different jurisdictions. Ethics, and the resulting laws, will vary largely from country to country and regions as it does now with Europe taking a more ethical approach, China being less interested in privacy, and the United States somewhere in the middle. Some countries are simply more interested in privacy, with others looking towards national security and even economic advancement. The issue with this, as Scriffignano shares, is that machine learning really doesn’t have any geographic boundaries. What might be unacceptable in one location might be perfectly fine in another. The models will be built and then be available for use in other regions. It might be very difficult to control the spread of models developed in one region with less care for privacy that might be used in another location with higher regard for data ethics.
On the podcast, Scriffignano also shares his dislike of anthropomorphizing AI. Taking a much more practical approach, Scriffignano reminds us that our current evolution of AI functions by means of algorithms and processes. Scriffignano uses Artificial General Intelligence (AGI)
as an example from which to share his point of view. Our limitations start when we cannot ask the right question of the copious amount of data we possess. Scriffignano foresees a future in which professionals will work alongside AI and that we need not fear answering to robots or machines as long as we are vigilant about it. To achieve this outcome, we must be stringent and alert to data ethics and governance issues to allow this progress to be made without harm.
安東尼博士在節目上也表示，他討厭將AI擬人化。他抱持務實態度，提醒聽眾目前AI演化運作靠的是演算法及程式。他以「通用人工智慧」（Artificial General Intelligence，簡稱AGI）為例分享看法。當人們無法就手上的海量數據提出正確問題時，就不能開展下一步。安東尼博士預見未來的專業人士將與AI並肩；只要保持警醒，就無需恐懼機器人或機器將凌駕人類之上。人們需戒慎處理數據倫理與治理議題，方能安然推展進程，以竟全功。