Читать книгу: «Digital Transformation for Chiefs and Owners. Volume 1. Immersion», страница 4

Шрифт:

Neural networks, machine and deep learning (ML & DL), speech and text recognition systems

So, we’re getting to the future – neural networks, artificial intelligence, machine revolutions and other horror stories.

Neural networks are perhaps the most interesting technology. With the support of the Internet of Things, 5G and Big Data, it will bring revolutionary changes to our lives.

Additionally, artificial intelligence is any mathematical method that can simulate human intelligence.

Oh as our favorite advertisers and marketers are satisfied… Now any, the simplest neuronetwork can proudly be called «Artificial Intelligence».

However, artificial intelligence is still divided into strong and weak. In 2019, scientists came close to creating a strong AI, the equivalent of human consciousness. This ability not only to distinguish a pen from a pencil or a cat from a dog (according to this principle all neural networks work, it is weak AI), but also to navigate changing conditions, choose specific solutions, model and predict the development of the situation.

A strong AI will be indispensable in intelligent transportation and transportation systems, cognitive assistants. However, this is the future, and what is now?

Now there are learning neural networks. An artificial neural network is a mathematical model modeled on the neural networks that make up the brains of living things. Such systems learn to perform tasks by treating them without specific programming for specific applications. This can be found in Yandex Music, Tesla autopilots, referral systems for doctors and managers.

Therefore, here are the two main trends:

– machine learning (ML – machine learning);

– deep learning (DL – deep learning).

Machine learning is statistical methods that enable computers to improve the quality of the task with experience and training. So it’s about how the neural networks of living organisms work.

Deep learning is not only learning a machine with the help of a person who says what is right and what is not, but also self-learning systems. This is the simultaneous use of different methods of training and data analysis.

However, how do these neural networks teach? What’s the magic?

Actually, in fact, nothing. It’s like training a dog. Neuronetworks show, for example, a picture and say that it is depicted. The neural network must then respond, and if the answer is wrong, it is corrected. An approximate algorithm is given below.

As a result, it turns out that each «neuron» of such a network learns to recognize, refers to it this picture, or rather its part, or not.

Example of neural network operation in image recognition


Neural networks and machine learning apply:

– for forecasting and decision making;

– image recognition and generation, including «pictures» and voice recordings;

– complex data analysis without clear relationships;

– process streamlining.

The application value of this can be seen in the examples of the creation of unmanned cars (decision-making), the search for illegal content (data analysis), the prediction of diseases (pattern recognition and linkage search). At the same time, on the haip it is pattern recognition and generative models (chatGPT, midjourney, etc.). However, business problems are still poorly solved. At the same time, 9 out of 10 students now go to study exactly on pattern recognition and machine vision.

The AI + IoT link deserves special attention:

– AI receives net big data (about them in the next section) in which there are no human factor errors to learn and search relationships;

– IoT’s effectiveness increases as it becomes possible to create predictive (predictive) analytics and early detection of deviations.

Okay, this is all a theory. I want to share a real example of how neuronetworks can be used in business.

In the summer of 2021, I was approached by an entrepreneur from the realtor sector. He is engaged in the rental of real estate, including daily rent. Its goal is to increase the pool of rented apartments and change the status of an entrepreneur to a full-fledged organization. The nearest plans are to launch the site and mobile application.

I happen to be a client myself. And at our meeting I noticed a very big problem – the long preparation of the contract: it takes up to 30 minutes for the registration of all the details and signing. This is both the limitation of the loss generating system and the inconvenience for the customer.

Imagine that you want to spend time with a girl, but you have to wait half an hour for your passport details to be entered into the contract, checked and signed.

Now there is only one option to eliminate this inconvenience – ask for passport photos in advance and manually enter all the data into the template of the contract. As you can imagine, that’s not very convenient either.

How can digital tools help solve this problem, but also provide the basis for working with data and analytics?

Neural networks. The client sends photo passports, the neural network recognizes data and enters the template or database. It remains only to print out the ready contract or to sign in electronic form. Additionally, the advantage here is that all passports are standardized. The series and the number are always printed in the same color and font, the division code too, and the list of issuing units is not very large. To teach such a neuronetwork can be easy and fast. Cope even student in the thesis. As a result, the business saves on development, and the student gets a current thesis. Besides, every time we make a mistake, the neural net gets smarter.

As a result, instead of 30 minutes, the signing of the contract takes about 5. That is, with an eight-hour working day, 1 person will be able to conclude not 8 contracts (30 minutes for registration and 30 minutes for the road), but 13—14. Additionally, this is with a conservative approach – without electronic signature, access to the apartment through a mobile app and smart locks. However, I believe that immediately implement «fancy» solutions and do not need. There’s a high probability of spending money on something that doesn’t create value or reduce costs. This will be the next step after the client receives the result and competence.

Restrictions

Personally, I see the following limitations in this direction.

– Quality and quantity of data. Neuronets are demanding on quality and quantity of source data. However, this problem is being solved. If previously it was necessary to listen to several hours of audio recordings to synthesize your speech, now only a few minutes. Additionally, the next generation will only take a few seconds. However, they still need a lot of tagged and structured data. Additionally, every mistake affects the ultimate quality of the trained model.

– The quality of the «teachers». Neuronetworks teach people. Additionally, there are a lot of limitations: who teaches what, on what data, for what.

– Ethical component. I mean the eternal dispute of who to shoot down the autopilot in a desperate situation: an adult, a child or a pensioner. There are countless such disputes. There is no ethics, good or evil for artificial intelligence.

So, for example, during the test mission, the drone under the control of the AI set the task of destroying the enemy’s air defence systems. If successful, the AI would receive points for passing the test. The final decision whether the target would be destroyed would have to be made by the UAV operator. During a training mission, he ordered the drone not to destroy the target. In the end, AI decided to kill the cameraman because the man was preventing him from doing his job.

After the incident, the AI was taught that killing the operator was wrong and points would be removed for such actions. The AI then decided to destroy the communication tower used to communicate with the drone so that the operator could not interfere with it.

– Neural networks cannot evaluate data for reality and logic.

– The readiness of people. We must expect a huge resistance of people whose work will be taken by the networks.

– Fear of the unknown. Sooner or later, the neural networks will become smarter than us. Additionally, people are afraid of this, which means that they will retard development and impose numerous restrictions.

– Unpredictability. Sometimes it all goes as intended, and sometimes (even if the neural network does its job well) even the creators struggle to understand how the algorithms work. Lack of predictability makes it extremely difficult to correct and correct errors in neural network algorithms.

– Activity constraint. AI algorithms are good for performing targeted tasks, but do not generalize their knowledge. Unlike humans, an AI trained to play chess cannot play another similar game, such as checkers. In addition, even in-depth training is not good at processing data that deviates from his teaching examples. To use the same ChatGPT effectively, you need to be an industry expert from the beginning and formulate a conscious and clear request, and then check the correctness of the answer.

– Costs of creation and operation. To create neuronetworks requires a lot of money. According to a report by Guosheng Securities, the cost of learning the natural language processing model GPT-3 is about $1.4 million. It may take $2 million to learn a larger model. For example, ChatGPT only requires over 30,000 NVIDIA A100 GPUs to handle all user requests. Electricity will cost about $50,000 a day. Team and resources (money, equipment) are required to ensure their «vital activity». It is also necessary to consider the cost of engineers for escort.

P.S.

Machine learning is moving towards an increasingly low threshold of entry. Very soon it will be as a website builder, where basic application does not need special knowledge and skills.

Creation of neural networks and data-companies is already developing on the model of «service as a service», for example, DSaaS – Data Science as a Service.

The introduction to machine learning can begin with AUTO ML, its free version, or DSaaS with initial audit, consulting and data markup. At the same time, even data markup can be obtained for free. All this reduces the threshold of entry.

The branch neuronetworks will be created and the direction of recommendatory networks, so-called digital advisers or solutions of the class «support and decision-making system (DSS) for various business tasks» will be developed more actively.

I discussed the AI issue in detail in a separate series of articles available via QR and link.


AI

Big Data (Big Data)

Big data (big data) is the cumulative name for structured and unstructured data. Additionally, in volumes that are simply impossible to handle manually.

Often this is still understood as tools and approaches to work with such data: how to structure, analyze and use for specific tasks and purposes.

Unstructured data is information that has no predefined structure or is not organized in a specific order.

Application Field

– Process Optimization. For example, big banks use big data to train a chat bot – a program that can replace a live employee with simple questions, and if necessary, will switch to a specialist. Or the detection of losses generated by these processes.

– Forecasting. By analysing big sales data, companies can predict customer behaviour and customer demand depending on the season or the location of goods on the shelf. They are also used to predict equipment failures.

– Model Construction. The analysis of data on equipment helps to build models of the most profitable operation or economic models of production activities.

– Sources of Big Data Collection

– Social – all uploaded photos and sent messages, calls, in general everything that a person does on the Internet.

– Machine – generated by machines, sensors and the «Internet of things»: smartphones, smart speakers, light bulbs and smart home systems, video cameras in the streets, weather satellites.

– Transactions – purchases, transfers of money, deliveries of goods and operations with ATMs.

– Corporate databases and archives. Although some sources do not assign them to Big Data. Here there are disputes. Additionally, the main problem – non-compliance with the criteria of «renewability» of data. More about this a little below.

Big Data Categories

– Structured data. Have a related table and tag structure. For example, Excel tables that are linked together.

– Semi-structured or loosely structured data. They do not correspond to the strict structure of tables and relationships but have «labels» that separate semantic elements and provide a hierarchical structure of records. Like information in e-mails.

– Unstructured data. They have no structure, order, hierarchy at all. For example, plain text, like in this book, is image files, audio and video.

Such data is processed on the basis of special algorithms: first, the data is filtered according to the conditions that the researcher sets, sorted and distributed among individual computers (nodes). The nodes then calculate their data blocks in parallel and transmit the result of the computation to the next stage.

Big data feature

According to different sources, big data have three, four and, according to some opinions, five, six or even eight components. However, let’s focus on what I think is the most sensible concept of four components.

– Volume (volume): Information should be a lot. Usually speak of quantity from 2 terabytes. Companies can collect a huge amount of information, the size of which becomes a critical factor in analytics.

– Velocity (speed): data must be updated, otherwise they become obsolete and lose value. Almost everything that happens around us (search queries, social networks) produces new data, many of which can be used for analysis.

– Variety (variety): generated information is heterogeneous and can be presented in different formats: video, text, tables, numerical sequences, sensor readings.

– Veracity (reliability): the quality of the data analysed. They must be reliable and valuable for analysis, so that they can be trusted. Low-fidelity data also contain a high percentage of meaningless information, which is called noise and has no value.

Restrictions on the Big Data Implementation

The main limitation is the quality of the raw data, critical thinking (what do we want to see? What pain? – This is done ontological models), the right selection of competencies. Well, and most importantly – people. Data-Scientists are engaged in work with the data. Additionally, there is one common joke: 90% of the data-scientists are data-satanists.

Digital doppelgangers

A digital double is a digital/virtual model of any object, system, process or person. In its conception, it accurately reproduces the shape and actions of the physical original and is synchronized with it. The error between the double and the real object must not exceed 5%.

It must be understood that it is almost impossible to create an absolute digital counterpart, so it is important to determine which domain is rationally modelled.

The concept of the digital counterpart was first described in 2002 by Michael Grieves, a professor at the University of Michigan. In the book «The Origin of Digital Doubles» he divided them into three main parts:

1) physical product in real space;

2) virtual product in virtual space;

3) data and information that combine virtual and physical products.

The digital double itself can be:

– prototype – the analogue of the real object in the virtual world, which contains all the data for the production of the original;

– a copy – a history of operation and data about all characteristics of the physical object, including the 3D model, the copy operates in parallel with the original;

– an aggregated double – a combined system of a digital double and a real object that can be controlled and shared from a single information space.

The development of artificial intelligence and the cheapening of the Internet of Things have made technology the most advanced. Digital doubles began to receive «clean» big data about the behaviour of real objects, it became possible to predict equipment failures long before accidents. Although the latter thesis is quite controversial, this direction is actively developing.

As a result, the digital double is a synergy of 3D technologies, including augmented or virtual reality, artificial intelligence, the Internet of Things. It’s a synthesis of several technologies and basic sciences.

The digital counterparts themselves can be divided into four levels.

• The double of the individual assembly unit simulates the most critical assembly unit. It can be a specific bearing, motor brushes, stator winding or pump motor. In general, the one that has the greatest risk of failure.

• The twin of the unit simulates the operation of the entire unit. For example, the gas turbine unit or the entire pump.

• The production system double simulates several assets linked together: the production line or the entire plant.

• Process counterpart – this is no longer about «hardware» but about process modelling. For example, when implementing MES- or APS-systems. We’ll talk about them in the next chapter.

What problems can digital duplicate technology solve?

• It becomes possible to reduce the number of changes and costs already at the stage of designing the equipment or plant, which allows to significantly reduce costs at the remaining stages of the life cycle. Additionally, it also avoids critical errors, which cannot be changed at the stage of operation.


The sooner an error is detected, the cheaper it is to fix it


In addition to cost increases, there is less room for error correction over time


– By collecting, visualizing and analyzing data, it is possible to take preventive measures before serious accidents and damage to equipment.

– Optimize maintenance costs while increasing overall reliability. The ability to predict failures allows to repair the equipment on the actual condition, and not on the «calendar». It is not necessary to keep a large amount of equipment in stock, that is, to freeze working capital.


The use of DC in combination with big data and neural networks and the way from reporting and monitoring to predictive analysis and accident prevention systems


Build the most efficient operating regimes and minimize production costs. The longer the accumulation of data and the deeper the analytics, the more efficient optimization will be.

It is very important not to confuse the types of forecasting. Lately, working with the market of various IT solutions, I constantly see confusion in the concepts of predictive analytics and machine detection of anomalies in the operation of equipment. That is, using machine detection of deviations, they speak about the introduction of a new, predictive approach to the organization of service.

On the one hand, both neural networks actually work. When machine detection of anomalies of the neuronet also finds deviations, which allows to perform maintenance to a serious failure and replace only worn-out element.

However, let’s take a closer look at the definition of predictive analytics.

A predictive (or predictive, predictive) analysis is a prediction based on historical data.

So, it’s the ability to predict equipment failures before the abnormality happens. When the operational performance is still normal, but already begin to develop trends to deviation.

If you go to a very domestic level, the detection of anomalies – it is when you have a change of pressure and you are warned about it before you have a headache or begin to have heart problems. And predictive analytics is when things are still normal, but you have changed your diet, your sleep quality or something, respectively, the processes in your body that will subsequently lead to an increase in pressure.

As a result, the main difference is the depth of the dive, the availability of the skills and the horizon of prediction. Anomaly detection is a short-term prediction to avoid a crisis. To do this, you do not need to study historical data for a long period of time, for example, several years.

A full-fledged predictive analysis is a long-term prediction. You get more time to make decisions and work out measures: plan the purchase of new equipment or spare parts, call a repair team at a lower price or change the mode of operation of the equipment to prevent any deviations.

That’s what I think, but maybe there are alternative opinions, especially from marketers. The most important constraint I see at the moment is the complexity and cost of technology. Creating mathematical models is long and expensive, and the risk of error is high. It is necessary to combine technical knowledge about the object, practical experience, knowledge in modelling and visualization, observance of standards in real objects. Not all technical solutions are justified, as not every company has all competencies.

So, I think it’s useful for the industry to start with accident analysis, to identify the critical components of the assets and to model them. That is, to use an approach from the system constraint theory.

This will, first, minimize the risk of errors. Second, to enter this direction at a lower cost and to get an effect on which you can rely in the future. Third, accumulate expertise in working with data, making decisions based on them and «complicating» models. Having your own data competence is one of the key conditions for successful digitalization.

It is worth remembering that for now it is a new technology. Additionally, on the same cycle Gartner, it must pass the «valley of disappointment». Then later, when the digital competencies become more common and the neural networks become more massive, we’re going to use the digital counterparts to the full.

Бесплатный фрагмент закончился.

Бесплатно
488 ₽

Начислим

+15

Покупайте книги и получайте бонусы в Литрес, Читай-городе и Буквоеде.

Участвовать в бонусной программе
Возрастное ограничение:
18+
Дата выхода на Литрес:
19 июня 2024
Объем:
276 стр. 61 иллюстрация
ISBN:
9785006410169
Правообладатель:
Издательские решения
Формат скачивания:
Текст PDF
Средний рейтинг 0 на основе 0 оценок
Текст PDF
Средний рейтинг 0 на основе 0 оценок
Текст, доступен аудиоформат
Средний рейтинг 4,6 на основе 68 оценок
Подкаст
Средний рейтинг 0 на основе 0 оценок
Подкаст
Средний рейтинг 0 на основе 0 оценок
Подкаст
Средний рейтинг 0 на основе 0 оценок
Подкаст
Средний рейтинг 0 на основе 0 оценок
Текст
Средний рейтинг 0 на основе 0 оценок
По подписке
Текст
Средний рейтинг 0 на основе 0 оценок
Текст
Средний рейтинг 0 на основе 0 оценок
Текст
Средний рейтинг 0 на основе 0 оценок