Fashion provides functional purpose and captures mysterious and elusive aspects of being human. Fashion expresses and invokes human emotion and creativity. This industry intertwines how we look and sometimes even how we feel. Fashion has always been forward-looking, grabbing onto new technologies as they arise. Artificial intelligence is no exception, moving as quickly as fashion does.
Artificial intelligence (AI) is a field of computer science that looks at the logic behind human intelligence. The field seeks ways to understand how we think and to re-create this intelligence in machines. Because of its nature, AI extends across human activities, making it relevant in different ways to every industry.
The intersection of fashion and AI is a rich and expansive space that is just beginning to be explored. As AI continues to develop, it becomes harder to comprehend for nontechnical followers. The challenge of comprehension stands in the way of meaningful developments between these two fields. In this article, I will discuss this relationship at the same time as I try to visualize this relationship visually.
In “The State of Fashion 2018,” a report by McKinsey & Company and The Business of Fashion, 75% of retailers plan to invest in artificial intelligence over 2018 and 2019. It is changing the way the fashion industry does business across the entire fashion value chain. Providing customized experiences and better forecasting is just the start.
Currently, up to 30% of activities in 60% of occupations across all industries can be automated. Implementing some of this automation and reskilling the current workforce will still take time. At this rate, there is no question that artificial intelligence will significantly impact the way we work.
Artificial intelligence has become a confusing term. Machine learning, deep learning, and artificial intelligence are terms often used interchangeably, which may leave to question, what is the difference? Machine learning is a way of achieving AI. In 1959, it was defined by Arthur Samuel as “the ability to learn without being explicitly programmed.”
Usually this is done through “training.” Deep learning is an approach to machine learning, which usually involves large neural networks. Machine learning makes up a large portion of artificial intelligence being applied in businesses today. The goals of machine learning are to automate processes in order to decrease human effort, and to discover complex patterns that humans cannot interpret on their own.
This analogy is not perfect, but you can think of it this way: machine learning is to programming as the sewing machine is to sewing. Before the advent of the sewing machine, every stitch was sewn by hand. Once the sewing machine was introduced, sewing became faster, because not every stitch was handled by a human. With machine learning, we can build programs that handle far more complexity without having to hand code every detail.
Ultimately, however, seams can’t sew themselves, and machine learning continues to require a human hand to make it work. In machine learning, machines are used to identify patterns in data and frequently predict the values of nonexistent data, often correlating to events happening in the future. Machine learning encompasses many methods for learning from data and makes up a large portion of research happening in artificial intelligence today.
What Is Intelligence? Albert Einstein once said: “The true sign of intelligence is not knowledge but imagination”. While we intuitively know what intelligence is, it turns out to be difficult to summarize or formally define. There are many theories and definitions about what makes humans intelligent. How to measure intelligence has been argued by philosophers for centuries.
Shane Legg and Marcus Hutter collected over 70 experts’ definitions of intelligence in a paper called “A Collection of Definitions of Intelligence.” In an effort to derive a single definition, they came up with this: Intelligence measures an agent’s ability to achieve goals in a wide range of environments. In artificial intelligence, systems are often designed to mimic behaviors of the human mind.
Researchers look to the human mind as a model of intelligence. The original goal of reconstructing human intelligence in machines requires teaching machines to carry out many complex functions. Reasoning, problem solving, memory recall, planning, learning, processing natural language, perception, manipulation, social intelligence, and creativity are all part of reaching this goal.
The Turing Test How can we know if a machine is intelligent? The Turing test (TT) was proposed by Alan Turing in 1950 as one of the first tests of intelligence in machines. It is a challenge to understand whether a machine acts like a human. To pass the test, a human interrogator asks questions to the machine. If the human interrogator cannot distinguish which responses are from a human and which are from a machine, the machine passes the test.
The Turing test has appeared time and time again in popular science-fiction movies over the past 40 years. Ex Machina and Blade Runner are examples. It is one of many “Are we there yet?” checkpoints for the field.
Making mental connections is our most crucial learning tool, the essence of human intelligence; to forge links; to go beyond the given; to see patterns, relationships, context. —Marilyn Ferguson, author
Understanding human behavior is complicated because humans do not always act rationally or logically. We can improve a machine’s ability to predict human behavior by searching for patterns. These patterns help to discover and define trends. By analyzing these trends and modeling them with algorithms, machines can mimic human responses to certain inputs.
Then, when encountering these inputs in real-world contexts, they are able to respond accordingly. What Is Learning? If we could simplify human learning, we might say that humans take information from their environment, relate it to something, and then learn or act. These inputs could be something they see, smell, taste, hear, feel, or even their interpretation of a mood or tone.
That information is related to prior knowledge a person has about the world, making a connection. From there, a human might act on their new knowledge, explore, or innovate. Machines are given input in the form of data. The machine interprets that data and learns from it. Then the machine evaluates that data before outputting the information that has been defined as useful to a human to interpret.
Where does the data come from? Machines are collecting data through hardware inputs as well as software programs. You can think of the hardware as the body, and software as the mind of a machine. Hardware addresses the area of machine perception, and software addresses the idea of both machine language and human language.
Machines can perceive the environment through sight, feeling, and hearing via sensors. Sensors are a part of a machine’s hardware system. They measure physical events like temperature, pressure, force, acceleration, sound, and light.
In fact, your phone can measure almost all these things. Phones sense through small electronics called microelectromechanical systems (MEMS). The microphone, camera, inertial measurement units (or IMUs, which help track position), and proximity sensors are all examples of MEMs. These sensors can also be found in various Internet of Things (IoT) devices.
In collaboration with these sensors, software systems on a machine can do things like interpret when a phone is upside down or right side up, measure human locomotion, and detect faces or sounds.
Human languages are critical for communication. We use words and phrases, combining them in multiple ways in order to express ideas and emotions. Machines use machine languages to define models and parameters. Human language and machine perception both provide inputs in the form of data for machines to use to learn from.
An important distinction exists between machine languages and human languages. Machine languages are written in code. Originally, this code was only a series of 0s and 1s, or binary. Different combinations of 0s and 1s encode different information to machines. Over time, humans have created programming languages that interface between human language and machine language to make the job of coding easier.
The output of a machine is most useful when it can be interpreted by a human, which makes human language a useful concept for machines to understand. Topics in Artificial Intelligence The role of the computer is not to displace human creativity but rather to amplify it.
—Ray Kurzweil, The Age of Intelligent Machines
Successfully applying AI today requires understanding which techniques should be used for solving a given problem. There is currently no single algorithm that will provide value in every aspect of the fashion industry. The term AI as an overarching category can be confusing because it often leads people to believe that AI is a mysterious black box that can solve any problem. In reality, it is made up of several application areas, tools, and techniques.
Not every topic in the field of AI is covered in this book. Because the categories often overlap, this book is written with more information being introduced cumulatively as you read through. The phrasing “application areas” refers to the specific areas where machine learning tools and techniques can be applied. Natural language processing, computer vision, predictive analytics, and robotics might use some of the same techniques like neural networks to solve different types of problems. Those application areas can be further extended to industry applications.
Machine language and human language meet in natural language processing (NLP). NLP is a way for computers to comprehend human languages. Every day, our interactions on the Web—things we post on social media, text messages we write, and so forth—contribute to an ever-expanding mass of data. Of this, it is estimated that 80% of the 2.5 quintillion bytes of data created every day is unstructured data.
This is written in free form, unorganized and historically hard to parse. We can use NLP to understand the content and context of this unstructured data, unlocking a rich treasure trove of information about ourselves. Natural language processing is applied in multiple product categories including conversational shopping, AI customer service chatbots, and virtual assistants and stylists.
Computer vision (CV) is used to process and analyze images and videos. CV automates tasks we might associate with the human visual system and more. Although computer vision is a field of its own, artificial intelligence has played a major role in recent progress. Computer vision is frequently used in fashion applications because the fashion industry is so visual. In the fashion industry, computer vision is being used in technologies such as visual search, smart mirrors, social shopping, trend forecasting, virtual reality, and augmented reality.
AI can identify upcoming trends faster than industry insiders to enhance the design process.
—Avery Baker, chief brand officer, Tommy Hilfiger
Predictive analytics uses a variety of methods that use historical information to predict events that will happen in the future. These methods range in complexity and include data mining, basic statistics, and machine learning.
In this article, predictive analytics show up in two other areas: recommender systems and demand forecasting. Recommender systems are part of predictive analytics. They seek to understand user or customer behavior and recommend products or services that the user is likely to like or purchase. Recommender systems have played a critical role for discovering products in e-commerce.
You’ll find them everywhere, from fashion retail web sites to behind the scenes in subscription box services. You’ll also notice them in other areas including music and video streaming on sites like Netflix or YouTube. Demand forecasting is used to optimize supply-chain planning. By predicting demand for products, the fashion industry can reduce overproduction, thereby cutting costs and reducing waste.
Robotics, especially in apparel manufacturing, is a unique area of study that requires domain expertise across fashion, mechanical engineering, and machine learning. Robots have been used in industrial settings for many years in manufacturing for automotive, aerospace, and other industries that deal with mostly rigid parts.
Robotic manufacturing in the fashion industry is still a nascent field because of the complexities involved with handling fabrics. Nonetheless, with improvements in computer vision and in the planning algorithms needed to perform complex tasks, robotics is being adopted in fashion.
Tools and techniques like neural networks, generative adversarial networks, and data mining are used across application areas. These methods are constantly changing and evolving to return higher quality results in industry.
Neural networks are a subcategory of machine learning. They were originally modeled after our understanding of the behavior of neurons in the human brain: in the brain, a single neuron takes in input, processes it, and sends output. Neuroscience has moved away from this idea. We now know brains don’t actually work like this, and the statistics behind neural networks in machine learning have been developed independently of neuroscience.
Neural networks are typically created with layers that compute information in parallel. They’re composed of interconnected nodes. Knowledge in these systems is represented by the patterns that are taken on by nodes passing information to each other.
The way people think about the composition of a neural network usually includes three basic parts:
• Input layers: Contain input data
• Hidden layers: Contain the synapse architecture
• Output layers: Provide results from the network
Within that framework, a neural network can take on many architectures. Not all neural networks are the same. In the implementation, training is also an important part of the process. Training involves sending data through the neural network. In this stage, the network is learning complex connections between inputs and desired outputs. In many instances, the network’s effectiveness is reliant on high-quality data.
Neural networks are frequently used in the application areas discussed earlier in this chapter. Understanding the basic mechanisms of neural networks helps provide a foundation for understanding how contemporary artificial intelligence works.
Unsupervised learning can be inefficient because machines must learn by themselves. What is obvious to us may not be obvious to a machine. Generative adversarial networks (GANs) are one way to increase the efficiency of unsupervised learning. GANs use two neural networks: one network generates results, and the other evaluates the accuracy of those results.
GANs are a more recently adopted technology in the machine learning space and have been proposed by companies such as Amazon as a method for creating AI fashion designers in 2017. These and other generative models are especially promising for creating unique new images as well as for filling in information from images that are incomplete or damaged.
Data is critical to any task in machine learning. Without data, the machine has nothing to train from. Data can include information such as video, images, and text. Data collection refers to the process of collecting data for analysis.
In many cases, data collection is just the beginning. What do you do with all the data? Data mining is about uncovering useful information in large amounts of data. For the fashion industry, social media can be a treasure trove for learning about the way customers feel about products and trends.
Different application areas, techniques, and tools have strengths and weaknesses at specific tasks. Application areas of AI are often targeted at addressing specific needs (for example, image- vs. language-related problems are addressed using computer vision and natural language processing respectively). Successfully applying AI today requires understanding which techniques and tools make sense for your application.
Computer vision is an inherently visual field of artificial intelligence and is applied in instances dealing with images and video. Natural language processing, on the other hand, deals with communicating between human languages and machine languages.
Artificial intelligence, and especially machine learning methods, use data and models to understand and make predictions about questions we don’t have the answers to.
If you’re not interested, you’re not interesting.
—Iris Apfel, interior designer, fashion icon
Natural language processing (NLP) plays a critical role in human-machine communication. Billions of gigabytes of data are being created by users around the world every day. Most of this content is created in unstructured formats, making it unusable using regular programming techniques. With NLP, this unstructured data can be interpreted by machines without the requirements of strict data structures.
In the fashion industry, natural language processing has been used in applications like conversational commerce, chatbots, AI-based stylists, image and trend classification, and micro-moment shopping.
Finding garments you like on the Internet can be really hard, requiring sifting through tens, hundreds, even thousands of listings. One of the most pervasive concepts to gain traction in fashion retail is conversational commerce, which took hold in late 2017. By integrating product information into a chat interface, brands are able to reduce friction during product discovery and provide highly personalized experiences to consumers searching for products, information, and customer service.
Natural language processing has been studied by computer scientists since the 1950s. Computer scientist Alan Turing thought the ability to use human language to be an important determinant of intelligence in machines. He later created the Turing test as a measure of machine intelligence.
A machine passes the Turing test if it can fool people into believing it is a human. At that time and throughout the 1960s, the first chatterbots were created, exemplifying the power of natural language–based interfaces.
What makes you think I am entitled to my own opinion?
—ELIZA, chatbot
One of the most famous examples during this time was ELIZA. ELIZA was one of the first programs to pass a restricted version of the Turing test. Simulating a Rogerian psychotherapist, the program processed user inputs, saving them in memory and recalling them during conversation.
