Unpacking black-box models | MIT News

Modern machine-learning models, such as neural networks, are often referred to as “black boxes” because they are so complex that even the researchers who design them can’t fully understand how they make predictions. To provide some insights, researchers use explanation methods that seek to describe individual model decisions. For example, they may highlight words in …

Unpacking black-box models | MIT News Read More »