Widely used large language models often function as “black boxes,” making their outputs difficult to predict or control and limiting their safe deployment in some sensitive domains, like healthcare, education and public services. A Cornell research team is working to allow users to better guide AI outputs.