This has stopped being a merely academic issue for me.
I was just in a kickoff meeting with my tech team at my college to explore what fun we're going to have integrating AI into our site search. (looks like it's gonna be Google).
We've had a prototype built so we can test what its returns look like.
- It returns some stuff with zero citations (even in debug mode), so we have no idea if it's just making stuff up.
- It ranks under-the-fold stuff over above-the-fold stuff (eg. it pulls from a weird sub-sub paragraph containing the keywords before pulling from the h1 title containing the keywords.)
- It pulls from documents, such as PDFs (which we've asked it not to), including documents that are, like, 5 years old.
Here's the real kicker: not only do we not have any ability to
change what or how it finds and returns references, but we don't even get to know
how it is deciding what's important. It is literally* a black box.
*
figuratively
Our only option is to rebuild our thousands of pages to be "data-centric". Whatever
that means.
Well, what it means is sacrifice as many trial-and-error chickens on the algorithm's altar as necessary, until it magically spits out the results we want.