Generative AI models are trained by enormous datasets built from publicly available online data sourced by web crawling bots. However, websites can restrict these bots with exclusion protocols like Robots.txt, which has historically been used to help users navigate search engine results.

As generative AI continues to grow, organizations have been seeking to use exclusion protocols to intentionally wall off their data. How will this affect the future of AI training?

Read about the implications in an expert Q&A on @ieeespectrum

Author

Senior Member Coordinator and Past Section Treasurer. IEEE Senior Member and active volunteer for more than 20 years in the IEEE Puerto Rico and Caribbean Section.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.