How can I programmatically provide site structure and a URL path to Google search?
You can provide the site structure and URL path to Google Search by using Sitemap and Robots.txt files.
A Sitemap is an XML file that lists the pages of your website and provides information about each page, such as when it was last updated, how frequently it changes, and its relative importance to other pages on your site. You can submit your sitemap to Google Search Console to help Google discover and crawl your pages.
A Robots.txt file is a simple text file that tells web robots (including Googlebots) which pages or sections of your site to crawl and which pages to ignore.
By using these two files, you can provide Google with the information it needs to better understand the structure of your site and improve the visibility of your pages in search results.