add new URLs to robots.txt to prevent crawling
[debiancodesearch.git] / cmd / 
treee631ae06740b9ccbd931b2aaea2f0254b9e3dad0
drwxr-xr-x   ..
drwxr-xr-x - compute-ranking
drwxr-xr-x - dcs-batch-helper
drwxr-xr-x - dcs-compute-ranking
drwxr-xr-x - dcs-debmirror
drwxr-xr-x - dcs-feeder
drwxr-xr-x - dcs-index-backend
drwxr-xr-x - dcs-index
drwxr-xr-x - dcs-package-importer
drwxr-xr-x - dcs-reshard
drwxr-xr-x - dcs-rs-cleanup
drwxr-xr-x - dcs-rs-deploy
drwxr-xr-x - dcs-source-backend
drwxr-xr-x - dcs-tail-fedmsg
drwxr-xr-x - dcs-unpack
drwxr-xr-x - dcs-varz-to-influxdb
drwxr-xr-x - dcs-web
drwxr-xr-x - index-backend
drwxr-xr-x - source-backend