Godwin
9 years ago
10 changed files with 20 additions and 46 deletions
@ -1,21 +0,0 @@ |
|||||
SitemapGenerator::Sitemap.default_host = "http://hackingoff.com" |
|
||||
|
|
||||
SitemapGenerator::Sitemap.create do |
|
||||
routes = Rails.application.routes.routes.map do |route| |
|
||||
{alias: route.name, path: route.path.spec.to_s, controller: route.defaults[:controller], action: route.defaults[:action]} |
|
||||
end |
|
||||
|
|
||||
# Set a list of controllers you don't want to generate routes for. |
|
||||
# /rails/info in particular maps to something inaccessible. |
|
||||
# redirects have a nil controller. This prevents duplicate content penalties. |
|
||||
banned_controllers = ["rails/info", nil] |
|
||||
routes.reject! {|route| banned_controllers.include?(route[:controller])} |
|
||||
|
|
||||
# sitemap_generator includes root by default; prevent duplication |
|
||||
routes.reject! {|route| route[:path] == '/'} |
|
||||
|
|
||||
routes.each {|route| add route[:path][0..-11]} # Strips off '(.:format) |
|
||||
|
|
||||
# Notice the below if you're hosting Jekyll/Octopress in a subdirectory |
|
||||
# or otherwise want to index content outside of Rails' routes. |
|
||||
# add_to_index '/path/sitemap.xml' |
|
@ -1 +0,0 @@ |
|||||
Delayed::Worker.logger = Logger.new(File.join(Rails.root, 'log', 'delayed_job.log')) |
|
@ -1,5 +0,0 @@ |
|||||
# See http://www.robotstxt.org/wc/norobots.html for documentation on how to use the robots.txt file |
|
||||
# |
|
||||
# To ban all spiders from the entire site uncomment the next two lines: |
|
||||
# User-agent: * |
|
||||
# Disallow: / |
|
Loading…
Reference in new issue