require 'set' module Moft class Site attr_accessor :config, :layouts, :posts, :pages, :static_files, :categories, :exclude, :include, :static_includes, :source, :dest, :lsi, :pygments, :permalink_style, :tags, :time, :future, :safe, :plugins, :limit_posts, :show_drafts, :keep_files, :baseurl attr_accessor :converters, :generators # Public: Initialize a new Site. # # config - A Hash containing site configuration details. def initialize(config) self.config = config.clone self.safe = config['safe'] self.source = File.expand_path(config['source']) self.dest = File.expand_path(config['destination']) self.plugins = plugins_path self.lsi = config['lsi'] self.pygments = config['pygments'] self.baseurl = config['baseurl'] self.permalink_style = config['permalink'].to_sym self.exclude = config['exclude'] self.include = config['include'] self.future = config['future'] self.show_drafts = config['show_drafts'] self.limit_posts = config['limit_posts'] self.keep_files = config['keep_files'] self.static_includes = config['static_includes'] self.reset self.setup end # Public: Read, process, and write this Site to output. # # Returns nothing. def process self.reset self.read self.generate self.render self.cleanup self.write end # Reset Site details. # # Returns nothing def reset self.time = if self.config['time'] Time.parse(self.config['time'].to_s) else Time.now end self.layouts = {} self.posts = [] self.pages = [] self.static_files = [] self.categories = Hash.new { |hash, key| hash[key] = [] } self.tags = Hash.new { |hash, key| hash[key] = [] } if self.limit_posts < 0 raise ArgumentError, "limit_posts must not be a negative number" end end # Load necessary libraries, plugins, converters, and generators. # # Returns nothing. def setup # require 'classifier' if self.lsi # Check that the destination dir isn't the source dir or a directory # parent to the source dir. if self.source =~ /^#{self.dest}/ raise FatalException.new "Destination directory cannot be or contain the Source directory." end # If safe mode is off, load in any Ruby files under the plugins # directory. unless self.safe self.plugins.each do |plugins| Dir[File.join(plugins, "**/*.rb")].each do |f| require f end end end self.converters = instantiate_subclasses(Moft::Converter) self.generators = instantiate_subclasses(Moft::Generator) end # Internal: Setup the plugin search path # # Returns an Array of plugin search paths def plugins_path if (config['plugins'] == Moft::Configuration::DEFAULTS['plugins']) [File.join(self.source, config['plugins'])] else Array(config['plugins']).map { |d| File.expand_path(d) } end end # Read Site data from disk and load it into internal data structures. # # Returns nothing. def read self.read_static_includes self.read_layouts self.read_directories end def read_static_includes self.static_includes.each do |s| self.read_directories(s) end end # Read all the files in / and create a new Layout object # with each one. # # Returns nothing. def read_layouts base = File.join(self.source, self.config['layouts']) return unless File.exists?(base) entries = [] Dir.chdir(base) { entries = filter_entries(Dir['*.*']) } entries.each do |f| name = f.split(".")[0..-2].join(".") self.layouts[name] = Layout.new(self, base, f) end end # Recursively traverse directories to find posts, pages and static files # that will become part of the site according to the rules in # filter_entries. # # dir - The String relative path of the directory to read. Default: ''. # # Returns nothing. def read_directories(dir = '') base = File.join(self.source, dir) entries = Dir.chdir(base) { filter_entries(Dir.entries('.')) } self.read_posts(dir) if self.show_drafts self.read_drafts(dir) end self.posts.sort! # limit the posts if :limit_posts option is set if limit_posts > 0 limit = self.posts.length < limit_posts ? self.posts.length : limit_posts self.posts = self.posts[-limit, limit] end entries.each do |f| f_abs = File.join(base, f) f_rel = File.join(dir, f) if File.directory?(f_abs) next if self.dest.sub(/\/$/, '') == f_abs read_directories(f_rel) else first3 = File.open(f_abs) { |fd| fd.read(3) } if first3 == "---" # file appears to have a YAML header so process it as a page pages << Page.new(self, self.source, dir, f) else # otherwise treat it as a static file static_files << StaticFile.new(self, self.source, dir, f) end end end end # Read all the files in //_posts and create a new Post # object with each one. # # dir - The String relative path of the directory to read. # # Returns nothing. def read_posts(dir) entries = get_entries(dir, '_posts') # first pass processes, but does not yet render post content entries.each do |f| if Post.valid?(f) post = Post.new(self, self.source, dir, f) if post.published && (self.future || post.date <= self.time) aggregate_post_info(post) end end end end # Read all the files in //_drafts and create a new Post # object with each one. # # dir - The String relative path of the directory to read. # # Returns nothing. def read_drafts(dir) entries = get_entries(dir, '_drafts') # first pass processes, but does not yet render draft content entries.each do |f| if Draft.valid?(f) draft = Draft.new(self, self.source, dir, f) aggregate_post_info(draft) end end end # Run each of the Generators. # # Returns nothing. def generate self.generators.each do |generator| generator.generate(self) end end # Render the site to the destination. # # Returns nothing. def render payload = site_payload self.posts.each do |post| post.render(self.layouts, payload) end self.pages.each do |page| page.render(self.layouts, payload) end self.categories.values.map { |ps| ps.sort! { |a, b| b <=> a } } self.tags.values.map { |ps| ps.sort! { |a, b| b <=> a } } rescue Errno::ENOENT => e # ignore missing layout dir end # Remove orphaned files and empty directories in destination. # # Returns nothing. def cleanup # all files and directories in destination, including hidden ones dest_files = Set.new Dir.glob(File.join(self.dest, "**", "*"), File::FNM_DOTMATCH) do |file| if self.keep_files.length > 0 dest_files << file unless file =~ /\/\.{1,2}$/ || file =~ keep_file_regex else dest_files << file unless file =~ /\/\.{1,2}$/ end end # files to be written files = Set.new self.posts.each do |post| files << post.destination(self.dest) end self.pages.each do |page| files << page.destination(self.dest) end self.static_files.each do |sf| files << sf.destination(self.dest) end # adding files' parent directories dirs = Set.new files.each { |file| dirs << File.dirname(file) } files.merge(dirs) obsolete_files = dest_files - files FileUtils.rm_rf(obsolete_files.to_a) end # Private: creates a regular expression from the keep_files array # # Examples # ['.git','.svn'] creates the following regex: /\/(\.git|\/.svn)/ # # Returns the regular expression def keep_file_regex or_list = self.keep_files.join("|") pattern = "\/(#{or_list.gsub(".", "\.")})" Regexp.new pattern end # Write static files, pages, and posts. # # Returns nothing. def write self.posts.each do |post| post.write(self.dest) end self.pages.each do |page| page.write(self.dest) end self.static_files.each do |sf| sf.write(self.dest) end end # Construct a Hash of Posts indexed by the specified Post attribute. # # post_attr - The String name of the Post attribute. # # Examples # # post_attr_hash('categories') # # => { 'tech' => [, ], # # 'ruby' => [] } # # Returns the Hash: { attr => posts } where # attr - One of the values for the requested attribute. # posts - The Array of Posts with the given attr value. def post_attr_hash(post_attr) # Build a hash map based on the specified post attribute ( post attr => # array of posts ) then sort each array in reverse order. hash = Hash.new { |hsh, key| hsh[key] = Array.new } self.posts.each { |p| p.send(post_attr.to_sym).each { |t| hash[t] << p } } hash.values.map { |sortme| sortme.sort! { |a, b| b <=> a } } hash end # The Hash payload containing site-wide data. # # Returns the Hash: { "site" => data } where data is a Hash with keys: # "time" - The Time as specified in the configuration or the # current time if none was specified. # "posts" - The Array of Posts, sorted chronologically by post date # and then title. # "pages" - The Array of all Pages. # "html_pages" - The Array of HTML Pages. # "categories" - The Hash of category values and Posts. # See Site#post_attr_hash for type info. # "tags" - The Hash of tag values and Posts. # See Site#post_attr_hash for type info. def site_payload {"site" => self.config.merge({ "time" => self.time, "posts" => self.posts.sort { |a, b| b <=> a }, "pages" => self.pages, "html_pages" => self.pages.reject { |page| !page.html? }, "categories" => post_attr_hash('categories'), "tags" => post_attr_hash('tags')})} end # Filter out any files/directories that are hidden or backup files (start # with "." or "#" or end with "~"), or contain site content (start with "_"), # or are excluded in the site configuration, unless they are web server # files such as '.htaccess'. # # entries - The Array of String file/directory entries to filter. # # Returns the Array of filtered entries. def filter_entries(entries) entries.reject do |e| unless self.include.glob_include?(e) ['.','_', '#'].include?(e[0..0]) || e[-1..-1] == '~' || self.exclude.glob_include?(e) || (File.symlink?(e) && self.safe) end end end # Get the implementation class for the given Converter. # # klass - The Class of the Converter to fetch. # # Returns the Converter instance implementing the given Converter. def getConverterImpl(klass) matches = self.converters.select { |c| c.class == klass } if impl = matches.first impl else raise "Converter implementation not found for #{klass}" end end # Create array of instances of the subclasses of the class or module # passed in as argument. # # klass - class or module containing the subclasses which should be # instantiated # # Returns array of instances of subclasses of parameter def instantiate_subclasses(klass) klass.subclasses.select do |c| !self.safe || c.safe end.sort.map do |c| c.new(self.config) end end # Read the entries from a particular directory for processing # # dir - The String relative path of the directory to read # subfolder - The String directory to read # # Returns the list of entries to process def get_entries(dir, subfolder) base = File.join(self.source, dir, subfolder) return [] unless File.exists?(base) entries = Dir.chdir(base) { filter_entries(Dir['**/*']) } entries.delete_if { |e| File.directory?(File.join(base, e)) } end # Aggregate post information # # post - The Post object to aggregate information for # # Returns nothing def aggregate_post_info(post) self.posts << post post.categories.each { |c| self.categories[c] << post } post.tags.each { |c| self.tags[c] << post } end end end