X-Git-Url: https://git.netwichtig.de/gitweb/?a=blobdiff_plain;ds=inline;f=data%2Frbot%2Fplugins%2Fsearch.rb;h=27f9519c37b4d5e3bec5bbda080380526371bed9;hb=6fe052181de09dd4a618241e600e2473f6706343;hp=c44100b1abaa106ba5367f683223b599200906ff;hpb=4911de2a2b6bc373ef7d5e73f34a14133c07b5e5;p=user%2Fhenk%2Fcode%2Fruby%2Frbot.git diff --git a/data/rbot/plugins/search.rb b/data/rbot/plugins/search.rb index c44100b1..27f9519c 100644 --- a/data/rbot/plugins/search.rb +++ b/data/rbot/plugins/search.rb @@ -4,34 +4,83 @@ Net::HTTP.version_1_2 GOOGLE_WAP_LINK = /(.*?)<\/a>/im +class ::String + def ircify_html + txt = self + + # bold and strong -> bold + txt.gsub!(/<\/?(?:b|strong)\s*>/, "#{Bold}") + + # italic, emphasis and underline -> underline + txt.gsub!(/<\/?(?:i|em|u)\s*>/, "#{Underline}") + + ## This would be a nice addition, but the results are horrible + ## Maybe make it configurable? + # txt.gsub!(/<\/?a( [^>]*)?>/, "#{Reverse}") + + # Paragraph and br tags are converted to whitespace. + txt.gsub!(/<\/?(p|br)\s*\/?\s*>/, ' ') + txt.gsub!("\n", ' ') + + # All other tags are just removed + txt.gsub!(/<[^>]+>/, '') + + # Remove double formatting options, since they only waste bytes + txt.gsub!(/#{Bold}\s*#{Bold}/,"") + txt.gsub!(/#{Underline}\s*#{Underline}/,"") + + # And finally whitespace is squeezed + txt.gsub!(/\s+/, ' ') + + # Decode entities and strip whitespace + return Utils.decode_html_entities(txt).strip! + end +end + class SearchPlugin < Plugin + BotConfig.register BotConfigIntegerValue.new('google.hits', + :default => 3, + :desc => "Number of hits to return from Google searches") + BotConfig.register BotConfigIntegerValue.new('google.first_par', + :default => 0, + :desc => "When set to n > 0, the bot will return the first paragraph from the first n search hits") + BotConfig.register BotConfigIntegerValue.new('wikipedia.hits', + :default => 3, + :desc => "Number of hits to return from Wikipedia searches") + BotConfig.register BotConfigIntegerValue.new('wikipedia.first_par', + :default => 1, + :desc => "When set to n > 0, the bot will return the first paragraph from the first n wikipedia search hits") + def help(plugin, topic="") case topic - when "search" - "search => search google for " - when "google" - "google => search google for " + when "search", "google" + "#{topic} => search google for " when "wp" "wp [] => search for on Wikipedia. You can select a national to only search the national Wikipedia" else - "search (or: google ) => search google for | wp => search for on Wikipedia" + "search (or: google ) => search google for | wp => search for on Wikipedia" end end def google(m, params) what = params[:words].to_s searchfor = URI.escape what + # This method is also called by other methods to restrict searching to some sites if params[:site] site = "site:#{params[:site]}+" else site = "" end + # It is also possible to choose a filter to remove constant parts from the titles + # e.g.: "Wikipedia, the free encyclopedia" when doing Wikipedia searches + filter = params[:filter] || "" url = "http://www.google.com/wml/search?q=#{site}#{searchfor}" + hits = params[:hits] || @bot.config['google.hits'] begin - wml = @bot.httputil.get(url) + wml = @bot.httputil.get_cached(url) rescue => e m.reply "error googling for #{what}" return @@ -41,17 +90,74 @@ class SearchPlugin < Plugin m.reply "no results found for #{what}" return end - results = results[0...3].map { |res| - "#{res[0]}. #{Bold}#{Utils.decode_html_entities res[2].strip}#{Bold}: #{URI.unescape res[1].strip}" + urls = Array.new + results = results[0...hits].map { |res| + n = res[0] + t = Utils.decode_html_entities res[2].gsub(filter, '').strip + u = URI.unescape res[1] + urls.push(u) + "#{n}. #{Bold}#{t}#{Bold}: #{u}" }.join(" | ") - m.reply "Results for #{what}: #{results}" + m.reply "Results for #{what}: #{results}", :split_at => /\s+\|\s+/ + + first_pars = params[:firstpar] || @bot.config['google.first_par'] + + idx = 0 + while first_pars > 0 and urls.length > 0 + url.replace(urls.shift) + idx += 1 + + # FIXME what happens if some big file is returned? We should share + # code with the url plugin to only retrieve partial file content! + xml = @bot.httputil.get_cached(url) + if xml.nil? + debug "Unable to retrieve #{url}" + next + end + # We get the first par after the first main heading, if possible + header_found = xml.match(/]*)?>(.*?)<\/h1>/im) + txt = String.new + if header_found + debug "Found header: #{header_found[1].inspect}" + while txt.empty? + header_found = $' + candidate = header_found[/]*)?>.*?<\/p>/im] + break unless candidate + txt.replace candidate.ircify_html + end + end + # If we haven't found a first par yet, try to get it from the whole + # document + if txt.empty? + header_found = xml + while txt.empty? + candidate = header_found[/]*)?>.*?<\/p>/im] + break unless candidate + txt.replace candidate.ircify_html + header_found = $' + end + end + # Nothing yet, try title + if txt.empty? + debug "No first par found\n#{xml}" + # FIXME only do this if the 'url' plugin is loaded + txt.replace @bot.plugins['url'].get_title_from_html(xml) + next if txt.empty? + end + m.reply "[#{idx}] #{txt}", :overlong => :truncate + first_pars -=1 + end end def wikipedia(m, params) lang = params[:lang] site = "#{lang.nil? ? '' : lang + '.'}wikipedia.org" + debug "Looking up things on #{site}" params[:site] = site + params[:filter] = / - Wikipedia.*$/ + params[:hits] = @bot.config['wikipedia.hits'] + params[:firstpar] = @bot.config['wikipedia.first_par'] return google(m, params) end end