From 6b38d320b00eee41316f8fc7be14ee7dd313dbb4 Mon Sep 17 00:00:00 2001 From: Gordon Shumway <39967334+oddluck@users.noreply.github.com> Date: Wed, 13 Mar 2019 02:50:05 -0400 Subject: [PATCH] use gsearch for searches --- WikiLeaf/plugin.py | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/WikiLeaf/plugin.py b/WikiLeaf/plugin.py index 732ac99..1e30479 100644 --- a/WikiLeaf/plugin.py +++ b/WikiLeaf/plugin.py @@ -15,6 +15,7 @@ import requests import re from bs4 import BeautifulSoup from fake_useragent import UserAgent +from gsearch.googlesearch import search try: from supybot.i18n import PluginInternationalization @@ -30,16 +31,14 @@ class WikiLeaf(callbacks.Plugin): def strain(self, irc, msg, args, strain): """ - Returns strain information from WikiLeaf. Searches powered by DuckDuckGo. + Returns strain information from WikiLeaf. Search powered by Google. """ - searchurl = "https://duckduckgo.com/html/?q={0} site: wikileaf.com/strain".format(strain) - ua = UserAgent() - header = {'User-Agent':str(ua.random)} try: - data = requests.get(searchurl, headers=header) - soup = BeautifulSoup(data.text) - url = re.sub('\s+', '', soup.find("a", class_="result__url").getText()) - data = requests.get("https://{0}".format(url), headers=header) + ua = UserAgent() + header = {'User-Agent':str(ua.random)} + data = search("{0} site:wikileaf.com/strain/".format(strain), num_results=1) + title, url = data[0] + data = requests.get(url, headers=header) soup = BeautifulSoup(data.text) name = re.sub('\s+', ' ', soup.find("h1", itemprop="name").getText()) straininfo = re.sub('\s+', ' ', soup.find("div", class_="product-info-line cannabis").getText()) @@ -53,3 +52,4 @@ class WikiLeaf(callbacks.Plugin): strain = wrap(strain, ['text']) Class = WikiLeaf +