Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Unverified Commit 4a127b19 authored by d-tux's avatar d-tux Committed by GitHub
Browse files

Merge branch 'master' into engines/unsplash

parents b9ada93b 442772a6
Loading
Loading
Loading
Loading
+1 −0
Original line number Diff line number Diff line
@@ -32,6 +32,7 @@ RUN echo "@commuedge http://nl.alpinelinux.org/alpine/edge/community" >> /etc/ap
    openssl-dev \
    ca-certificates \
    tini@commuedge \
 && pip install --upgrade pip \
 && pip install --no-cache -r requirements.txt \
 && apk del \
    build-base \
+27225 −1

File changed.

Preview size limit exceeded, changes collapsed.

searx/engines/findx.py

deleted100644 → 0
+0 −115
Original line number Diff line number Diff line
"""
FindX (General, Images, Videos)

@website     https://www.findx.com
@provide-api no
@using-api   no
@results     HTML
@stable      no
@parse       url, title, content, embedded, img_src, thumbnail_src
"""

from dateutil import parser
from json import loads
import re

from lxml import html

from searx import logger
from searx.engines.xpath import extract_text
from searx.engines.youtube_noapi import base_youtube_url, embedded_url
from searx.url_utils import urlencode


paging = True
results_xpath = '//script[@id="initial-state"]'
search_url = 'https://www.findx.com/{category}?{q}'
type_map = {
    'none': 'web',
    'general': 'web',
    'images': 'images',
    'videos': 'videos',
}


def request(query, params):
    params['url'] = search_url.format(
        category=type_map[params['category']],
        q=urlencode({
            'q': query,
            'page': params['pageno']
        })
    )
    return params


def response(resp):
    dom = html.fromstring(resp.text)
    results_raw_json = dom.xpath(results_xpath)
    results_json = loads(extract_text(results_raw_json))

    if len(results_json['web']['results']) > 0:
        return _general_results(results_json['web']['results']['webSearch']['results'])

    if len(results_json['images']['results']) > 0:
        return _images_results(results_json['images']['results'])

    if len(results_json['video']['results']) > 0:
        return _videos_results(results_json['video']['results'])

    return []


def _general_results(general_results):
    results = []
    for result in general_results:
        results.append({
            'url': result['url'],
            'title': result['title'],
            'content': result['sum'],
        })
    return results


def _images_results(image_results):
    results = []
    for result in image_results:
        results.append({
            'url': result['sourceURL'],
            'title': result['title'],
            'content': result['source'],
            'thumbnail_src': _extract_url(result['assets']['thumb']['url']),
            'img_src': _extract_url(result['assets']['file']['url']),
            'template': 'images.html',
        })
    return results


def _videos_results(video_results):
    results = []
    for result in video_results:
        if not result['kind'].startswith('youtube'):
            logger.warn('Unknown video kind in findx: {}'.format(result['kind']))
            continue

        description = result['snippet']['description']
        if len(description) > 300:
            description = description[:300] + '...'

        results.append({
            'url': base_youtube_url + result['id'],
            'title': result['snippet']['title'],
            'content': description,
            'thumbnail': _extract_url(result['snippet']['thumbnails']['default']['url']),
            'publishedDate': parser.parse(result['snippet']['publishedAt']),
            'embedded': embedded_url.format(videoid=result['id']),
            'template': 'videos.html',
        })
    return results


def _extract_url(url):
    matching = re.search('(/https?://[^)]+)', url)
    if matching:
        return matching.group(0)[1:]
    return ''
+13 −18
Original line number Diff line number Diff line
@@ -218,24 +218,6 @@ engines:
    shortcut : fd
    disabled : True

  - name : findx
    engine : findx
    shortcut : fx
    categories : general
    disabled : True

  - name : findx images
    engine : findx
    shortcut : fxi
    categories : images
    disabled : True

  - name : findx videos
    engine : findx
    shortcut : fxv
    categories : videos
    disabled : True

  - name : flickr
    categories : images
    shortcut : fl
@@ -728,6 +710,19 @@ engines:
    shortcut : du
    disabled : True

  - name : seznam
    shortcut: szn
    engine: xpath
    paging : True
    search_url : https://search.seznam.cz/?q={query}&count=10&from={pageno}
    results_xpath: //div[@class="Page-content"]//div[@class="Result "]
    url_xpath : ./h3/a/@href
    title_xpath : ./h3
    content_xpath : .//p[@class="Result-description"]
    first_page_num : 0
    page_size : 10
    disabled : True

#  - name : yacy
#    engine : yacy
#    shortcut : ya
+3 −1
Original line number Diff line number Diff line
@@ -27,12 +27,14 @@ def fetch_supported_languages():
        if hasattr(engines[engine_name], 'fetch_supported_languages'):
            try:
                engines_languages[engine_name] = engines[engine_name].fetch_supported_languages()
                if type(engines_languages[engine_name]) == list:
                    engines_languages[engine_name] = sorted(engines_languages[engine_name])
            except Exception as e:
                print(e)

    # write json file
    with io.open(engines_languages_file, "w", encoding="utf-8") as f:
        dump(engines_languages, f, ensure_ascii=False)
        dump(engines_languages, f, ensure_ascii=False, indent=4, separators=(',', ': '))

    return engines_languages