Displaying 1 to 20 from 34 results

elasticsearch-dsl-py - High level Python client for Elasticsearch

  •    Python

Elasticsearch DSL is a high-level library whose aim is to help with writing and running queries against Elasticsearch. It is built on top of the official low-level client (elasticsearch-py).It provides a more convenient and idiomatic way to write and manipulate queries. It stays close to the Elasticsearch JSON DSL, mirroring its terminology and structure. It exposes the whole range of the DSL from Python either directly using defined classes or a queryset-like expressions.

elasticsearch-py - Official Python low-level client for Elasticsearch.

  •    Python

Official low-level client for Elasticsearch. Its goal is to provide common ground for all Elasticsearch-related code in Python; because of this it tries to be opinion-free and very extendable.For a more high level client library with more limited scope, have a look at elasticsearch-dsl - a more pythonic library sitting on top of elasticsearch-py.

tribler - Privacy enhanced BitTorrent client with P2P content discovery

  •    Python

Towards making Bittorrent anonymous and impossible to shut down. The aim of Tribler is giving anonymous access to online (streaming) videos. We are trying to make privacy, strong cryptography and authentication the Internet norm.




datasketch - MinHash, LSH, LSH Forest, Weighted MinHash, HyperLogLog, HyperLogLog++

  •    Python

datasketch gives you probabilistic data structures that can process and search very large amount of data super fast, with little loss of accuracy. datasketch must be used with Python 2.7 or above and NumPy 1.11 or above. Scipy is optional, but with it the LSH initialization can be much faster.

Memacs - What did I do on February 14th 2007? Visualize your (digital) life in Org-mode

  •    Python

Memacs extracts metadata (subjects, timestamps, contact information, …) from many different existing data sources (file names, emails, tweets, bookmarks, …) on your computer and generates files which are readable by GNU Emacs with Org-mode. Memacs - as the central component of the system - is a hub for all the connectors that add data from individual data sources. Those connectors are called Memacs modules or short module.

Squzer - Distributed Web Crawler

  •    Python

Squzer is the Declum's open-source, extensible, scale, multithreaded and quality web crawler project entirely written in the Python language.

scout - RESTful search server written in Python, powered by SQLite.

  •    Python

scout is a RESTful search server written in Python. The search is powered by SQLite's full-text search extension, and the web application utilizes the Flask framework.Scout can be installed from PyPI using pip or from source using git. Should you install from PyPI you will run the latest version, whereas installing from git ensures you have the latest changes.


sir - Transfer data from a MusicBrainz database to a Solr server

  •    Python

This repository contains the code necessary to index documents from a MusicBrainz database into a Solr search server. The documentation contains detailed information on how to use it.

mongodb-chemistry - Ideas for chemical similarity searches in MongoDB.

  •    Python

Chemical similarity search implementation in MongoDB, with performance analysis. See this blog post for more information.

splunk-webframework - Splunk Web Framework

  •    Python

The Splunk Web Framework lets developers quickly create custom Splunk apps by using prebuilt components, styles, templates, and reusable samples, and by adding custom logic, interactions, and UI. Applications developed with the Web Framework work seamlessly side by side with current Splunk applications. The Splunk Web Framework uses the Django web framework, the Splunk SDK for Python, and the Splunk SDK for JavaScript. The Web Framework also depends on a few JavaScript libraries for the client-side of code, such as Backbone.js for eventing, and jQuery for working with the document object model (DOM).

greptile - Fast grep implementation in python, with recursive search and replace

  •    Python

Because it exclusively uses generators, reptile never allocates big lists, it always processes one line in one file at a time. you can do a search on big files and large directories like / recursively without memory overhead.

algoliasearch-client-python - Algolia Search API Client for Python

  •    Python

Algolia Search is a hosted full-text, numerical, and faceted search engine capable of delivering realtime results from the first keystroke. The Algolia Search API Client for Python lets you easily use the Algolia Search REST API from your Python code.

search - A wrapper around Google's full text search API for App Engine

  •    Python

Thor is a wrapper for Google App Engine's search API that uses Django-like syntax for defining documents, and searching and filtering search indexes. From a basic standpoint, that's all there is to it. There is various filtering and ordering that can be applied to search queries, refer to the reference for the Index class for more in-depth example queries.

ffind - A sane replacement for find

  •    Python

ffind allows quick and easy recursive search for files in the command line. Very convenient to find a file you don't know exactly where it is or how it's called in a jungle of directories. If you have deal with Unix find, it replaces the cumbersome find . -name '*FILE_PATTERN*' with ffind FILE_PATTERN (plus more niceties).

Products.ZCatalog - Zope's indexing and search solution.

  •    Python

The ZCatalog is Zope's built in search engine. It allows you to categorize and search all kinds of Zope objects. It comes with a variety of indexes for different types of data.

docker_registry_cli - Docker Registry CLI - Provides search functionality for Docker Registry (UI and CLI)

  •    Python

Docker Registry CLI - Currently ONLY Supports the Search capability via Catalog API in the new version of Docker Registry v2.

KB-python-API - Python API for KB data-services

  •    Python

KB-Python-API is a simple API for Python, the API provides easy access to free and CC-BY-NC-ND datasets provided by the National Library of the Netherlands (KB). It relies on the back-end infrastructure of the KB which consists of an SRU and OAI-MPH service. The KB Python API makes it easy to interact with historical data, for more information on the provided datasets and data-rights take a look at the DataServices page of the KB.