Gleam is a high performance and efficient distributed execution system, and also simple, generic, flexible and easy to customize.Gleam is built in Go, and the user defined computation can be written in Go, Unix pipe tools, or any streaming programs.
distributed-computing map-reduce distributed-systems distributed-system distributedIntroducing cross platform actor support between Go and C#.Can I use this? The Go implementation is still in beta, there are users using Proto Actor for Go in production already. But be aware that the API might change over time until 1.0.
grpc actor-model actors cross-platform protobuf akka clustering distributed-computingThis is a bare bones example of TensorFlow, a machine learning package published by Google. You will not find a simpler introduction to it. In each example, a straight line is fit to some data. Values for the slope and y-intercept of the line that best fit the data are determined using gradient descent. If you do not know about gradient descent, check out the Wikipedia page.
tensorflow tensorflow-tutorials distributed-computing simple big-data linear-regression tensorflow-examples tensorflow-exercisesSchematically, elephas works as follows. Elephas brings deep learning with Keras to Spark. Elephas intends to keep the simplicity and high usability of Keras, thereby allowing for fast prototyping of distributed models, which can be run on massive data sets. For an introductory example, see the following iPython notebook.
spark keras neural-networks deep-learning distributed-computingThe University of California holds the copyright on all BOINC source code. By submitting contributions to the BOINC code, you irrevocably assign all right, title, and interest, including copyright and all copyright rights, in such contributions to The Regents of the University of California, who may then use the code for any purpose that it desires. BOINC is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.
boinc distributed-computing volunteer-computing high-performance-computing citizen-science scientific-computing grid-computing science high-throughput-computing c-plus-plusThis project is experimental and the APIs are not considered stable. Fiber is a Python distributed computing library for modern computer clusters.
machine-learning sandbox multiprocessing distributed-computingAccelerated deep learning R&D
infrastructure machine-learning natural-language-processing information-retrieval research reinforcement-learning computer-vision deep-learning text-classification distributed-computing image-processing pytorch image-classification metric-learning recommender-system object-detection image-segmentation reproducibility text-segmentationUltra-fast, distributed, cross-platform actors.This is the .NET repository for Proto Actor.
actors clustering proto-actor akka distributed-systems distributed-computingA library for distributed computation. See documentation for more details.
pydata dask distributed-computingMooseFS is a Petabyte Open Source Network Distributed File System. It is easy to deploy and maintain, fault tolerant, highly performing, easily scalable, POSIX compliant. MooseFS Linux Client uses FUSE. MooseFS macOS Client uses FUSE for macOS.
dfs software-defined-storage posix filesystem file-system distributed-file-system clustering distributed-storage distributed-computing fuse big-data snapshot storage-tiering high-availability scalability storage moosefs hadoop posix-compliantRain is an open-source distributed computational framework for processing of large-scale task-based pipelines. Rain aims to lower the entry barrier to the world of distributed computing. Our intention is to provide a light yet robust distributed framework that features an intuitive Python API, straightforward installation and deployment with insightful monitoring on top.
pipelines distributed-computing python-api workflowsIt's a bit mind bending in that it's like using a higher-level list monad, but it's very, very cool. For beginning Haskellers, what would be really useful is a visualisation of what happens when you do various distributed/parallel stuff. It's almost shocking how effortlessly you can run computations across threads/nodes. The cool part is the composability in the distributed setting. You can make higher-order monadic functions that allow you to compose & reuse a long chain of distributed transactions via wormhole and teleport. Another benefit is that the transaction becomes first class and you can see exactly what's going on in one place instead of distributing the logic across actors making the code equivalent to event callbacks, as you've stated.
distributed-computing web events threading composition backtracking composability haskell transaction concurrencySmartSql = MyBatis + Cache(Memory | Redis) + R/W Splitting +Dynamic Repository + Diagnostics ...... SmartSql draws on MyBatis's ideas, uses XML to manage SQL, and provides several filter tags to eliminate various if/else judgment branches at the code level. SmartSql will manage your SQL and filter the tags to maintain your various conditional judgments at the code level to make your code more beautiful.
orm dotnet-core cross-platform high-performance distributed-computing redis zookeeper read-write-separation cache sql smartsql dotnetcore mybatis sqlserver mysql sqlite oracle postgresql skywalkingCouler aims to provide a unified interface for constructing and managing workflows on different workflow engines, such as Argo Workflows, Tekton Pipelines, and Apache Airflow. Couler is included in CNCF Cloud Native Landscape and LF AI Landscape.
kubernetes machine-learning workflow-engine scheduler distributed-computing cloud-native workflow-management unified-interface workflow-automation unified-api apache-airflow kubeflow argo-workflows tekton-pipelinesLizardFS Software Defined Storage is a distributed, parallel, scalable, fault-tolerant, Geo-Redundant and highly available file system. It spreads data over a number of physical servers, making it visible to an end user as a single file system.
distributed-filesystem filesystem nas distributed-systems distributed-computing distributed-storage fault-tolerance snapshot qos erasure-coding replication replicas geo-replication hsm hierarchical-storage hadoopJPPF enables applications with large processing power requirements to be run on any number of computers, in order to dramatically reduce their processing time. This is done by splitting an application into smaller parts that can be executed simultaneously on different machines.
parallel-processing distributed-computing clusterThe purpose of the future package is to provide a very simple and uniform way of evaluating R expressions asynchronously using various resources available to the user. In programming, a future is an abstraction for a value that may be available at some point in the future. The state of a future can either be unresolved or resolved. As soon as it is resolved, the value is available instantaneously. If the value is queried while the future is still unresolved, the current process is blocked until the future is resolved. It is possible to check whether a future is resolved or not without blocking. Exactly how and when futures are resolved depends on what strategy is used to evaluate them. For instance, a future can be resolved using a sequential strategy, which means it is resolved in the current R session. Other strategies may be to resolve futures asynchronously, for instance, by evaluating expressions in parallel on the current machine or concurrently on a compute cluster.
r cran parallel-processing parallel-computing distributed-computing hpc-clusters hpc promises futures asynchronous programming parallelizationThis repository contains the core Holochain libraries and binaries. This is the most recent and well maintained version of Holochain with a refactored state model (you may see references to it as Holochain RSM).
blockchain distributed-computing p2p holochain dapps-development holochain-rustParacel is a distributed computational framework, designed for many machine learning problems: Logistic Regression, SVD, Matrix Factorization(BFGS, sgd, als, cg), LDA, Lasso... Firstly, paracel splits both massive dataset and massive parameter space. Unlike Mapreduce-Like Systems, paracel offers a simple communication model, allowing you to work with a global and distributed key-value storage, which is called parameter server.
machine-learning distributed-computing graph c-plus-plusThis markdown file contains a list of academic papers (and other works) in the field of distributed consensus. Many of the papers listed below fit into more than one section. However, for simplicity, each paper is listed only in the most relevant section. Where possible, open access links for each paper have been provided. Contributions are welcome. This section lists theoretical results relating to distributed consensus.
distributed-systems networked-systems consensus-algorithm paxos atomic-broadcast-protocol zookeeper fault-tolerance distributed-computing
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.