node-filestorage - Storage for storing uploaded files in node.js

  •        21

Directory contains max 1000 files. Each file has .data extension and each file contains internal META information (2 kB). Each directory contains the config file with the informations about all files. The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

https://github.com/petersirka/node-filestorage

Tags
Implementation
License
Platform

   




Related Projects

httpinvoke - A no-dependencies HTTP client library for browsers and Node

  •    Javascript

A no-dependencies HTTP client library for browsers and Node.js with a promise-based or Node.js-style callback-based API to progress events, text and binary file upload and download, partial response body, request and response headers, status code. Any one, two or three arguments can be skipped, except the url.

skipper - Streaming multi-uploads for Sails/Express - supports disk, S3, gridfs, and custom file adapters

  •    Javascript

Skipper makes it easy to implement streaming file uploads to disk, S3, or any supported file upload adapters.The following example assumes skipper is already installed as the body parser in your Express or Sails app. It receives one or more files from a file parameter named avatar using the default, built-in file adapter (skipper-disk). This streams the file(s) to the default upload directory .tmp/uploads/ on the server's local disk.

NeatUpload

  •    ASPNET

The NeatUpload ™ ASP.NET component allows developers to stream uploaded files to storage (filesystem or database) and allows users to monitor upload progress. It is open source and works under Mono's XSP/mod_mono as well as Microsoft's ASP.NET implementation.

PGStorage

  •    Python

PGStorage is a Zope / ZODB backend that persists to a PostgreSQL database. It can take the place of FileStorage, ZEO, and ZRS.

Meteor-Files - Upload files via DDP or HTTP to Meteor server FS, AWS, GridFS, DropBox or Google Drive

  •    Javascript

Award winning, extremely fast and robust package for file uploading, managing and streaming (Audio & Video & Images), with support of server's file system (FS) or third party storage, like: AWS, DropBox, Google Storage, Google Drive, GridFS or any other with API. All PRs are always welcome on dev branch. Please, always give expressive description to your changes and additions.


tus-node-server - Node.js tus server

  •    Javascript

tus is a new open protocol for resumable uploads built on HTTP. This is the tus protocol 1.0.0 node.js server implementation. โš ๏ธ Attention: We currently lack the resources to properly maintain tus-node-server. This has the unfortunate consequence that this project is in rather bad condition (out-dated dependencies, no tests for the S3 storage, no resumable uploads for the GCS storage etc). If you want to help us with tus-node-server, we are more than happy to assist you and welcome new contributors. In the meantime, we can recommend tusd as a reliable and production-tested tus server. Of course, you can use tus-node-server if it serves your purpose.

s3-upload-stream - A Node.js module for streaming data to Amazon S3 via the multipart upload API

  •    Javascript

A pipeable write stream which uploads to Amazon S3 using the multipart file upload API. NOTE: This module is deprecated after the 2.1.0 release of the AWS SDK on Dec 9, 2014, which added S3.upload(). I highly recommend switching away from this module and using the official method supported by AWS.

Azure Blob Pusher

  •    

File uploader allows bulk upload of files from client to Azure blob storage. FileSystemWatcher notices new files in configured directory. SQL Express database tracks files as they are pushed up to blob storage and then processed. Local Directory is cleaned up after acknowle...

megacmd - A command-line client for mega.co.nz storage service

  •    Go

I'm (@ncw) doing light maintenance on megacmd, but mainly on the underlying go-mega library which I'm using to provide a mega backend for rclone. Mega (mega.co.nz) is an excellent free storage service which provides 50 GB of free storage space. It has a web based user interface to upload and download files. Megacmd is a command-line tool for performing file and directory transfer between local directories and mega service. Features of megacmd are much similar to s3cmd utility, which is used to perform file transfer to Amazon S3.

Bebop - Simple discussion board / forum web application.

  •    Go

Bebop is a simple discussion board / forum web application.Create a new empty database (MySQL ะพr PostgreSQL) that will be used as a data store and a database user with all privileges granted on this database.

s3-plugin-webpack - Uploads files to s3 after complete

  •    Javascript

I notice a lot of people are setting the directory option when the files are part of their build. Please don't set directory if you're uploading your build. Using the directory option reads the files after compilation to upload instead of from the build process. include and exclude rules behave similarly to Webpack's loader options. In addition to a RegExp you can pass a function which will be called with the path as its first argument. Returning a truthy value will match the rule. You can also pass an Array of rules, all of which must pass for the file to be included or excluded.

ZipStream-PHP - Fork of pablotron's zip streaming library.

  •    PHP

You can also add comments, modify file timestamps, and customize (or disable) the HTTP headers. It is also possible to specify the storage method when adding files, the current default storage method is 'deflate' i.e files are stored with Compression mode 0x08. See the class file for details. Please take a look at the CONTRIBUTOR-README.md File.

casync - Content-Addressable Data Synchronization Tool

  •    C

Encoding: Let's take a large linear data stream, split it into variable-sized chunks (the size of each being a function of the chunk's contents), and store these chunks in individual, compressed files in some directory, each file named after a strong hash value of its contents, so that the hash value may be used to as key for retrieving the full chunk data. Let's call this directory a "chunk store". At the same time, generate a "chunk index" file that lists these chunk hash values plus their respective chunk sizes in a simple linear array. The chunking algorithm is supposed to create variable, but similarly sized chunks from the data stream, and do so in a way that the same data results in the same chunks even if placed at varying offsets. For more information see this blog story. Decoding: Let's take the chunk index file, and reassemble the large linear data stream by concatenating the uncompressed chunks retrieved from the chunk store, keyed by the listed chunk hash values.

stream-handbook - how to write node programs with streams

  •    Javascript

Now you will have a stream-handbook command that will open this readme file in your $PAGER. Otherwise, you may continue reading this document as you are presently doing.Streams come to us from the earliest days of unix and have proven themselves over the decades as a dependable way to compose large systems out of small components that do one thing well. In unix, streams are implemented by the shell with | pipes. In node, the built-in stream module is used by the core libraries and can also be used by user-space modules. Similar to unix, the node stream module's primary composition operator is called .pipe() and you get a backpressure mechanism for free to throttle writes for slow consumers.

angular-express-seed - A great starting point for writing AngularJS apps backed by an Express-powered node

  •    Javascript

Start an awesome app with AngularJS on the front, Express + Node on the back. This project is an application skeleton for a typical AngularJS web app for those who want to use Node to serve their app.The seed contains angular libraries, test libraries and a bunch of scripts all preconfigured for instant web development gratification. Just clone the repo (or download the zip/tarball) and you're ready to develop your application.

arc - :paperclip: Flexible file upload and attachment library for Elixir

  •    Elixir

Arc is a flexible file upload library for Elixir with straightforward integrations for Amazon S3 and ImageMagick. Browse the readme below, or jump to a full example.

frisbee - :dog2: Modern fetch-based alternative to axios/superagent/request

  •    Javascript

tldr; Stripe-inspired API wrapper for WHATWG's fetch() method for making simple HTTP requests (alternative to superagent, request, axios).If you're using node-fetch, you need node-fetch@v1.5.3 to use form-data with files properly (due to https://github.com/bitinn/node-fetch/issues/102) If you experience form file upload issues, please see https://github.com/facebook/react-native/issues/7564#issuecomment-266323928.

node-express-mongoose-demo - A simple demo app using express, mongoose, passport for beginners

  •    Javascript

This is a demo node.js application illustrating various features used in everyday web development, with a fine touch of best practices. The demo app is a blog application where users (signing up using facebook, twitter, github and simple registrations) can create an article, delete an article and add comments on the article. and replace the values there. In production env, it is not safe to keep the ids and secrets in a file, so you need to set it up via commandline. If you are using heroku checkout how environment variables are set here.

crate - ๐Ÿ‘•๐Ÿ‘–๐Ÿ“ฆ A sample web and mobile application built with Node, Express, React, React Native, Redux and GraphQL

  •    Javascript

๐Ÿ‘•๐Ÿ‘–๐Ÿ“ฆ A sample web and mobile application built with Node, Express, React, React Native, Redux and GraphQL. Very basic replica of stitchfix.com / krate.in (allows users to get monthly subscription of trendy clothes and accessories).