summaryrefslogtreecommitdiff
path: root/node_modules/asar/README.md
diff options
context:
space:
mode:
authorLinuxWizard42 <computerwizard@linuxmail.org>2022-10-12 23:08:57 +0300
committerLinuxWizard42 <computerwizard@linuxmail.org>2022-10-12 23:08:57 +0300
commit726b81b19251674e149ccfbb1abacbd837fc6db0 (patch)
treefbdbb227dc01357eb76e8222d76185bc124c5ca6 /node_modules/asar/README.md
parent34f0890e175698940d49238097579f44e4d78c89 (diff)
downloadFlashRunner-726b81b19251674e149ccfbb1abacbd837fc6db0.tar.gz
FlashRunner-726b81b19251674e149ccfbb1abacbd837fc6db0.tar.zst
Removed files that should not have been included in git
Diffstat (limited to 'node_modules/asar/README.md')
-rw-r--r--node_modules/asar/README.md215
1 files changed, 0 insertions, 215 deletions
diff --git a/node_modules/asar/README.md b/node_modules/asar/README.md
deleted file mode 100644
index b9f2568..0000000
--- a/node_modules/asar/README.md
+++ /dev/null
@@ -1,215 +0,0 @@
-# asar - Electron Archive
-
-[![CircleCI build status](https://circleci.com/gh/electron/asar/tree/main.svg?style=shield)](https://circleci.com/gh/electron/asar/tree/main)
-[![npm version](http://img.shields.io/npm/v/asar.svg)](https://npmjs.org/package/asar)
-
-Asar is a simple extensive archive format, it works like `tar` that concatenates
-all files together without compression, while having random access support.
-
-## Features
-
-* Support random access
-* Use JSON to store files' information
-* Very easy to write a parser
-
-## Command line utility
-
-### Install
-
-This module requires Node 10 or later.
-
-```bash
-$ npm install --engine-strict asar
-```
-
-### Usage
-
-```bash
-$ asar --help
-
- Usage: asar [options] [command]
-
- Commands:
-
- pack|p <dir> <output>
- create asar archive
-
- list|l <archive>
- list files of asar archive
-
- extract-file|ef <archive> <filename>
- extract one file from archive
-
- extract|e <archive> <dest>
- extract archive
-
-
- Options:
-
- -h, --help output usage information
- -V, --version output the version number
-
-```
-
-#### Excluding multiple resources from being packed
-
-Given:
-```
- app
-(a) ├── x1
-(b) ├── x2
-(c) ├── y3
-(d) │   ├── x1
-(e) │   └── z1
-(f) │   └── x2
-(g) └── z4
-(h) └── w1
-```
-
-Exclude: a, b
-```bash
-$ asar pack app app.asar --unpack-dir "{x1,x2}"
-```
-
-Exclude: a, b, d, f
-```bash
-$ asar pack app app.asar --unpack-dir "**/{x1,x2}"
-```
-
-Exclude: a, b, d, f, h
-```bash
-$ asar pack app app.asar --unpack-dir "{**/x1,**/x2,z4/w1}"
-```
-
-## Using programatically
-
-### Example
-
-```javascript
-const asar = require('asar');
-
-const src = 'some/path/';
-const dest = 'name.asar';
-
-await asar.createPackage(src, dest);
-console.log('done.');
-```
-
-Please note that there is currently **no** error handling provided!
-
-### Transform
-You can pass in a `transform` option, that is a function, which either returns
-nothing, or a `stream.Transform`. The latter will be used on files that will be
-in the `.asar` file to transform them (e.g. compress).
-
-```javascript
-const asar = require('asar');
-
-const src = 'some/path/';
-const dest = 'name.asar';
-
-function transform (filename) {
- return new CustomTransformStream()
-}
-
-await asar.createPackageWithOptions(src, dest, { transform: transform });
-console.log('done.');
-```
-
-## Using with grunt
-
-There is also an unofficial grunt plugin to generate asar archives at [bwin/grunt-asar][grunt-asar].
-
-## Format
-
-Asar uses [Pickle][pickle] to safely serialize binary value to file, there is
-also a [node.js binding][node-pickle] of `Pickle` class.
-
-The format of asar is very flat:
-
-```
-| UInt32: header_size | String: header | Bytes: file1 | ... | Bytes: file42 |
-```
-
-The `header_size` and `header` are serialized with [Pickle][pickle] class, and
-`header_size`'s [Pickle][pickle] object is 8 bytes.
-
-The `header` is a JSON string, and the `header_size` is the size of `header`'s
-`Pickle` object.
-
-Structure of `header` is something like this:
-
-```json
-{
- "files": {
- "tmp": {
- "files": {}
- },
- "usr" : {
- "files": {
- "bin": {
- "files": {
- "ls": {
- "offset": "0",
- "size": 100,
- "executable": true,
- "integrity": {
- "algorithm": "SHA256",
- "hash": "...",
- "blockSize": 1024,
- "blocks": ["...", "..."]
- }
- },
- "cd": {
- "offset": "100",
- "size": 100,
- "executable": true,
- "integrity": {
- "algorithm": "SHA256",
- "hash": "...",
- "blockSize": 1024,
- "blocks": ["...", "..."]
- }
- }
- }
- }
- }
- },
- "etc": {
- "files": {
- "hosts": {
- "offset": "200",
- "size": 32,
- "integrity": {
- "algorithm": "SHA256",
- "hash": "...",
- "blockSize": 1024,
- "blocks": ["...", "..."]
- }
- }
- }
- }
- }
-}
-```
-
-`offset` and `size` records the information to read the file from archive, the
-`offset` starts from 0 so you have to manually add the size of `header_size` and
-`header` to the `offset` to get the real offset of the file.
-
-`offset` is a UINT64 number represented in string, because there is no way to
-precisely represent UINT64 in JavaScript `Number`. `size` is a JavaScript
-`Number` that is no larger than `Number.MAX_SAFE_INTEGER`, which has a value of
-`9007199254740991` and is about 8PB in size. We didn't store `size` in UINT64
-because file size in Node.js is represented as `Number` and it is not safe to
-convert `Number` to UINT64.
-
-`integrity` is an object consisting of a few keys:
-* A hashing `algorithm`, currently only `SHA256` is supported.
-* A hex encoded `hash` value representing the hash of the entire file.
-* An array of hex encoded hashes for the `blocks` of the file. i.e. for a blockSize of 4KB this array contains the hash of every block if you split the file into N 4KB blocks.
-* A integer value `blockSize` representing the size in bytes of each block in the `blocks` hashes above
-
-[pickle]: https://chromium.googlesource.com/chromium/src/+/master/base/pickle.h
-[node-pickle]: https://www.npmjs.org/package/chromium-pickle
-[grunt-asar]: https://github.com/bwin/grunt-asar