summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/npm/node_modules/@fastify/busboy/lib
diff options
context:
space:
mode:
Diffstat (limited to 'g4f/Provider/npm/node_modules/@fastify/busboy/lib')
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.d.ts196
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.js85
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/multipart.js306
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/urlencoded.js190
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/Decoder.js54
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/basename.js14
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/decodeText.js26
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/getLimit.js16
-rw-r--r--g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/parseParams.js87
9 files changed, 0 insertions, 974 deletions
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.d.ts b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.d.ts
deleted file mode 100644
index 91b64483..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.d.ts
+++ /dev/null
@@ -1,196 +0,0 @@
-// Definitions by: Jacob Baskin <https://github.com/jacobbaskin>
-// BendingBender <https://github.com/BendingBender>
-// Igor Savin <https://github.com/kibertoad>
-
-/// <reference types="node" />
-
-import * as http from 'http';
-import { Readable, Writable } from 'stream';
-export { Dicer } from "../deps/dicer/lib/dicer";
-
-export const Busboy: BusboyConstructor;
-export default Busboy;
-
-export interface BusboyConfig {
- /**
- * These are the HTTP headers of the incoming request, which are used by individual parsers.
- */
- headers: BusboyHeaders;
- /**
- * `highWaterMark` to use for this Busboy instance.
- * @default WritableStream default.
- */
- highWaterMark?: number | undefined;
- /**
- * highWaterMark to use for file streams.
- * @default ReadableStream default.
- */
- fileHwm?: number | undefined;
- /**
- * Default character set to use when one isn't defined.
- * @default 'utf8'
- */
- defCharset?: string | undefined;
- /**
- * Detect if a Part is a file.
- *
- * By default a file is detected if contentType
- * is application/octet-stream or fileName is not
- * undefined.
- *
- * Modify this to handle e.g. Blobs.
- */
- isPartAFile?: (fieldName: string | undefined, contentType: string | undefined, fileName: string | undefined) => boolean;
- /**
- * If paths in the multipart 'filename' field shall be preserved.
- * @default false
- */
- preservePath?: boolean | undefined;
- /**
- * Various limits on incoming data.
- */
- limits?:
- | {
- /**
- * Max field name size (in bytes)
- * @default 100 bytes
- */
- fieldNameSize?: number | undefined;
- /**
- * Max field value size (in bytes)
- * @default 1MB
- */
- fieldSize?: number | undefined;
- /**
- * Max number of non-file fields
- * @default Infinity
- */
- fields?: number | undefined;
- /**
- * For multipart forms, the max file size (in bytes)
- * @default Infinity
- */
- fileSize?: number | undefined;
- /**
- * For multipart forms, the max number of file fields
- * @default Infinity
- */
- files?: number | undefined;
- /**
- * For multipart forms, the max number of parts (fields + files)
- * @default Infinity
- */
- parts?: number | undefined;
- /**
- * For multipart forms, the max number of header key=>value pairs to parse
- * @default 2000
- */
- headerPairs?: number | undefined;
-
- /**
- * For multipart forms, the max size of a header part
- * @default 81920
- */
- headerSize?: number | undefined;
- }
- | undefined;
-}
-
-export type BusboyHeaders = { 'content-type': string } & http.IncomingHttpHeaders;
-
-export interface BusboyFileStream extends
- Readable {
-
- truncated: boolean;
-
- /**
- * The number of bytes that have been read so far.
- */
- bytesRead: number;
-}
-
-export interface Busboy extends Writable {
- addListener<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- addListener(event: string | symbol, listener: (...args: any[]) => void): this;
-
- on<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- on(event: string | symbol, listener: (...args: any[]) => void): this;
-
- once<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- once(event: string | symbol, listener: (...args: any[]) => void): this;
-
- removeListener<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- removeListener(event: string | symbol, listener: (...args: any[]) => void): this;
-
- off<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- off(event: string | symbol, listener: (...args: any[]) => void): this;
-
- prependListener<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- prependListener(event: string | symbol, listener: (...args: any[]) => void): this;
-
- prependOnceListener<Event extends keyof BusboyEvents>(event: Event, listener: BusboyEvents[Event]): this;
-
- prependOnceListener(event: string | symbol, listener: (...args: any[]) => void): this;
-}
-
-export interface BusboyEvents {
- /**
- * Emitted for each new file form field found.
- *
- * * Note: if you listen for this event, you should always handle the `stream` no matter if you care about the
- * file contents or not (e.g. you can simply just do `stream.resume();` if you want to discard the contents),
- * otherwise the 'finish' event will never fire on the Busboy instance. However, if you don't care about **any**
- * incoming files, you can simply not listen for the 'file' event at all and any/all files will be automatically
- * and safely discarded (these discarded files do still count towards `files` and `parts` limits).
- * * If a configured file size limit was reached, `stream` will both have a boolean property `truncated`
- * (best checked at the end of the stream) and emit a 'limit' event to notify you when this happens.
- *
- * @param listener.transferEncoding Contains the 'Content-Transfer-Encoding' value for the file stream.
- * @param listener.mimeType Contains the 'Content-Type' value for the file stream.
- */
- file: (
- fieldname: string,
- stream: BusboyFileStream,
- filename: string,
- transferEncoding: string,
- mimeType: string,
- ) => void;
- /**
- * Emitted for each new non-file field found.
- */
- field: (
- fieldname: string,
- value: string,
- fieldnameTruncated: boolean,
- valueTruncated: boolean,
- transferEncoding: string,
- mimeType: string,
- ) => void;
- finish: () => void;
- /**
- * Emitted when specified `parts` limit has been reached. No more 'file' or 'field' events will be emitted.
- */
- partsLimit: () => void;
- /**
- * Emitted when specified `files` limit has been reached. No more 'file' events will be emitted.
- */
- filesLimit: () => void;
- /**
- * Emitted when specified `fields` limit has been reached. No more 'field' events will be emitted.
- */
- fieldsLimit: () => void;
- error: (error: unknown) => void;
-}
-
-export interface BusboyConstructor {
- (options: BusboyConfig): Busboy;
-
- new(options: BusboyConfig): Busboy;
-}
-
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.js
deleted file mode 100644
index 8794bebf..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/main.js
+++ /dev/null
@@ -1,85 +0,0 @@
-'use strict'
-
-const WritableStream = require('node:stream').Writable
-const { inherits } = require('node:util')
-const Dicer = require('../deps/dicer/lib/Dicer')
-
-const MultipartParser = require('./types/multipart')
-const UrlencodedParser = require('./types/urlencoded')
-const parseParams = require('./utils/parseParams')
-
-function Busboy (opts) {
- if (!(this instanceof Busboy)) { return new Busboy(opts) }
-
- if (typeof opts !== 'object') {
- throw new TypeError('Busboy expected an options-Object.')
- }
- if (typeof opts.headers !== 'object') {
- throw new TypeError('Busboy expected an options-Object with headers-attribute.')
- }
- if (typeof opts.headers['content-type'] !== 'string') {
- throw new TypeError('Missing Content-Type-header.')
- }
-
- const {
- headers,
- ...streamOptions
- } = opts
-
- this.opts = {
- autoDestroy: false,
- ...streamOptions
- }
- WritableStream.call(this, this.opts)
-
- this._done = false
- this._parser = this.getParserByHeaders(headers)
- this._finished = false
-}
-inherits(Busboy, WritableStream)
-
-Busboy.prototype.emit = function (ev) {
- if (ev === 'finish') {
- if (!this._done) {
- this._parser?.end()
- return
- } else if (this._finished) {
- return
- }
- this._finished = true
- }
- WritableStream.prototype.emit.apply(this, arguments)
-}
-
-Busboy.prototype.getParserByHeaders = function (headers) {
- const parsed = parseParams(headers['content-type'])
-
- const cfg = {
- defCharset: this.opts.defCharset,
- fileHwm: this.opts.fileHwm,
- headers,
- highWaterMark: this.opts.highWaterMark,
- isPartAFile: this.opts.isPartAFile,
- limits: this.opts.limits,
- parsedConType: parsed,
- preservePath: this.opts.preservePath
- }
-
- if (MultipartParser.detect.test(parsed[0])) {
- return new MultipartParser(this, cfg)
- }
- if (UrlencodedParser.detect.test(parsed[0])) {
- return new UrlencodedParser(this, cfg)
- }
- throw new Error('Unsupported Content-Type.')
-}
-
-Busboy.prototype._write = function (chunk, encoding, cb) {
- this._parser.write(chunk, cb)
-}
-
-module.exports = Busboy
-module.exports.default = Busboy
-module.exports.Busboy = Busboy
-
-module.exports.Dicer = Dicer
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/multipart.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/multipart.js
deleted file mode 100644
index ad242db2..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/multipart.js
+++ /dev/null
@@ -1,306 +0,0 @@
-'use strict'
-
-// TODO:
-// * support 1 nested multipart level
-// (see second multipart example here:
-// http://www.w3.org/TR/html401/interact/forms.html#didx-multipartform-data)
-// * support limits.fieldNameSize
-// -- this will require modifications to utils.parseParams
-
-const { Readable } = require('node:stream')
-const { inherits } = require('node:util')
-
-const Dicer = require('../../deps/dicer/lib/Dicer')
-
-const parseParams = require('../utils/parseParams')
-const decodeText = require('../utils/decodeText')
-const basename = require('../utils/basename')
-const getLimit = require('../utils/getLimit')
-
-const RE_BOUNDARY = /^boundary$/i
-const RE_FIELD = /^form-data$/i
-const RE_CHARSET = /^charset$/i
-const RE_FILENAME = /^filename$/i
-const RE_NAME = /^name$/i
-
-Multipart.detect = /^multipart\/form-data/i
-function Multipart (boy, cfg) {
- let i
- let len
- const self = this
- let boundary
- const limits = cfg.limits
- const isPartAFile = cfg.isPartAFile || ((fieldName, contentType, fileName) => (contentType === 'application/octet-stream' || fileName !== undefined))
- const parsedConType = cfg.parsedConType || []
- const defCharset = cfg.defCharset || 'utf8'
- const preservePath = cfg.preservePath
- const fileOpts = { highWaterMark: cfg.fileHwm }
-
- for (i = 0, len = parsedConType.length; i < len; ++i) {
- if (Array.isArray(parsedConType[i]) &&
- RE_BOUNDARY.test(parsedConType[i][0])) {
- boundary = parsedConType[i][1]
- break
- }
- }
-
- function checkFinished () {
- if (nends === 0 && finished && !boy._done) {
- finished = false
- self.end()
- }
- }
-
- if (typeof boundary !== 'string') { throw new Error('Multipart: Boundary not found') }
-
- const fieldSizeLimit = getLimit(limits, 'fieldSize', 1 * 1024 * 1024)
- const fileSizeLimit = getLimit(limits, 'fileSize', Infinity)
- const filesLimit = getLimit(limits, 'files', Infinity)
- const fieldsLimit = getLimit(limits, 'fields', Infinity)
- const partsLimit = getLimit(limits, 'parts', Infinity)
- const headerPairsLimit = getLimit(limits, 'headerPairs', 2000)
- const headerSizeLimit = getLimit(limits, 'headerSize', 80 * 1024)
-
- let nfiles = 0
- let nfields = 0
- let nends = 0
- let curFile
- let curField
- let finished = false
-
- this._needDrain = false
- this._pause = false
- this._cb = undefined
- this._nparts = 0
- this._boy = boy
-
- const parserCfg = {
- boundary,
- maxHeaderPairs: headerPairsLimit,
- maxHeaderSize: headerSizeLimit,
- partHwm: fileOpts.highWaterMark,
- highWaterMark: cfg.highWaterMark
- }
-
- this.parser = new Dicer(parserCfg)
- this.parser.on('drain', function () {
- self._needDrain = false
- if (self._cb && !self._pause) {
- const cb = self._cb
- self._cb = undefined
- cb()
- }
- }).on('part', function onPart (part) {
- if (++self._nparts > partsLimit) {
- self.parser.removeListener('part', onPart)
- self.parser.on('part', skipPart)
- boy.hitPartsLimit = true
- boy.emit('partsLimit')
- return skipPart(part)
- }
-
- // hack because streams2 _always_ doesn't emit 'end' until nextTick, so let
- // us emit 'end' early since we know the part has ended if we are already
- // seeing the next part
- if (curField) {
- const field = curField
- field.emit('end')
- field.removeAllListeners('end')
- }
-
- part.on('header', function (header) {
- let contype
- let fieldname
- let parsed
- let charset
- let encoding
- let filename
- let nsize = 0
-
- if (header['content-type']) {
- parsed = parseParams(header['content-type'][0])
- if (parsed[0]) {
- contype = parsed[0].toLowerCase()
- for (i = 0, len = parsed.length; i < len; ++i) {
- if (RE_CHARSET.test(parsed[i][0])) {
- charset = parsed[i][1].toLowerCase()
- break
- }
- }
- }
- }
-
- if (contype === undefined) { contype = 'text/plain' }
- if (charset === undefined) { charset = defCharset }
-
- if (header['content-disposition']) {
- parsed = parseParams(header['content-disposition'][0])
- if (!RE_FIELD.test(parsed[0])) { return skipPart(part) }
- for (i = 0, len = parsed.length; i < len; ++i) {
- if (RE_NAME.test(parsed[i][0])) {
- fieldname = parsed[i][1]
- } else if (RE_FILENAME.test(parsed[i][0])) {
- filename = parsed[i][1]
- if (!preservePath) { filename = basename(filename) }
- }
- }
- } else { return skipPart(part) }
-
- if (header['content-transfer-encoding']) { encoding = header['content-transfer-encoding'][0].toLowerCase() } else { encoding = '7bit' }
-
- let onData,
- onEnd
-
- if (isPartAFile(fieldname, contype, filename)) {
- // file/binary field
- if (nfiles === filesLimit) {
- if (!boy.hitFilesLimit) {
- boy.hitFilesLimit = true
- boy.emit('filesLimit')
- }
- return skipPart(part)
- }
-
- ++nfiles
-
- if (!boy._events.file) {
- self.parser._ignore()
- return
- }
-
- ++nends
- const file = new FileStream(fileOpts)
- curFile = file
- file.on('end', function () {
- --nends
- self._pause = false
- checkFinished()
- if (self._cb && !self._needDrain) {
- const cb = self._cb
- self._cb = undefined
- cb()
- }
- })
- file._read = function (n) {
- if (!self._pause) { return }
- self._pause = false
- if (self._cb && !self._needDrain) {
- const cb = self._cb
- self._cb = undefined
- cb()
- }
- }
- boy.emit('file', fieldname, file, filename, encoding, contype)
-
- onData = function (data) {
- if ((nsize += data.length) > fileSizeLimit) {
- const extralen = fileSizeLimit - nsize + data.length
- if (extralen > 0) { file.push(data.slice(0, extralen)) }
- file.truncated = true
- file.bytesRead = fileSizeLimit
- part.removeAllListeners('data')
- file.emit('limit')
- return
- } else if (!file.push(data)) { self._pause = true }
-
- file.bytesRead = nsize
- }
-
- onEnd = function () {
- curFile = undefined
- file.push(null)
- }
- } else {
- // non-file field
- if (nfields === fieldsLimit) {
- if (!boy.hitFieldsLimit) {
- boy.hitFieldsLimit = true
- boy.emit('fieldsLimit')
- }
- return skipPart(part)
- }
-
- ++nfields
- ++nends
- let buffer = ''
- let truncated = false
- curField = part
-
- onData = function (data) {
- if ((nsize += data.length) > fieldSizeLimit) {
- const extralen = (fieldSizeLimit - (nsize - data.length))
- buffer += data.toString('binary', 0, extralen)
- truncated = true
- part.removeAllListeners('data')
- } else { buffer += data.toString('binary') }
- }
-
- onEnd = function () {
- curField = undefined
- if (buffer.length) { buffer = decodeText(buffer, 'binary', charset) }
- boy.emit('field', fieldname, buffer, false, truncated, encoding, contype)
- --nends
- checkFinished()
- }
- }
-
- /* As of node@2efe4ab761666 (v0.10.29+/v0.11.14+), busboy had become
- broken. Streams2/streams3 is a huge black box of confusion, but
- somehow overriding the sync state seems to fix things again (and still
- seems to work for previous node versions).
- */
- part._readableState.sync = false
-
- part.on('data', onData)
- part.on('end', onEnd)
- }).on('error', function (err) {
- if (curFile) { curFile.emit('error', err) }
- })
- }).on('error', function (err) {
- boy.emit('error', err)
- }).on('finish', function () {
- finished = true
- checkFinished()
- })
-}
-
-Multipart.prototype.write = function (chunk, cb) {
- const r = this.parser.write(chunk)
- if (r && !this._pause) {
- cb()
- } else {
- this._needDrain = !r
- this._cb = cb
- }
-}
-
-Multipart.prototype.end = function () {
- const self = this
-
- if (self.parser.writable) {
- self.parser.end()
- } else if (!self._boy._done) {
- process.nextTick(function () {
- self._boy._done = true
- self._boy.emit('finish')
- })
- }
-}
-
-function skipPart (part) {
- part.resume()
-}
-
-function FileStream (opts) {
- Readable.call(this, opts)
-
- this.bytesRead = 0
-
- this.truncated = false
-}
-
-inherits(FileStream, Readable)
-
-FileStream.prototype._read = function (n) {}
-
-module.exports = Multipart
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/urlencoded.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/urlencoded.js
deleted file mode 100644
index 6f5f7846..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/types/urlencoded.js
+++ /dev/null
@@ -1,190 +0,0 @@
-'use strict'
-
-const Decoder = require('../utils/Decoder')
-const decodeText = require('../utils/decodeText')
-const getLimit = require('../utils/getLimit')
-
-const RE_CHARSET = /^charset$/i
-
-UrlEncoded.detect = /^application\/x-www-form-urlencoded/i
-function UrlEncoded (boy, cfg) {
- const limits = cfg.limits
- const parsedConType = cfg.parsedConType
- this.boy = boy
-
- this.fieldSizeLimit = getLimit(limits, 'fieldSize', 1 * 1024 * 1024)
- this.fieldNameSizeLimit = getLimit(limits, 'fieldNameSize', 100)
- this.fieldsLimit = getLimit(limits, 'fields', Infinity)
-
- let charset
- for (var i = 0, len = parsedConType.length; i < len; ++i) { // eslint-disable-line no-var
- if (Array.isArray(parsedConType[i]) &&
- RE_CHARSET.test(parsedConType[i][0])) {
- charset = parsedConType[i][1].toLowerCase()
- break
- }
- }
-
- if (charset === undefined) { charset = cfg.defCharset || 'utf8' }
-
- this.decoder = new Decoder()
- this.charset = charset
- this._fields = 0
- this._state = 'key'
- this._checkingBytes = true
- this._bytesKey = 0
- this._bytesVal = 0
- this._key = ''
- this._val = ''
- this._keyTrunc = false
- this._valTrunc = false
- this._hitLimit = false
-}
-
-UrlEncoded.prototype.write = function (data, cb) {
- if (this._fields === this.fieldsLimit) {
- if (!this.boy.hitFieldsLimit) {
- this.boy.hitFieldsLimit = true
- this.boy.emit('fieldsLimit')
- }
- return cb()
- }
-
- let idxeq; let idxamp; let i; let p = 0; const len = data.length
-
- while (p < len) {
- if (this._state === 'key') {
- idxeq = idxamp = undefined
- for (i = p; i < len; ++i) {
- if (!this._checkingBytes) { ++p }
- if (data[i] === 0x3D/* = */) {
- idxeq = i
- break
- } else if (data[i] === 0x26/* & */) {
- idxamp = i
- break
- }
- if (this._checkingBytes && this._bytesKey === this.fieldNameSizeLimit) {
- this._hitLimit = true
- break
- } else if (this._checkingBytes) { ++this._bytesKey }
- }
-
- if (idxeq !== undefined) {
- // key with assignment
- if (idxeq > p) { this._key += this.decoder.write(data.toString('binary', p, idxeq)) }
- this._state = 'val'
-
- this._hitLimit = false
- this._checkingBytes = true
- this._val = ''
- this._bytesVal = 0
- this._valTrunc = false
- this.decoder.reset()
-
- p = idxeq + 1
- } else if (idxamp !== undefined) {
- // key with no assignment
- ++this._fields
- let key; const keyTrunc = this._keyTrunc
- if (idxamp > p) { key = (this._key += this.decoder.write(data.toString('binary', p, idxamp))) } else { key = this._key }
-
- this._hitLimit = false
- this._checkingBytes = true
- this._key = ''
- this._bytesKey = 0
- this._keyTrunc = false
- this.decoder.reset()
-
- if (key.length) {
- this.boy.emit('field', decodeText(key, 'binary', this.charset),
- '',
- keyTrunc,
- false)
- }
-
- p = idxamp + 1
- if (this._fields === this.fieldsLimit) { return cb() }
- } else if (this._hitLimit) {
- // we may not have hit the actual limit if there are encoded bytes...
- if (i > p) { this._key += this.decoder.write(data.toString('binary', p, i)) }
- p = i
- if ((this._bytesKey = this._key.length) === this.fieldNameSizeLimit) {
- // yep, we actually did hit the limit
- this._checkingBytes = false
- this._keyTrunc = true
- }
- } else {
- if (p < len) { this._key += this.decoder.write(data.toString('binary', p)) }
- p = len
- }
- } else {
- idxamp = undefined
- for (i = p; i < len; ++i) {
- if (!this._checkingBytes) { ++p }
- if (data[i] === 0x26/* & */) {
- idxamp = i
- break
- }
- if (this._checkingBytes && this._bytesVal === this.fieldSizeLimit) {
- this._hitLimit = true
- break
- } else if (this._checkingBytes) { ++this._bytesVal }
- }
-
- if (idxamp !== undefined) {
- ++this._fields
- if (idxamp > p) { this._val += this.decoder.write(data.toString('binary', p, idxamp)) }
- this.boy.emit('field', decodeText(this._key, 'binary', this.charset),
- decodeText(this._val, 'binary', this.charset),
- this._keyTrunc,
- this._valTrunc)
- this._state = 'key'
-
- this._hitLimit = false
- this._checkingBytes = true
- this._key = ''
- this._bytesKey = 0
- this._keyTrunc = false
- this.decoder.reset()
-
- p = idxamp + 1
- if (this._fields === this.fieldsLimit) { return cb() }
- } else if (this._hitLimit) {
- // we may not have hit the actual limit if there are encoded bytes...
- if (i > p) { this._val += this.decoder.write(data.toString('binary', p, i)) }
- p = i
- if ((this._val === '' && this.fieldSizeLimit === 0) ||
- (this._bytesVal = this._val.length) === this.fieldSizeLimit) {
- // yep, we actually did hit the limit
- this._checkingBytes = false
- this._valTrunc = true
- }
- } else {
- if (p < len) { this._val += this.decoder.write(data.toString('binary', p)) }
- p = len
- }
- }
- }
- cb()
-}
-
-UrlEncoded.prototype.end = function () {
- if (this.boy._done) { return }
-
- if (this._state === 'key' && this._key.length > 0) {
- this.boy.emit('field', decodeText(this._key, 'binary', this.charset),
- '',
- this._keyTrunc,
- false)
- } else if (this._state === 'val') {
- this.boy.emit('field', decodeText(this._key, 'binary', this.charset),
- decodeText(this._val, 'binary', this.charset),
- this._keyTrunc,
- this._valTrunc)
- }
- this.boy._done = true
- this.boy.emit('finish')
-}
-
-module.exports = UrlEncoded
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/Decoder.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/Decoder.js
deleted file mode 100644
index 7917678c..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/Decoder.js
+++ /dev/null
@@ -1,54 +0,0 @@
-'use strict'
-
-const RE_PLUS = /\+/g
-
-const HEX = [
- 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0,
- 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0,
- 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0
-]
-
-function Decoder () {
- this.buffer = undefined
-}
-Decoder.prototype.write = function (str) {
- // Replace '+' with ' ' before decoding
- str = str.replace(RE_PLUS, ' ')
- let res = ''
- let i = 0; let p = 0; const len = str.length
- for (; i < len; ++i) {
- if (this.buffer !== undefined) {
- if (!HEX[str.charCodeAt(i)]) {
- res += '%' + this.buffer
- this.buffer = undefined
- --i // retry character
- } else {
- this.buffer += str[i]
- ++p
- if (this.buffer.length === 2) {
- res += String.fromCharCode(parseInt(this.buffer, 16))
- this.buffer = undefined
- }
- }
- } else if (str[i] === '%') {
- if (i > p) {
- res += str.substring(p, i)
- p = i
- }
- this.buffer = ''
- ++p
- }
- }
- if (p < len && this.buffer === undefined) { res += str.substring(p) }
- return res
-}
-Decoder.prototype.reset = function () {
- this.buffer = undefined
-}
-
-module.exports = Decoder
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/basename.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/basename.js
deleted file mode 100644
index db588199..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/basename.js
+++ /dev/null
@@ -1,14 +0,0 @@
-'use strict'
-
-module.exports = function basename (path) {
- if (typeof path !== 'string') { return '' }
- for (var i = path.length - 1; i >= 0; --i) { // eslint-disable-line no-var
- switch (path.charCodeAt(i)) {
- case 0x2F: // '/'
- case 0x5C: // '\'
- path = path.slice(i + 1)
- return (path === '..' || path === '.' ? '' : path)
- }
- }
- return (path === '..' || path === '.' ? '' : path)
-}
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/decodeText.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/decodeText.js
deleted file mode 100644
index ee376062..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/decodeText.js
+++ /dev/null
@@ -1,26 +0,0 @@
-'use strict'
-
-// Node has always utf-8
-const utf8Decoder = new TextDecoder('utf-8')
-const textDecoders = new Map([
- ['utf-8', utf8Decoder],
- ['utf8', utf8Decoder]
-])
-
-function decodeText (text, textEncoding, destEncoding) {
- if (text) {
- if (textDecoders.has(destEncoding)) {
- try {
- return textDecoders.get(destEncoding).decode(Buffer.from(text, textEncoding))
- } catch (e) { }
- } else {
- try {
- textDecoders.set(destEncoding, new TextDecoder(destEncoding))
- return textDecoders.get(destEncoding).decode(Buffer.from(text, textEncoding))
- } catch (e) { }
- }
- }
- return text
-}
-
-module.exports = decodeText
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/getLimit.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/getLimit.js
deleted file mode 100644
index cb64fd67..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/getLimit.js
+++ /dev/null
@@ -1,16 +0,0 @@
-'use strict'
-
-module.exports = function getLimit (limits, name, defaultLimit) {
- if (
- !limits ||
- limits[name] === undefined ||
- limits[name] === null
- ) { return defaultLimit }
-
- if (
- typeof limits[name] !== 'number' ||
- isNaN(limits[name])
- ) { throw new TypeError('Limit ' + name + ' is not a valid number') }
-
- return limits[name]
-}
diff --git a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/parseParams.js b/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/parseParams.js
deleted file mode 100644
index f9214180..00000000
--- a/g4f/Provider/npm/node_modules/@fastify/busboy/lib/utils/parseParams.js
+++ /dev/null
@@ -1,87 +0,0 @@
-'use strict'
-
-const decodeText = require('./decodeText')
-
-const RE_ENCODED = /%([a-fA-F0-9]{2})/g
-
-function encodedReplacer (match, byte) {
- return String.fromCharCode(parseInt(byte, 16))
-}
-
-function parseParams (str) {
- const res = []
- let state = 'key'
- let charset = ''
- let inquote = false
- let escaping = false
- let p = 0
- let tmp = ''
-
- for (var i = 0, len = str.length; i < len; ++i) { // eslint-disable-line no-var
- const char = str[i]
- if (char === '\\' && inquote) {
- if (escaping) { escaping = false } else {
- escaping = true
- continue
- }
- } else if (char === '"') {
- if (!escaping) {
- if (inquote) {
- inquote = false
- state = 'key'
- } else { inquote = true }
- continue
- } else { escaping = false }
- } else {
- if (escaping && inquote) { tmp += '\\' }
- escaping = false
- if ((state === 'charset' || state === 'lang') && char === "'") {
- if (state === 'charset') {
- state = 'lang'
- charset = tmp.substring(1)
- } else { state = 'value' }
- tmp = ''
- continue
- } else if (state === 'key' &&
- (char === '*' || char === '=') &&
- res.length) {
- if (char === '*') { state = 'charset' } else { state = 'value' }
- res[p] = [tmp, undefined]
- tmp = ''
- continue
- } else if (!inquote && char === ';') {
- state = 'key'
- if (charset) {
- if (tmp.length) {
- tmp = decodeText(tmp.replace(RE_ENCODED, encodedReplacer),
- 'binary',
- charset)
- }
- charset = ''
- } else if (tmp.length) {
- tmp = decodeText(tmp, 'binary', 'utf8')
- }
- if (res[p] === undefined) { res[p] = tmp } else { res[p][1] = tmp }
- tmp = ''
- ++p
- continue
- } else if (!inquote && (char === ' ' || char === '\t')) { continue }
- }
- tmp += char
- }
- if (charset && tmp.length) {
- tmp = decodeText(tmp.replace(RE_ENCODED, encodedReplacer),
- 'binary',
- charset)
- } else if (tmp) {
- tmp = decodeText(tmp, 'binary', 'utf8')
- }
-
- if (res[p] === undefined) {
- if (tmp) { res[p] = tmp }
- } else { res[p][1] = tmp }
-
- return res
-}
-
-module.exports = parseParams