Processors
processors
import { AutoProcessor, read_audio } from '@xenova/transformers';
let processor = await AutoProcessor.from_pretrained('openai/whisper-tiny.en');
let audio = await read_audio('https://boincai.com/datasets/Narsil/asr_dummy/resolve/main/mlk.flac', 16000);
let { input_features } = await processor(audio);
// Tensor {
// data: Float32Array(240000) [0.4752984642982483, 0.5597258806228638, 0.56434166431427, ...],
// dims: [1, 80, 3000],
// type: 'float32',
// size: 240000,
// }processors.FeatureExtractor β <code> Callable </code>
new FeatureExtractor(config)
Param
Type
Description
processors.ImageFeatureExtractor β <code> FeatureExtractor </code>
new ImageFeatureExtractor(config)
Param
Type
Description
imageFeatureExtractor.thumbnail(image, size, [resample]) β <code> Promise. < RawImage > </code>
Param
Type
Default
Description
imageFeatureExtractor.preprocess(image) β <code> Promise. < PreprocessedImage > </code>
Param
Type
Description
imageFeatureExtractor._call(images, ...args) β <code> Promise. < ImageFeatureExtractorResult > </code>
Param
Type
Description
processors.DetrFeatureExtractor β <code> ImageFeatureExtractor </code>
detrFeatureExtractor._call(urls) β <code> Promise. < DetrFeatureExtractorResult > </code>
Param
Type
Description
detrFeatureExtractor.post_process_object_detection() : <code> post_process_object_detection </code>
detrFeatureExtractor.remove_low_and_no_objects(class_logits, mask_logits, object_mask_threshold, num_labels) β <code> * </code>
Param
Type
Description
detrFeatureExtractor.check_segment_validity(mask_labels, mask_probs, k, mask_threshold, overlap_mask_area_threshold) β <code> * </code>
Param
Type
Default
Description
detrFeatureExtractor.compute_segments(mask_probs, pred_scores, pred_labels, mask_threshold, overlap_mask_area_threshold, label_ids_to_fuse, target_size) β <code> * </code>
Param
Type
Default
Description
detrFeatureExtractor.post_process_panoptic_segmentation(outputs, [threshold], [mask_threshold], [overlap_mask_area_threshold], [label_ids_to_fuse], [target_sizes]) β <code> Array. < {segmentation: Tensor, segments_info: Array < {id: number, label_id: number, score: number} > } > </code>
Param
Type
Default
Description
processors.Processor β <code> Callable </code>
new Processor(feature_extractor)
Param
Type
Description
processor._call(input, ...args) β <code> Promise. < any > </code>
Param
Type
Description
processors.WhisperProcessor β <code> Processor </code>
whisperProcessor._call(audio) β <code> Promise. < any > </code>
Param
Type
Description
processors.AutoProcessor
AutoProcessor.from_pretrained(pretrained_model_name_or_path, options) β <code> Promise. < Processor > </code>
Param
Type
Description
processors~center_to_corners_format(arr) β <code> Array. < number > </code>
Param
Type
Description
processors~post_process_object_detection(outputs) β <code> Array. < Object > </code>
Param
Type
Description
post_process_object_detection~box : <code> Array. < number > </code>
processors~HeightWidth : <code> * </code>
processors~ImageFeatureExtractorResult : <code> object </code>
Name
Type
Description
processors~PreprocessedImage : <code> object </code>
Name
Type
Description
processors~DetrFeatureExtractorResult : <code> object </code>
Name
Type
processors~SamImageProcessorResult : <code> object </code>
Name
Type
Last updated