Bark
Bark
Overview
Bark is a transformer-based text-to-speech model proposed by Suno AI in suno-ai/bark.
Bark is made of 4 main models:
- BarkSemanticModel (also referred to as the βtextβ model): a causal auto-regressive transformer model that takes as input tokenized text, and predicts semantic text tokens that capture the meaning of the text. 
- BarkCoarseModel (also referred to as the βcoarse acousticsβ model): a causal autoregressive transformer, that takes as input the results of the BarkSemanticModel model. It aims at predicting the first two audio codebooks necessary for EnCodec. 
- BarkFineModel (the βfine acousticsβ model), this time a non-causal autoencoder transformer, which iteratively predicts the last codebooks based on the sum of the previous codebooks embeddings. 
- having predicted all the codebook channels from the EncodecModel, Bark uses it to decode the output audio array. 
It should be noted that each of the first three modules can support conditional speaker embeddings to condition the output sound according to specific predefined voice.
Optimizing Bark
Bark can be optimized with just a few extra lines of code, which significantly reduces its memory footprint and accelerates inference.
Using half-precision
You can speed up inference and reduce memory footprint by 50% simply by loading the model in half-precision.
Copied
from transformers import BarkModel
import torch
device = "cuda" if torch.cuda.is_available() else "cpu"
model = BarkModel.from_pretrained("suno/bark-small", torch_dtype=torch.float16).to(device)Using π Better Transformer
Better Transformer is an π Optimum feature that performs kernel fusion under the hood. You can gain 20% to 30% in speed with zero performance degradation. It only requires one line of code to export the model to π Better Transformer:
Copied
model =  model.to_bettertransformer()Note that π Optimum must be installed before using this feature. Hereβs how to install it.
Using CPU offload
As mentioned above, Bark is made up of 4 sub-models, which are called up sequentially during audio generation. In other words, while one sub-model is in use, the other sub-models are idle.
If youβre using a CUDA device, a simple solution to benefit from an 80% reduction in memory footprint is to offload the GPUβs submodels when theyβre idle. This operation is called CPU offloading. You can use it with one line of code.
Copied
model.enable_cpu_offload()Note that π Accelerate must be installed before using this feature. Hereβs how to install it.
Combining optimizaton techniques
You can combine optimization techniques, and use CPU offload, half-precision and π Better Transformer all at once.
Copied
from transformers import BarkModel
import torch
device = "cuda" if torch.cuda.is_available() else "cpu"
# load in fp16
model = BarkModel.from_pretrained("suno/bark-small", torch_dtype=torch.float16).to(device)
# convert to bettertransformer
model = BetterTransformer.transform(model, keep_original_model=False)
# enable CPU offload
model.enable_cpu_offload()Find out more on inference optimization techniques here.
Tips
Suno offers a library of voice presets in a number of languages here. These presets are also uploaded in the hub here or here.
Copied
>>> from transformers import AutoProcessor, BarkModel
>>> processor = AutoProcessor.from_pretrained("suno/bark")
>>> model = BarkModel.from_pretrained("suno/bark")
>>> voice_preset = "v2/en_speaker_6"
>>> inputs = processor("Hello, my dog is cute", voice_preset=voice_preset)
>>> audio_array = model.generate(**inputs)
>>> audio_array = audio_array.cpu().numpy().squeeze()Bark can generate highly realistic, multilingual speech as well as other audio - including music, background noise and simple sound effects.
Copied
>>> # Multilingual speech - simplified Chinese
>>> inputs = processor("ζδΊΊηοΌζδΌθ―΄δΈζ")
>>> # Multilingual speech - French - let's use a voice_preset as well
>>> inputs = processor("Incroyable! Je peux gΓ©nΓ©rer du son.", voice_preset="fr_speaker_5")
>>> # Bark can also generate music. You can help it out by adding music notes around your lyrics.
>>> inputs = processor("βͺ Hello, my dog is cute βͺ")
>>> audio_array = model.generate(**inputs)
>>> audio_array = audio_array.cpu().numpy().squeeze()The model can also produce nonverbal communications like laughing, sighing and crying.
Copied
>>> # Adding non-speech cues to the input text
>>> inputs = processor("Hello uh ... [clears throat], my dog is cute [laughter]")
>>> audio_array = model.generate(**inputs)
>>> audio_array = audio_array.cpu().numpy().squeeze()To save the audio, simply take the sample rate from the model config and some scipy utility:
Copied
>>> from scipy.io.wavfile import write as write_wav
>>> # save audio to disk, but first take the sample rate from the model config
>>> sample_rate = model.generation_config.sample_rate
>>> write_wav("bark_generation.wav", sample_rate, audio_array)This model was contributed by Yoach Lacombe (ylacombe) and Sanchit Gandhi (sanchit-gandhi). The original code can be found here.
BarkConfig
class transformers.BarkConfig
( semantic_config: typing.Dict = Nonecoarse_acoustics_config: typing.Dict = Nonefine_acoustics_config: typing.Dict = Nonecodec_config: typing.Dict = Noneinitializer_range = 0.02**kwargs )
Parameters
- semantic_config (BarkSemanticConfig, optional) β Configuration of the underlying semantic sub-model. 
- coarse_acoustics_config (BarkCoarseConfig, optional) β Configuration of the underlying coarse acoustics sub-model. 
- fine_acoustics_config (BarkFineConfig, optional) β Configuration of the underlying fine acoustics sub-model. 
- codec_config (AutoConfig, optional) β Configuration of the underlying codec sub-model. - Example β 
This is the configuration class to store the configuration of a BarkModel. It is used to instantiate a Bark model according to the specified sub-models configurations, defining the model architecture.
Instantiating a configuration with the defaults will yield a similar configuration to that of the Bark suno/bark architecture.
Configuration objects inherit from PretrainedConfig and can be used to control the model outputs. Read the documentation from PretrainedConfig for more information.
from_sub_model_configs
( semantic_config: BarkSemanticConfigcoarse_acoustics_config: BarkCoarseConfigfine_acoustics_config: BarkFineConfigcodec_config: PretrainedConfig**kwargs ) β BarkConfig
Returns
An instance of a configuration object
Instantiate a BarkConfig (or a derived class) from bark sub-models configuration.
BarkProcessor
class transformers.BarkProcessor
( tokenizerspeaker_embeddings = None )
Parameters
- tokenizer (PreTrainedTokenizer) β An instance of PreTrainedTokenizer. 
- speaker_embeddings ( - Dict[Dict[str]], optional, defaults to- None) β Optional nested speaker embeddings dictionary. The first level contains voice preset names (e.g- "en_speaker_4"). The second level contains- "semantic_prompt",- "coarse_prompt"and- "fine_prompt"embeddings. The values correspond to the path of the corresponding- np.ndarray. See here for a list of- voice_preset_names.
Constructs a Bark processor which wraps a text tokenizer and optional Bark voice presets into a single processor.
__call__
( text = Nonevoice_preset = Nonereturn_tensors = 'pt'max_length = 256add_special_tokens = Falsereturn_attention_mask = Truereturn_token_type_ids = False**kwargs ) β Tuple(BatchEncoding, BatchFeature)
Parameters
- text ( - str,- List[str],- List[List[str]]) β The sequence or batch of sequences to be encoded. Each sequence can be a string or a list of strings (pretokenized string). If the sequences are provided as list of strings (pretokenized), you must set- is_split_into_words=True(to lift the ambiguity with a batch of sequences).
- voice_preset ( - str,- Dict[np.ndarray]) β The voice preset, i.e the speaker embeddings. It can either be a valid voice_preset name, e.g- "en_speaker_1", or directly a dictionnary of- np.ndarrayembeddings for each submodel of- Bark. Or it can be a valid file name of a local- .npzsingle voice preset.
- return_tensors ( - stror TensorType, optional) β If set, will return tensors of a particular framework. Acceptable values are:- 'pt': Return PyTorch- torch.Tensorobjects.
- 'np': Return NumPy- np.ndarrayobjects.
 
Returns
Tuple(BatchEncoding, BatchFeature)
A tuple composed of a BatchEncoding, i.e the output of the tokenizer and a BatchFeature, i.e the voice preset with the right tensors type.
Main method to prepare for the model one or several sequences(s). This method forwards the text and kwargs arguments to the AutoTokenizerβs __call__() to encode the text. The method also proposes a voice preset which is a dictionary of arrays that conditions Barkβs output. kwargs arguments are forwarded to the tokenizer and to cached_file method if voice_preset is a valid filename.
from_pretrained
( pretrained_processor_name_or_pathspeaker_embeddings_dict_path = 'speaker_embeddings_path.json'**kwargs )
Parameters
- pretrained_model_name_or_path ( - stror- os.PathLike) β This can be either:- a string, the model id of a pretrained BarkProcessor hosted inside a model repo on boincai.com. Valid model ids can be located at the root-level, like - bert-base-uncased, or namespaced under a user or organization name, like- dbmdz/bert-base-german-cased.
- a path to a directory containing a processor saved using the save_pretrained() method, e.g., - ./my_model_directory/.
 
- speaker_embeddings_dict_path ( - str, optional, defaults to- "speaker_embeddings_path.json") β The name of the- .jsonfile containing the speaker_embeddings dictionnary located in- pretrained_model_name_or_path. If- None, no speaker_embeddings is loaded. **kwargs β Additional keyword arguments passed along to both- ~tokenization_utils_base.PreTrainedTokenizer.from_pretrained.
Instantiate a Bark processor associated with a pretrained model.
save_pretrained
( save_directoryspeaker_embeddings_dict_path = 'speaker_embeddings_path.json'speaker_embeddings_directory = 'speaker_embeddings'push_to_hub: bool = False**kwargs )
Parameters
- save_directory ( - stror- os.PathLike) β Directory where the tokenizer files and the speaker embeddings will be saved (directory will be created if it does not exist).
- speaker_embeddings_dict_path ( - str, optional, defaults to- "speaker_embeddings_path.json") β The name of the- .jsonfile that will contains the speaker_embeddings nested path dictionnary, if it exists, and that will be located in- pretrained_model_name_or_path/speaker_embeddings_directory.
- speaker_embeddings_directory ( - str, optional, defaults to- "speaker_embeddings/") β The name of the folder in which the speaker_embeddings arrays will be saved.
- push_to_hub ( - bool, optional, defaults to- False) β Whether or not to push your model to the BOINC AI model hub after saving it. You can specify the repository you want to push to with- repo_id(will default to the name of- save_directoryin your namespace). kwargs β Additional key word arguments passed along to the push_to_hub() method.
Saves the attributes of this processor (tokenizerβ¦) in the specified directory so that it can be reloaded using the from_pretrained() method.
BarkModel
class transformers.BarkModel
( config )
Parameters
- config (BarkConfig) β Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the from_pretrained() method to load the model weights. 
The full Bark model, a text-to-speech model composed of 4 sub-models:
- BarkSemanticModel (also referred to as the βtextβ model): a causal auto-regressive transformer model that takes as input tokenized text, and predicts semantic text tokens that capture the meaning of the text. 
- BarkCoarseModel (also refered to as the βcoarse acousticsβ model), also a causal autoregressive transformer, that takes into input the results of the last model. It aims at regressing the first two audio codebooks necessary to - encodec.
- BarkFineModel (the βfine acousticsβ model), this time a non-causal autoencoder transformer, which iteratively predicts the last codebooks based on the sum of the previous codebooks embeddings. 
- having predicted all the codebook channels from the EncodecModel, Bark uses it to decode the output audio array. 
It should be noted that each of the first three modules can support conditional speaker embeddings to condition the output sound according to specific predefined voice.
This model inherits from PreTrainedModel. Check the superclass documentation for the generic methods the library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads etc.)
This model is also a PyTorch torch.nn.Module subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior.
generate
( input_ids: typing.Optional[torch.Tensor] = Nonehistory_prompt: typing.Union[typing.Dict[str, torch.Tensor], NoneType] = None**kwargs ) β torch.LongTensor
Parameters
- input_ids ( - Optional[torch.Tensor]of shape (batch_size, seq_len), optional) β Input ids. Will be truncated up to 256 tokens. Note that the output audios will be as long as the longest generation among the batch.
- history_prompt ( - Optional[Dict[str,torch.Tensor]], optional) β Optional- Barkspeaker prompt. Note that for now, this model takes only one speaker prompt per batch.
- kwargs (optional) β Remaining dictionary of keyword arguments. Keyword arguments are of two types: - Without a prefix, they will be entered as - **kwargsfor the- generatemethod of each sub-model.
- With a semantic_, coarse_, fine_ prefix, they will be input for the - generatemethod of the semantic, coarse and fine respectively. It has the priority over the keywords without a prefix.
 - This means you can, for example, specify a generation strategy for all sub-models except one. 
Returns
torch.LongTensor
Output generated audio.
Generates audio from an input prompt and an additional optional Bark speaker prompt.
Example:
Copied
>>> from transformers import AutoProcessor, BarkModel
>>> processor = AutoProcessor.from_pretrained("suno/bark-small")
>>> model = BarkModel.from_pretrained("suno/bark-small")
>>> # To add a voice preset, you can pass `voice_preset` to `BarkProcessor.__call__(...)`
>>> voice_preset = "v2/en_speaker_6"
>>> inputs = processor("Hello, my dog is cute, I need him in my life", voice_preset=voice_preset)
>>> audio_array = model.generate(**inputs, semantic_max_new_tokens=100)
>>> audio_array = audio_array.cpu().numpy().squeeze()enable_cpu_offload
( gpu_id: typing.Optional[int] = 0 )
Parameters
- gpu_id ( - int, optional, defaults to 0) β GPU id on which the sub-models will be loaded and offloaded.
Offloads all sub-models to CPU using accelerate, reducing memory usage with a low impact on performance. This method moves one whole sub-model at a time to the GPU when it is used, and the sub-model remains in GPU until the next sub-model runs.
BarkSemanticModel
class transformers.BarkSemanticModel
( config )
Parameters
- config (BarkSemanticConfig) β Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the from_pretrained() method to load the model weights. 
Bark semantic (or text) model. It shares the same architecture as the coarse model. It is a GPT-2 like autoregressive model with a language modeling head on top. This model inherits from PreTrainedModel. Check the superclass documentation for the generic methods the library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads etc.)
This model is also a PyTorch torch.nn.Module subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior.
forward
( input_ids: typing.Optional[torch.Tensor] = Nonepast_key_values: typing.Optional[typing.Tuple[torch.FloatTensor]] = Noneattention_mask: typing.Optional[torch.Tensor] = Noneposition_ids: typing.Optional[torch.Tensor] = Nonehead_mask: typing.Optional[torch.Tensor] = Nonelabels: typing.Optional[torch.LongTensor] = Noneinput_embeds: typing.Optional[torch.Tensor] = Noneuse_cache: typing.Optional[bool] = Noneoutput_attentions: typing.Optional[bool] = Noneoutput_hidden_states: typing.Optional[bool] = Nonereturn_dict: typing.Optional[bool] = None )
Parameters
- input_ids ( - torch.LongTensorof shape- (batch_size, sequence_length)) β Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide it. Indices can be obtained using AutoTokenizer. See PreTrainedTokenizer.encode() and PreTrainedTokenizer.call() for details. What are input IDs?
- past_key_values ( - tuple(tuple(torch.FloatTensor)), optional, returned when- use_cacheis passed or when- config.use_cache=True) β Tuple of- tuple(torch.FloatTensor)of length- config.n_layers, with each tuple having 2 tensors of shape- (batch_size, num_heads, sequence_length, embed_size_per_head).- Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see - past_key_valuesinput) to speed up sequential decoding.- If - past_key_valuesare used, the user can optionally input only the last- decoder_input_ids(those that donβt have their past key value states given to this model) of shape- (batch_size, 1)instead of all- input_idsof shape- (batch_size, sequence_length).
- attention_mask ( - torch.Tensorof shape- (batch_size, sequence_length), optional) β Mask to avoid performing attention on padding token indices. Mask values selected in- [0, 1]:- 1 for tokens that are not masked, 
- 0 for tokens that are masked. 
 
- position_ids ( - torch.LongTensorof shape- (batch_size, sequence_length), optional) β Indices of positions of each input sequence tokens in the position embeddings. Selected in the range- [0, config.max_position_embeddings - 1].
- head_mask ( - torch.Tensorof shape- (encoder_layers, encoder_attention_heads), optional) β Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in- [0, 1]:- 1 indicates the head is not masked, 
- 0 indicates the head is masked. 
 
- input_embeds ( - torch.FloatTensorof shape- (batch_size, input_sequence_length, hidden_size), optional) β Optionally, instead of passing- input_idsyou can choose to directly pass an embedded representation. Here, due to- Barkparticularities, if- past_key_valuesis used,- input_embedswill be ignored and you have to use- input_ids. If- past_key_valuesis not used and- use_cacheis set to- True,- input_embedsis used in priority instead of- input_ids.
- use_cache ( - bool, optional) β If set to- True,- past_key_valueskey value states are returned and can be used to speed up decoding (see- past_key_values).
- output_attentions ( - bool, optional) β Whether or not to return the attentions tensors of all attention layers. See- attentionsunder returned tensors for more detail.
- output_hidden_states ( - bool, optional) β Whether or not to return the hidden states of all layers. See- hidden_statesunder returned tensors for more detail.
- return_dict ( - bool, optional) β Whether or not to return a ModelOutput instead of a plain tuple.
The BarkCausalModel forward method, overrides the __call__ special method.
Although the recipe for forward pass needs to be defined within this function, one should call the Module instance afterwards instead of this since the former takes care of running the pre and post processing steps while the latter silently ignores them.
BarkCoarseModel
class transformers.BarkCoarseModel
( config )
Parameters
- config (BarkCoarseConfig) β Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the from_pretrained() method to load the model weights. 
Bark coarse acoustics model. It shares the same architecture as the semantic (or text) model. It is a GPT-2 like autoregressive model with a language modeling head on top. This model inherits from PreTrainedModel. Check the superclass documentation for the generic methods the library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads etc.)
This model is also a PyTorch torch.nn.Module subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior.
forward
( input_ids: typing.Optional[torch.Tensor] = Nonepast_key_values: typing.Optional[typing.Tuple[torch.FloatTensor]] = Noneattention_mask: typing.Optional[torch.Tensor] = Noneposition_ids: typing.Optional[torch.Tensor] = Nonehead_mask: typing.Optional[torch.Tensor] = Nonelabels: typing.Optional[torch.LongTensor] = Noneinput_embeds: typing.Optional[torch.Tensor] = Noneuse_cache: typing.Optional[bool] = Noneoutput_attentions: typing.Optional[bool] = Noneoutput_hidden_states: typing.Optional[bool] = Nonereturn_dict: typing.Optional[bool] = None )
Parameters
- input_ids ( - torch.LongTensorof shape- (batch_size, sequence_length)) β Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide it. Indices can be obtained using AutoTokenizer. See PreTrainedTokenizer.encode() and PreTrainedTokenizer.call() for details. What are input IDs?
- past_key_values ( - tuple(tuple(torch.FloatTensor)), optional, returned when- use_cacheis passed or when- config.use_cache=True) β Tuple of- tuple(torch.FloatTensor)of length- config.n_layers, with each tuple having 2 tensors of shape- (batch_size, num_heads, sequence_length, embed_size_per_head).- Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see - past_key_valuesinput) to speed up sequential decoding.- If - past_key_valuesare used, the user can optionally input only the last- decoder_input_ids(those that donβt have their past key value states given to this model) of shape- (batch_size, 1)instead of all- input_idsof shape- (batch_size, sequence_length).
- attention_mask ( - torch.Tensorof shape- (batch_size, sequence_length), optional) β Mask to avoid performing attention on padding token indices. Mask values selected in- [0, 1]:- 1 for tokens that are not masked, 
- 0 for tokens that are masked. 
 
- position_ids ( - torch.LongTensorof shape- (batch_size, sequence_length), optional) β Indices of positions of each input sequence tokens in the position embeddings. Selected in the range- [0, config.max_position_embeddings - 1].
- head_mask ( - torch.Tensorof shape- (encoder_layers, encoder_attention_heads), optional) β Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in- [0, 1]:- 1 indicates the head is not masked, 
- 0 indicates the head is masked. 
 
- input_embeds ( - torch.FloatTensorof shape- (batch_size, input_sequence_length, hidden_size), optional) β Optionally, instead of passing- input_idsyou can choose to directly pass an embedded representation. Here, due to- Barkparticularities, if- past_key_valuesis used,- input_embedswill be ignored and you have to use- input_ids. If- past_key_valuesis not used and- use_cacheis set to- True,- input_embedsis used in priority instead of- input_ids.
- use_cache ( - bool, optional) β If set to- True,- past_key_valueskey value states are returned and can be used to speed up decoding (see- past_key_values).
- output_attentions ( - bool, optional) β Whether or not to return the attentions tensors of all attention layers. See- attentionsunder returned tensors for more detail.
- output_hidden_states ( - bool, optional) β Whether or not to return the hidden states of all layers. See- hidden_statesunder returned tensors for more detail.
- return_dict ( - bool, optional) β Whether or not to return a ModelOutput instead of a plain tuple.
The BarkCausalModel forward method, overrides the __call__ special method.
Although the recipe for forward pass needs to be defined within this function, one should call the Module instance afterwards instead of this since the former takes care of running the pre and post processing steps while the latter silently ignores them.
BarkFineModel
class transformers.BarkFineModel
( config )
Parameters
- config (BarkFineConfig) β Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the from_pretrained() method to load the model weights. 
Bark fine acoustics model. It is a non-causal GPT-like model with config.n_codes_total embedding layers and language modeling heads, one for each codebook. This model inherits from PreTrainedModel. Check the superclass documentation for the generic methods the library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads etc.)
This model is also a PyTorch torch.nn.Module subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior.
forward
( codebook_idx: intinput_ids: typing.Optional[torch.Tensor] = Noneattention_mask: typing.Optional[torch.Tensor] = Noneposition_ids: typing.Optional[torch.Tensor] = Nonehead_mask: typing.Optional[torch.Tensor] = Nonelabels: typing.Optional[torch.LongTensor] = Noneinput_embeds: typing.Optional[torch.Tensor] = Noneoutput_attentions: typing.Optional[bool] = Noneoutput_hidden_states: typing.Optional[bool] = Nonereturn_dict: typing.Optional[bool] = None )
Parameters
- codebook_idx ( - int) β Index of the codebook that will be predicted.
- input_ids ( - torch.LongTensorof shape- (batch_size, sequence_length, number_of_codebooks)) β Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide it. Initially, indices of the first two codebooks are obtained from the- coarsesub-model. The rest is predicted recursively by attending the previously predicted channels. The model predicts on windows of length 1024.
- attention_mask ( - torch.Tensorof shape- (batch_size, sequence_length), optional) β Mask to avoid performing attention on padding token indices. Mask values selected in- [0, 1]:- 1 for tokens that are not masked, 
- 0 for tokens that are masked. 
 
- position_ids ( - torch.LongTensorof shape- (batch_size, sequence_length), optional) β Indices of positions of each input sequence tokens in the position embeddings. Selected in the range- [0, config.max_position_embeddings - 1].
- head_mask ( - torch.Tensorof shape- (encoder_layers, encoder_attention_heads), optional) β Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in- [0, 1]:- 1 indicates the head is not masked, 
- 0 indicates the head is masked. 
 
- labels ( - torch.LongTensorof shape- (batch_size, sequence_length), optional) β NOT IMPLEMENTED YET.
- input_embeds ( - torch.FloatTensorof shape- (batch_size, input_sequence_length, hidden_size), optional) β Optionally, instead of passing- input_idsyou can choose to directly pass an embedded representation. If- past_key_valuesis used, optionally only the last- input_embedshave to be input (see- past_key_values). This is useful if you want more control over how to convert- input_idsindices into associated vectors than the modelβs internal embedding lookup matrix.
- output_attentions ( - bool, optional) β Whether or not to return the attentions tensors of all attention layers. See- attentionsunder returned tensors for more detail.
- output_hidden_states ( - bool, optional) β Whether or not to return the hidden states of all layers. See- hidden_statesunder returned tensors for more detail.
- return_dict ( - bool, optional) β Whether or not to return a ModelOutput instead of a plain tuple.
The BarkFineModel forward method, overrides the __call__ special method.
Although the recipe for forward pass needs to be defined within this function, one should call the Module instance afterwards instead of this since the former takes care of running the pre and post processing steps while the latter silently ignores them.
BarkCausalModel
class transformers.BarkCausalModel
( config )
forward
( input_ids: typing.Optional[torch.Tensor] = Nonepast_key_values: typing.Optional[typing.Tuple[torch.FloatTensor]] = Noneattention_mask: typing.Optional[torch.Tensor] = Noneposition_ids: typing.Optional[torch.Tensor] = Nonehead_mask: typing.Optional[torch.Tensor] = Nonelabels: typing.Optional[torch.LongTensor] = Noneinput_embeds: typing.Optional[torch.Tensor] = Noneuse_cache: typing.Optional[bool] = Noneoutput_attentions: typing.Optional[bool] = Noneoutput_hidden_states: typing.Optional[bool] = Nonereturn_dict: typing.Optional[bool] = None )
Parameters
- input_ids ( - torch.LongTensorof shape- (batch_size, sequence_length)) β Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide it. Indices can be obtained using AutoTokenizer. See PreTrainedTokenizer.encode() and PreTrainedTokenizer.call() for details. What are input IDs?
- past_key_values ( - tuple(tuple(torch.FloatTensor)), optional, returned when- use_cacheis passed or when- config.use_cache=True) β Tuple of- tuple(torch.FloatTensor)of length- config.n_layers, with each tuple having 2 tensors of shape- (batch_size, num_heads, sequence_length, embed_size_per_head).- Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see - past_key_valuesinput) to speed up sequential decoding.- If - past_key_valuesare used, the user can optionally input only the last- decoder_input_ids(those that donβt have their past key value states given to this model) of shape- (batch_size, 1)instead of all- input_idsof shape- (batch_size, sequence_length).
- attention_mask ( - torch.Tensorof shape- (batch_size, sequence_length), optional) β Mask to avoid performing attention on padding token indices. Mask values selected in- [0, 1]:- 1 for tokens that are not masked, 
- 0 for tokens that are masked. 
 
- position_ids ( - torch.LongTensorof shape- (batch_size, sequence_length), optional) β Indices of positions of each input sequence tokens in the position embeddings. Selected in the range- [0, config.max_position_embeddings - 1].
- head_mask ( - torch.Tensorof shape- (encoder_layers, encoder_attention_heads), optional) β Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in- [0, 1]:- 1 indicates the head is not masked, 
- 0 indicates the head is masked. 
 
- input_embeds ( - torch.FloatTensorof shape- (batch_size, input_sequence_length, hidden_size), optional) β Optionally, instead of passing- input_idsyou can choose to directly pass an embedded representation. Here, due to- Barkparticularities, if- past_key_valuesis used,- input_embedswill be ignored and you have to use- input_ids. If- past_key_valuesis not used and- use_cacheis set to- True,- input_embedsis used in priority instead of- input_ids.
- use_cache ( - bool, optional) β If set to- True,- past_key_valueskey value states are returned and can be used to speed up decoding (see- past_key_values).
- output_attentions ( - bool, optional) β Whether or not to return the attentions tensors of all attention layers. See- attentionsunder returned tensors for more detail.
- output_hidden_states ( - bool, optional) β Whether or not to return the hidden states of all layers. See- hidden_statesunder returned tensors for more detail.
- return_dict ( - bool, optional) β Whether or not to return a ModelOutput instead of a plain tuple.
The BarkCausalModel forward method, overrides the __call__ special method.
Although the recipe for forward pass needs to be defined within this function, one should call the Module instance afterwards instead of this since the former takes care of running the pre and post processing steps while the latter silently ignores them.
BarkCoarseConfig
class transformers.BarkCoarseConfig
( block_size = 1024input_vocab_size = 10048output_vocab_size = 10048num_layers = 12num_heads = 12hidden_size = 768dropout = 0.0bias = Trueinitializer_range = 0.02use_cache = True**kwargs )
Parameters
- block_size ( - int, optional, defaults to 1024) β The maximum sequence length that this model might ever be used with. Typically set this to something large just in case (e.g., 512 or 1024 or 2048).
- input_vocab_size ( - int, optional, defaults to 10_048) β Vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the- inputs_idspassed when calling BarkCoarseModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- output_vocab_size ( - int, optional, defaults to 10_048) β Output vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the:- output_idswhen passing forward a BarkCoarseModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- num_layers ( - int, optional, defaults to 12) β Number of hidden layers in the given sub-model.
- num_heads ( - int, optional, defaults to 12) β Number of attention heads for each attention layer in the Transformer architecture.
- hidden_size ( - int, optional, defaults to 768) β Dimensionality of the βintermediateβ (often named feed-forward) layer in the architecture.
- dropout ( - float, optional, defaults to 0.0) β The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
- bias ( - bool, optional, defaults to- True) β Whether or not to use bias in the linear layers and layer norm layers.
- initializer_range ( - float, optional, defaults to 0.02) β The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
- use_cache ( - bool, optional, defaults to- True) β Whether or not the model should return the last key/values attentions (not used by all models).
This is the configuration class to store the configuration of a BarkCoarseModel. It is used to instantiate the model according to the specified arguments, defining the model architecture. Instantiating a configuration with the defaults will yield a similar configuration to that of the Bark suno/bark architecture.
Configuration objects inherit from PretrainedConfig and can be used to control the model outputs. Read the documentation from PretrainedConfig for more information.
Example:
Copied
>>> from transformers import BarkCoarseConfig, BarkCoarseModel
>>> # Initializing a Bark sub-module style configuration
>>> configuration = BarkCoarseConfig()
>>> # Initializing a model (with random weights) from the suno/bark style configuration
>>> model = BarkCoarseModel(configuration)
>>> # Accessing the model configuration
>>> configuration = model.configBarkFineConfig
class transformers.BarkFineConfig
( tie_word_embeddings = Truen_codes_total = 8n_codes_given = 1**kwargs )
Parameters
- block_size ( - int, optional, defaults to 1024) β The maximum sequence length that this model might ever be used with. Typically set this to something large just in case (e.g., 512 or 1024 or 2048).
- input_vocab_size ( - int, optional, defaults to 10_048) β Vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the- inputs_idspassed when calling BarkFineModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- output_vocab_size ( - int, optional, defaults to 10_048) β Output vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the:- output_idswhen passing forward a BarkFineModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- num_layers ( - int, optional, defaults to 12) β Number of hidden layers in the given sub-model.
- num_heads ( - int, optional, defaults to 12) β Number of attention heads for each attention layer in the Transformer architecture.
- hidden_size ( - int, optional, defaults to 768) β Dimensionality of the βintermediateβ (often named feed-forward) layer in the architecture.
- dropout ( - float, optional, defaults to 0.0) β The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
- bias ( - bool, optional, defaults to- True) β Whether or not to use bias in the linear layers and layer norm layers.
- initializer_range ( - float, optional, defaults to 0.02) β The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
- use_cache ( - bool, optional, defaults to- True) β Whether or not the model should return the last key/values attentions (not used by all models).
- n_codes_total ( - int, optional, defaults to 8) β The total number of audio codebooks predicted. Used in the fine acoustics sub-model.
- n_codes_given ( - int, optional, defaults to 1) β The number of audio codebooks predicted in the coarse acoustics sub-model. Used in the acoustics sub-models.
This is the configuration class to store the configuration of a BarkFineModel. It is used to instantiate the model according to the specified arguments, defining the model architecture. Instantiating a configuration with the defaults will yield a similar configuration to that of the Bark suno/bark architecture.
Configuration objects inherit from PretrainedConfig and can be used to control the model outputs. Read the documentation from PretrainedConfig for more information.
Example:
Copied
>>> from transformers import BarkFineConfig, BarkFineModel
>>> # Initializing a Bark sub-module style configuration
>>> configuration = BarkFineConfig()
>>> # Initializing a model (with random weights) from the suno/bark style configuration
>>> model = BarkFineModel(configuration)
>>> # Accessing the model configuration
>>> configuration = model.configBarkSemanticConfig
class transformers.BarkSemanticConfig
( block_size = 1024input_vocab_size = 10048output_vocab_size = 10048num_layers = 12num_heads = 12hidden_size = 768dropout = 0.0bias = Trueinitializer_range = 0.02use_cache = True**kwargs )
Parameters
- block_size ( - int, optional, defaults to 1024) β The maximum sequence length that this model might ever be used with. Typically set this to something large just in case (e.g., 512 or 1024 or 2048).
- input_vocab_size ( - int, optional, defaults to 10_048) β Vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the- inputs_idspassed when calling BarkSemanticModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- output_vocab_size ( - int, optional, defaults to 10_048) β Output vocabulary size of a Bark sub-model. Defines the number of different tokens that can be represented by the:- output_idswhen passing forward a BarkSemanticModel. Defaults to 10_048 but should be carefully thought with regards to the chosen sub-model.
- num_layers ( - int, optional, defaults to 12) β Number of hidden layers in the given sub-model.
- num_heads ( - int, optional, defaults to 12) β Number of attention heads for each attention layer in the Transformer architecture.
- hidden_size ( - int, optional, defaults to 768) β Dimensionality of the βintermediateβ (often named feed-forward) layer in the architecture.
- dropout ( - float, optional, defaults to 0.0) β The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
- bias ( - bool, optional, defaults to- True) β Whether or not to use bias in the linear layers and layer norm layers.
- initializer_range ( - float, optional, defaults to 0.02) β The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
- use_cache ( - bool, optional, defaults to- True) β Whether or not the model should return the last key/values attentions (not used by all models).
This is the configuration class to store the configuration of a BarkSemanticModel. It is used to instantiate the model according to the specified arguments, defining the model architecture. Instantiating a configuration with the defaults will yield a similar configuration to that of the Bark suno/bark architecture.
Configuration objects inherit from PretrainedConfig and can be used to control the model outputs. Read the documentation from PretrainedConfig for more information.
Example:
Copied
>>> from transformers import BarkSemanticConfig, BarkSemanticModel
>>> # Initializing a Bark sub-module style configuration
>>> configuration = BarkSemanticConfig()
>>> # Initializing a model (with random weights) from the suno/bark style configuration
>>> model = BarkSemanticModel(configuration)
>>> # Accessing the model configuration
>>> configuration = model.configLast updated
